Hybrid Surface Mesh Adaptation for Climate Modeling
Khamayseh, Ahmed K; de Almeida, Valmor F; Hansen, Glen
2008-01-01
Solution-driven mesh adaptation is becoming quite popular for spatial error control in the numerical simulation of complex computational physics applications, such as climate modeling. Typically, spatial adaptation is achieved by element subdivision (h adaptation) with a primary goal of resolving the local length scales of interest. A second, less-popular method of spatial adaptivity is called "mesh motion" (r adaptation); the smooth repositioning of mesh node points aimed at resizing existing elements to capture the local length scales. This paper proposes an adaptation method based on a combination of both element subdivision and node point repositioning (rh adaptation). By combining these two methods using the notion of a mobility function, the proposed approach seeks to increase the flexibility and extensibility of mesh motion algorithms while providing a somewhat smoother transition between refined regions than is produced by element subdivision alone. Further, in an attempt to support the requirements of a very general class of climate simulation applications, the proposed method is designed to accommodate unstructured, polygonal mesh topologies in addition to the most popular mesh types.
Hybrid Surface Mesh Adaptation for Climate Modeling
Ahmed Khamayseh; Valmor de Almeida; Glen Hansen
2008-10-01
Solution-driven mesh adaptation is becoming quite popular for spatial error control in the numerical simulation of complex computational physics applications, such as climate modeling. Typically, spatial adaptation is achieved by element subdivision (h adaptation) with a primary goal of resolving the local length scales of interest. A second, less-popular method of spatial adaptivity is called “mesh motion” (r adaptation); the smooth repositioning of mesh node points aimed at resizing existing elements to capture the local length scales. This paper proposes an adaptation method based on a combination of both element subdivision and node point repositioning (rh adaptation). By combining these two methods using the notion of a mobility function, the proposed approach seeks to increase the flexibility and extensibility of mesh motion algorithms while providing a somewhat smoother transition between refined regions than is produced by element subdivision alone. Further, in an attempt to support the requirements of a very general class of climate simulation applications, the proposed method is designed to accommodate unstructured, polygonal mesh topologies in addition to the most popular mesh types.
Numerical modeling of seismic waves using frequency-adaptive meshes
NASA Astrophysics Data System (ADS)
Hu, Jinyin; Jia, Xiaofeng
2016-08-01
An improved modeling algorithm using frequency-adaptive meshes is applied to meet the computational requirements of all seismic frequency components. It automatically adopts coarse meshes for low-frequency computations and fine meshes for high-frequency computations. The grid intervals are adaptively calculated based on a smooth inversely proportional function of grid size with respect to the frequency. In regular grid-based methods, the uniform mesh or non-uniform mesh is used for frequency-domain wave propagators and it is fixed for all frequencies. A too coarse mesh results in inaccurate high-frequency wavefields and unacceptable numerical dispersion; on the other hand, an overly fine mesh may cause storage and computational overburdens as well as invalid propagation angles of low-frequency wavefields. Experiments on the Padé generalized screen propagator indicate that the Adaptive mesh effectively solves these drawbacks of regular fixed-mesh methods, thus accurately computing the wavefield and its propagation angle in a wide frequency band. Several synthetic examples also demonstrate its feasibility for seismic modeling and migration.
Particle systems for adaptive, isotropic meshing of CAD models
Levine, Joshua A.; Whitaker, Ross T.
2012-01-01
We present a particle-based approach for generating adaptive triangular surface and tetrahedral volume meshes from computer-aided design models. Input shapes are treated as a collection of smooth, parametric surface patches that can meet non-smoothly on boundaries. Our approach uses a hierarchical sampling scheme that places particles on features in order of increasing dimensionality. These particles reach a good distribution by minimizing an energy computed in 3D world space, with movements occurring in the parametric space of each surface patch. Rather than using a pre-computed measure of feature size, our system automatically adapts to both curvature as well as a notion of topological separation. It also enforces a measure of smoothness on these constraints to construct a sizing field that acts as a proxy to piecewise-smooth feature size. We evaluate our technique with comparisons against other popular triangular meshing techniques for this domain. PMID:23162181
Thermal-chemical Mantle Convection Models With Adaptive Mesh Refinement
NASA Astrophysics Data System (ADS)
Leng, W.; Zhong, S.
2008-12-01
In numerical modeling of mantle convection, resolution is often crucial for resolving small-scale features. New techniques, adaptive mesh refinement (AMR), allow local mesh refinement wherever high resolution is needed, while leaving other regions with relatively low resolution. Both computational efficiency for large- scale simulation and accuracy for small-scale features can thus be achieved with AMR. Based on the octree data structure [Tu et al. 2005], we implement the AMR techniques into the 2-D mantle convection models. For pure thermal convection models, benchmark tests show that our code can achieve high accuracy with relatively small number of elements both for isoviscous cases (i.e. 7492 AMR elements v.s. 65536 uniform elements) and for temperature-dependent viscosity cases (i.e. 14620 AMR elements v.s. 65536 uniform elements). We further implement tracer-method into the models for simulating thermal-chemical convection. By appropriately adding and removing tracers according to the refinement of the meshes, our code successfully reproduces the benchmark results in van Keken et al. [1997] with much fewer elements and tracers compared with uniform-mesh models (i.e. 7552 AMR elements v.s. 16384 uniform elements, and ~83000 tracers v.s. ~410000 tracers). The boundaries of the chemical piles in our AMR code can be easily refined to the scales of a few kilometers for the Earth's mantle and the tracers are concentrated near the chemical boundaries to precisely trace the evolvement of the boundaries. It is thus very suitable for our AMR code to study the thermal-chemical convection problems which need high resolution to resolve the evolvement of chemical boundaries, such as the entrainment problems [Sleep, 1988].
Algebraic turbulence modeling for unstructured and adaptive meshes
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1990-01-01
An algebraic turbulence model based on the Baldwin-Lomax model, has been implemented for use on unstructured grids. The implementation is based on the use of local background structured turbulence meshes. At each time-step, flow variables are interpolated from the unstructured mesh onto the background structured meshes, the turbulence model is executed on these meshes, and the resulting eddy viscosity values are interpolated back to the unstructured mesh. Modifications to the algebraic model were required to enable the treatment of more complicated flows, such as confluent boundary layers and wakes. The model is used in conjuction with an efficient unstructured multigrid finite-element Navier-Stokes solver in order to compute compressible turbulent flows on fully unstructured meshes. Solutions about single and multiple element airfoils are obtained and compared with experimental data.
NASA Technical Reports Server (NTRS)
Stapleton, Scott; Gries, Thomas; Waas, Anthony M.; Pineda, Evan J.
2014-01-01
Enhanced finite elements are elements with an embedded analytical solution that can capture detailed local fields, enabling more efficient, mesh independent finite element analysis. The shape functions are determined based on the analytical model rather than prescribed. This method was applied to adhesively bonded joints to model joint behavior with one element through the thickness. This study demonstrates two methods of maintaining the fidelity of such elements during adhesive non-linearity and cracking without increasing the mesh needed for an accurate solution. The first method uses adaptive shape functions, where the shape functions are recalculated at each load step based on the softening of the adhesive. The second method is internal mesh adaption, where cracking of the adhesive within an element is captured by further discretizing the element internally to represent the partially cracked geometry. By keeping mesh adaptations within an element, a finer mesh can be used during the analysis without affecting the global finite element model mesh. Examples are shown which highlight when each method is most effective in reducing the number of elements needed to capture adhesive nonlinearity and cracking. These methods are validated against analogous finite element models utilizing cohesive zone elements.
Parallel Adaptive Mesh Refinement
Diachin, L; Hornung, R; Plassmann, P; WIssink, A
2005-03-04
As large-scale, parallel computers have become more widely available and numerical models and algorithms have advanced, the range of physical phenomena that can be simulated has expanded dramatically. Many important science and engineering problems exhibit solutions with localized behavior where highly-detailed salient features or large gradients appear in certain regions which are separated by much larger regions where the solution is smooth. Examples include chemically-reacting flows with radiative heat transfer, high Reynolds number flows interacting with solid objects, and combustion problems where the flame front is essentially a two-dimensional sheet occupying a small part of a three-dimensional domain. Modeling such problems numerically requires approximating the governing partial differential equations on a discrete domain, or grid. Grid spacing is an important factor in determining the accuracy and cost of a computation. A fine grid may be needed to resolve key local features while a much coarser grid may suffice elsewhere. Employing a fine grid everywhere may be inefficient at best and, at worst, may make an adequately resolved simulation impractical. Moreover, the location and resolution of fine grid required for an accurate solution is a dynamic property of a problem's transient features and may not be known a priori. Adaptive mesh refinement (AMR) is a technique that can be used with both structured and unstructured meshes to adjust local grid spacing dynamically to capture solution features with an appropriate degree of resolution. Thus, computational resources can be focused where and when they are needed most to efficiently achieve an accurate solution without incurring the cost of a globally-fine grid. Figure 1.1 shows two example computations using AMR; on the left is a structured mesh calculation of a impulsively-sheared contact surface and on the right is the fuselage and volume discretization of an RAH-66 Comanche helicopter [35]. Note the
Parallel Adaptive Mesh Refinement Library
NASA Technical Reports Server (NTRS)
Mac-Neice, Peter; Olson, Kevin
2005-01-01
Parallel Adaptive Mesh Refinement Library (PARAMESH) is a package of Fortran 90 subroutines designed to provide a computer programmer with an easy route to extension of (1) a previously written serial code that uses a logically Cartesian structured mesh into (2) a parallel code with adaptive mesh refinement (AMR). Alternatively, in its simplest use, and with minimal effort, PARAMESH can operate as a domain-decomposition tool for users who want to parallelize their serial codes but who do not wish to utilize adaptivity. The package builds a hierarchy of sub-grids to cover the computational domain of a given application program, with spatial resolution varying to satisfy the demands of the application. The sub-grid blocks form the nodes of a tree data structure (a quad-tree in two or an oct-tree in three dimensions). Each grid block has a logically Cartesian mesh. The package supports one-, two- and three-dimensional models.
Advanced numerical methods in mesh generation and mesh adaptation
Lipnikov, Konstantine; Danilov, A; Vassilevski, Y; Agonzal, A
2010-01-01
Numerical solution of partial differential equations requires appropriate meshes, efficient solvers and robust and reliable error estimates. Generation of high-quality meshes for complex engineering models is a non-trivial task. This task is made more difficult when the mesh has to be adapted to a problem solution. This article is focused on a synergistic approach to the mesh generation and mesh adaptation, where best properties of various mesh generation methods are combined to build efficiently simplicial meshes. First, the advancing front technique (AFT) is combined with the incremental Delaunay triangulation (DT) to build an initial mesh. Second, the metric-based mesh adaptation (MBA) method is employed to improve quality of the generated mesh and/or to adapt it to a problem solution. We demonstrate with numerical experiments that combination of all three methods is required for robust meshing of complex engineering models. The key to successful mesh generation is the high-quality of the triangles in the initial front. We use a black-box technique to improve surface meshes exported from an unattainable CAD system. The initial surface mesh is refined into a shape-regular triangulation which approximates the boundary with the same accuracy as the CAD mesh. The DT method adds robustness to the AFT. The resulting mesh is topologically correct but may contain a few slivers. The MBA uses seven local operations to modify the mesh topology. It improves significantly the mesh quality. The MBA method is also used to adapt the mesh to a problem solution to minimize computational resources required for solving the problem. The MBA has a solid theoretical background. In the first two experiments, we consider the convection-diffusion and elasticity problems. We demonstrate the optimal reduction rate of the discretization error on a sequence of adaptive strongly anisotropic meshes. The key element of the MBA method is construction of a tensor metric from hierarchical edge
Conditional entropy maximization for PET image reconstruction using adaptive mesh model.
Zhu, Hongqing; Shu, Huazhong; Zhou, Jian; Dai, Xiubin; Luo, Limin
2007-04-01
Iterative image reconstruction algorithms have been widely used in the field of positron emission tomography (PET). However, such algorithms are sensitive to noise artifacts so that the reconstruction begins to degrade when the number of iterations is high. In this paper, we propose a new algorithm to reconstruct an image from the PET emission projection data by using the conditional entropy maximization and the adaptive mesh model. In a traditional tomography reconstruction method, the reconstructed image is directly computed in the pixel domain. Unlike this kind of methods, the proposed approach is performed by estimating the nodal values from the observed projection data in a mesh domain. In our method, the initial Delaunay triangulation mesh is generated from a set of randomly selected pixel points, and it is then modified according to the pixel intensity value of the estimated image at each iteration step in which the conditional entropy maximization is used. The advantage of using the adaptive mesh model for image reconstruction is that it provides a natural spatially adaptive smoothness mechanism. In experiments using the synthetic and clinical data, it is found that the proposed algorithm is more robust to noise compared to the common pixel-based MLEM algorithm and mesh-based MLEM with a fixed mesh structure. PMID:17368841
Unstructured mesh generation and adaptivity
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1995-01-01
An overview of current unstructured mesh generation and adaptivity techniques is given. Basic building blocks taken from the field of computational geometry are first described. Various practical mesh generation techniques based on these algorithms are then constructed and illustrated with examples. Issues of adaptive meshing and stretched mesh generation for anisotropic problems are treated in subsequent sections. The presentation is organized in an education manner, for readers familiar with computational fluid dynamics, wishing to learn more about current unstructured mesh techniques.
Ibrahim, Ahmad M.; Wilson, Paul P.H.; Sawan, Mohamed E.; Mosher, Scott W.; Peplow, Douglas E.; Wagner, John C.; Evans, Thomas M.; Grove, Robert E.
2015-06-30
The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as muchmore » geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, eliminating the need for a world-class super computer.« less
Ibrahim, Ahmad M.; Wilson, Paul P.H.; Sawan, Mohamed E.; Mosher, Scott W.; Peplow, Douglas E.; Wagner, John C.; Evans, Thomas M.; Grove, Robert E.
2015-06-30
The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, eliminating the need for a world-class super computer.
Towards a large-scale scalable adaptive heart model using shallow tree meshes
NASA Astrophysics Data System (ADS)
Krause, Dorian; Dickopf, Thomas; Potse, Mark; Krause, Rolf
2015-10-01
Electrophysiological heart models are sophisticated computational tools that place high demands on the computing hardware due to the high spatial resolution required to capture the steep depolarization front. To address this challenge, we present a novel adaptive scheme for resolving the deporalization front accurately using adaptivity in space. Our adaptive scheme is based on locally structured meshes. These tensor meshes in space are organized in a parallel forest of trees, which allows us to resolve complicated geometries and to realize high variations in the local mesh sizes with a minimal memory footprint in the adaptive scheme. We discuss both a non-conforming mortar element approximation and a conforming finite element space and present an efficient technique for the assembly of the respective stiffness matrices using matrix representations of the inclusion operators into the product space on the so-called shallow tree meshes. We analyzed the parallel performance and scalability for a two-dimensional ventricle slice as well as for a full large-scale heart model. Our results demonstrate that the method has good performance and high accuracy.
A Predictive Model of Fragmentation using Adaptive Mesh Refinement and a Hierarchical Material Model
Koniges, A E; Masters, N D; Fisher, A C; Anderson, R W; Eder, D C; Benson, D; Kaiser, T B; Gunney, B T; Wang, P; Maddox, B R; Hansen, J F; Kalantar, D H; Dixit, P; Jarmakani, H; Meyers, M A
2009-03-03
Fragmentation is a fundamental material process that naturally spans spatial scales from microscopic to macroscopic. We developed a mathematical framework using an innovative combination of hierarchical material modeling (HMM) and adaptive mesh refinement (AMR) to connect the continuum to microstructural regimes. This framework has been implemented in a new multi-physics, multi-scale, 3D simulation code, NIF ALE-AMR. New multi-material volume fraction and interface reconstruction algorithms were developed for this new code, which is leading the world effort in hydrodynamic simulations that combine AMR with ALE (Arbitrary Lagrangian-Eulerian) techniques. The interface reconstruction algorithm is also used to produce fragments following material failure. In general, the material strength and failure models have history vector components that must be advected along with other properties of the mesh during remap stage of the ALE hydrodynamics. The fragmentation models are validated against an electromagnetically driven expanding ring experiment and dedicated laser-based fragmentation experiments conducted at the Jupiter Laser Facility. As part of the exit plan, the NIF ALE-AMR code was applied to a number of fragmentation problems of interest to the National Ignition Facility (NIF). One example shows the added benefit of multi-material ALE-AMR that relaxes the requirement that material boundaries must be along mesh boundaries.
Parallel automated adaptive procedures for unstructured meshes
NASA Technical Reports Server (NTRS)
Shephard, M. S.; Flaherty, J. E.; Decougny, H. L.; Ozturan, C.; Bottasso, C. L.; Beall, M. W.
1995-01-01
Consideration is given to the techniques required to support adaptive analysis of automatically generated unstructured meshes on distributed memory MIMD parallel computers. The key areas of new development are focused on the support of effective parallel computations when the structure of the numerical discretization, the mesh, is evolving, and in fact constructed, during the computation. All the procedures presented operate in parallel on already distributed mesh information. Starting from a mesh definition in terms of a topological hierarchy, techniques to support the distribution, redistribution and communication among the mesh entities over the processors is given, and algorithms to dynamically balance processor workload based on the migration of mesh entities are given. A procedure to automatically generate meshes in parallel, starting from CAD geometric models, is given. Parallel procedures to enrich the mesh through local mesh modifications are also given. Finally, the combination of these techniques to produce a parallel automated finite element analysis procedure for rotorcraft aerodynamics calculations is discussed and demonstrated.
Impact of space-time mesh adaptation on solute transport modeling in porous media
NASA Astrophysics Data System (ADS)
Esfandiar, Bahman; Porta, Giovanni; Perotto, Simona; Guadagnini, Alberto
2015-02-01
We implement a space-time grid adaptation procedure to efficiently improve the accuracy of numerical simulations of solute transport in porous media in the context of model parameter estimation. We focus on the Advection Dispersion Equation (ADE) for the interpretation of nonreactive transport experiments in laboratory-scale heterogeneous porous media. When compared to a numerical approximation based on a fixed space-time discretization, our approach is grounded on a joint automatic selection of the spatial grid and the time step to capture the main (space-time) system dynamics. Spatial mesh adaptation is driven by an anisotropic recovery-based error estimator which enables us to properly select the size, shape, and orientation of the mesh elements. Adaptation of the time step is performed through an ad hoc local reconstruction of the temporal derivative of the solution via a recovery-based approach. The impact of the proposed adaptation strategy on the ability to provide reliable estimates of the key parameters of an ADE model is assessed on the basis of experimental solute breakthrough data measured following tracer injection in a nonuniform porous system. Model calibration is performed in a Maximum Likelihood (ML) framework upon relying on the representation of the ADE solution through a generalized Polynomial Chaos Expansion (gPCE). Our results show that the proposed anisotropic space-time grid adaptation leads to ML parameter estimates and to model results of markedly improved quality when compared to classical inversion approaches based on a uniform space-time discretization.
NASA Astrophysics Data System (ADS)
Papadakis, A. P.; Georghiou, G. E.; Metaxas, A. C.
2008-12-01
A new adaptive mesh generator has been developed and used in the analysis of high-pressure gas discharges, such as avalanches and streamers, reducing computational times and computer memory needs significantly. The new adaptive mesh generator developed, uses normalized error indicators, varying from 0 to 1, to guarantee optimal mesh resolution for all carriers involved in the analysis. Furthermore, it uses h- and r-refinement techniques such as mesh jiggling, edge swapping and node addition/removal to develop an element quality improvement algorithm that improves the mesh quality significantly and a fast and accurate algorithm for interpolation between meshes. Finally, the mesh generator is applied in the characterization of the transition from a single electron to the avalanche and streamer discharges in high-voltage, high-pressure gas discharges for dc 1 mm gaps, RF 1 cm point-plane gaps and parallel-plate 40 MHz configurations, in ambient atmospheric air.
Numerical modelling of tsunami generation by deformable submarine slides using mesh adaptivity
NASA Astrophysics Data System (ADS)
Smith, Rebecca; Parkinson, Samuel; Hill, Jon; Collins, Gareth; Piggott, Matthew
2014-05-01
Tsunamis generated by submarine slides are often under considered in comparison to earthquake generated tsunami, despite several recent examples. Tsunamigenic slides have generated waves that have caused significant damage and loss of life, for example the 1998 Papua New Guinea submarine mass failure resulted in a tsunami that devastated coastal villages and killed over 2,100 people. Numerical simulations of submarine slide generated waves can help us understand the nature of the waves that are generated, and identify the important factors in determining wave characteristics. There have not been many studies of tsunami generation by deformable submarine slides, largely because of the complexities and computational expense involved in modelling these large scale events. At large, real world, scales modelling of tsunami waves by the generation of slides is computationally challenging. Fluidity is an open source finite element code that is ideally suited to tackle this type of problem as it uses unstructured, adaptive meshes, which help to reduce the computational expense without losing accuracy in the results. Adaptive meshes change topology and resolution based on the current simulation state and as such can focus or reduce resolution when and where it is required. The model also allows a number of different numerical approaches to be taken to simulate the same problem within the same numerical framework. In this example we use multi-material approach, with both two materials (slide and water) and three materials (slide, water and air), alongside a density-driven sediment model approach. We will present results of validating Fluidity against benchmarks from experimental and other numerical studies, at different scales, for deformable underwater slides, and consider the utility of mesh adaptivity. We show good agreement to both laboratory results and other numerical models, both with a fixed mesh and a dynamically adaptive mesh, tracking important features of the
Cubit Adaptive Meshing Algorithm Library
2004-09-01
CAMAL (Cubit adaptive meshing algorithm library) is a software component library for mesh generation. CAMAL 2.0 includes components for triangle, quad and tetrahedral meshing. A simple Application Programmers Interface (API) takes a discrete boundary definition and CAMAL computes a quality interior unstructured grid. The triangle and quad algorithms may also import a geometric definition of a surface on which to define the grid. CAMALs triangle meshing uses a 3D space advancing front method, the quadmore » meshing algorithm is based upon Sandias patented paving algorithm and the tetrahedral meshing algorithm employs the GHS3D-Tetmesh component developed by INRIA, France.« less
Numerical modeling of landslide-generated tsunami using adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Wilson, Cian; Collins, Gareth; Desousa Costa, Patrick; Piggott, Matthew
2010-05-01
Landslides impacting into or occurring under water generate waves, which can have devastating environmental consequences. Depending on the characteristics of the landslide the waves can have significant amplitude and potentially propagate over large distances. Linear models of classical earthquake-generated tsunamis cannot reproduce the highly nonlinear generation mechanisms required to accurately predict the consequences of landslide-generated tsunamis. Also, laboratory-scale experimental investigation is limited to simple geometries and short time-scales before wave reflections contaminate the data. Computational fluid dynamics models based on the nonlinear Navier-Stokes equations can simulate landslide-tsunami generation at realistic scales. However, traditional chessboard-like structured meshes introduce superfluous resolution and hence the computing power required for such a simulation can be prohibitively high, especially in three dimensions. Unstructured meshes allow the grid spacing to vary rapidly from high resolution in the vicinity of small scale features to much coarser, lower resolution in other areas. Combining this variable resolution with dynamic mesh adaptivity allows such high resolution zones to follow features like the interface between the landslide and the water whilst minimising the computational costs. Unstructured meshes are also better suited to representing complex geometries and bathymetries allowing more realistic domains to be simulated. Modelling multiple materials, like water, air and a landslide, on an unstructured adaptive mesh poses significant numerical challenges. Novel methods of interface preservation must be considered and coupled to a flow model in such a way that ensures conservation of the different materials. Furthermore this conservation property must be maintained during successive stages of mesh optimisation and interpolation. In this paper we validate a new multi-material adaptive unstructured fluid dynamics model
Adaptive Mesh Expansion Model (AMEM) for Liver Segmentation from CT Image
Wang, Xuehu; Yang, Jian; Ai, Danni; Zheng, Yongchang; Tang, Songyuan; Wang, Yongtian
2015-01-01
This study proposes a novel adaptive mesh expansion model (AMEM) for liver segmentation from computed tomography images. The virtual deformable simplex model (DSM) is introduced to represent the mesh, in which the motion of each vertex can be easily manipulated. The balloon, edge, and gradient forces are combined with the binary image to construct the external force of the deformable model, which can rapidly drive the DSM to approach the target liver boundaries. Moreover, tangential and normal forces are combined with the gradient image to control the internal force, such that the DSM degree of smoothness can be precisely controlled. The triangular facet of the DSM is adaptively decomposed into smaller triangular components, which can significantly improve the segmentation accuracy of the irregularly sharp corners of the liver. The proposed method is evaluated on the basis of different criteria applied to 10 clinical data sets. Experiments demonstrate that the proposed AMEM algorithm is effective and robust and thus outperforms six other up-to-date algorithms. Moreover, AMEM can achieve a mean overlap error of 6.8% and a mean volume difference of 2.7%, whereas the average symmetric surface distance and the root mean square symmetric surface distance can reach 1.3 mm and 2.7 mm, respectively. PMID:25769030
Adaptive and Unstructured Mesh Cleaving
Bronson, Jonathan R.; Sastry, Shankar P.; Levine, Joshua A.; Whitaker, Ross T.
2015-01-01
We propose a new strategy for boundary conforming meshing that decouples the problem of building tetrahedra of proper size and shape from the problem of conforming to complex, non-manifold boundaries. This approach is motivated by the observation that while several methods exist for adaptive tetrahedral meshing, they typically have difficulty at geometric boundaries. The proposed strategy avoids this conflict by extracting the boundary conforming constraint into a secondary step. We first build a background mesh having a desired set of tetrahedral properties, and then use a generalized stenciling method to divide, or “cleave”, these elements to get a set of conforming tetrahedra, while limiting the impacts cleaving has on element quality. In developing this new framework, we make several technical contributions including a new method for building graded tetrahedral meshes as well as a generalization of the isosurface stuffing and lattice cleaving algorithms to unstructured background meshes. PMID:26137171
A Parallel Ocean Model With Adaptive Mesh Refinement Capability For Global Ocean Prediction
Herrnstein, A
2005-09-08
An ocean model with adaptive mesh refinement (AMR) capability is presented for simulating ocean circulation on decade time scales. The model closely resembles the LLNL ocean general circulation model with some components incorporated from other well known ocean models when appropriate. Spatial components are discretized using finite differences on a staggered grid where tracer and pressure variables are defined at cell centers and velocities at cell vertices (B-grid). Horizontal motion is modeled explicitly with leapfrog and Euler forward-backward time integration, and vertical motion is modeled semi-implicitly. New AMR strategies are presented for horizontal refinement on a B-grid, leapfrog time integration, and time integration of coupled systems with unequal time steps. These AMR capabilities are added to the LLNL software package SAMRAI (Structured Adaptive Mesh Refinement Application Infrastructure) and validated with standard benchmark tests. The ocean model is built on top of the amended SAMRAI library. The resulting model has the capability to dynamically increase resolution in localized areas of the domain. Limited basin tests are conducted using various refinement criteria and produce convergence trends in the model solution as refinement is increased. Carbon sequestration simulations are performed on decade time scales in domains the size of the North Atlantic and the global ocean. A suggestion is given for refinement criteria in such simulations. AMR predicts maximum pH changes and increases in CO{sub 2} concentration near the injection sites that are virtually unattainable with a uniform high resolution due to extremely long run times. Fine scale details near the injection sites are achieved by AMR with shorter run times than the finest uniform resolution tested despite the need for enhanced parallel performance. The North Atlantic simulations show a reduction in passive tracer errors when AMR is applied instead of a uniform coarse resolution. No
Adaptive Mesh Refinement in CTH
Crawford, David
1999-05-04
This paper reports progress on implementing a new capability of adaptive mesh refinement into the Eulerian multimaterial shock- physics code CTH. The adaptivity is block-based with refinement and unrefinement occurring in an isotropic 2:1 manner. The code is designed to run on serial, multiprocessor and massive parallel platforms. An approximate factor of three in memory and performance improvements over comparable resolution non-adaptive calculations has-been demonstrated for a number of problems.
Adaptive triangular mesh generation
NASA Technical Reports Server (NTRS)
Erlebacher, G.; Eiseman, P. R.
1984-01-01
A general adaptive grid algorithm is developed on triangular grids. The adaptivity is provided by a combination of node addition, dynamic node connectivity and a simple node movement strategy. While the local restructuring process and the node addition mechanism take place in the physical plane, the nodes are displaced on a monitor surface, constructed from the salient features of the physical problem. An approximation to mean curvature detects changes in the direction of the monitor surface, and provides the pulling force on the nodes. Solutions to the axisymmetric Grad-Shafranov equation demonstrate the capturing, by triangles, of the plasma-vacuum interface in a free-boundary equilibrium configuration.
Adaptive Mesh Refinement for Microelectronic Device Design
NASA Technical Reports Server (NTRS)
Cwik, Tom; Lou, John; Norton, Charles
1999-01-01
Finite element and finite volume methods are used in a variety of design simulations when it is necessary to compute fields throughout regions that contain varying materials or geometry. Convergence of the simulation can be assessed by uniformly increasing the mesh density until an observable quantity stabilizes. Depending on the electrical size of the problem, uniform refinement of the mesh may be computationally infeasible due to memory limitations. Similarly, depending on the geometric complexity of the object being modeled, uniform refinement can be inefficient since regions that do not need refinement add to the computational expense. In either case, convergence to the correct (measured) solution is not guaranteed. Adaptive mesh refinement methods attempt to selectively refine the region of the mesh that is estimated to contain proportionally higher solution errors. The refinement may be obtained by decreasing the element size (h-refinement), by increasing the order of the element (p-refinement) or by a combination of the two (h-p refinement). A successful adaptive strategy refines the mesh to produce an accurate solution measured against the correct fields without undue computational expense. This is accomplished by the use of a) reliable a posteriori error estimates, b) hierarchal elements, and c) automatic adaptive mesh generation. Adaptive methods are also useful when problems with multi-scale field variations are encountered. These occur in active electronic devices that have thin doped layers and also when mixed physics is used in the calculation. The mesh needs to be fine at and near the thin layer to capture rapid field or charge variations, but can coarsen away from these layers where field variations smoothen and charge densities are uniform. This poster will present an adaptive mesh refinement package that runs on parallel computers and is applied to specific microelectronic device simulations. Passive sensors that operate in the infrared portion of
Multiphase flow modelling of explosive volcanic eruptions using adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Jacobs, Christian T.; Collins, Gareth S.; Piggott, Matthew D.; Kramer, Stephan C.
2014-05-01
Explosive volcanic eruptions generate highly energetic plumes of hot gas and ash particles that produce diagnostic deposits and pose an extreme environmental hazard. The formation, dispersion and collapse of these volcanic plumes are complex multiscale processes that are extremely challenging to simulate numerically. Accurate description of particle and droplet aggregation, movement and settling requires a model capable of capturing the dynamics on a range of scales (from cm to km) and a model that can correctly describe the important multiphase interactions that take place. However, even the most advanced models of eruption dynamics to date are restricted by the fixed mesh-based approaches that they employ. The research presented herein describes the development of a compressible multiphase flow model within Fluidity, a combined finite element / control volume computational fluid dynamics (CFD) code, for the study of explosive volcanic eruptions. Fluidity adopts a state-of-the-art adaptive unstructured mesh-based approach to discretise the domain and focus numerical resolution only in areas important to the dynamics, while decreasing resolution where it is not needed as a simulation progresses. This allows the accurate but economical representation of the flow dynamics throughout time, and potentially allows large multi-scale problems to become tractable in complex 3D domains. The multiphase flow model is verified with the method of manufactured solutions, and validated by simulating published gas-solid shock tube experiments and comparing the numerical results against pressure gauge data. The application of the model considers an idealised 7 km by 7 km domain in which the violent eruption of hot gas and volcanic ash high into the atmosphere is simulated. Although the simulations do not correspond to a particular eruption case study, the key flow features observed in a typical explosive eruption event are successfully captured. These include a shock wave resulting
Modeling, mesh generation, and adaptive numerical methods for partial differential equations
Babuska, I.; Henshaw, W.D.; Oliger, J.E.; Flaherty, J.E.; Hopcroft, J.E.; Tezduyar, T.
1995-12-31
Mesh generation is one of the most time consuming aspects of computational solutions of problems involving partial differential equations. It is, furthermore, no longer acceptable to compute solutions without proper verification that specified accuracy criteria are being satisfied. Mesh generation must be related to the solution through computable estimates of discretization errors. Thus, an iterative process of alternate mesh and solution generation evolves in an adaptive manner with the end result that the solution is computed to prescribed specifications in an optimal, or at least efficient, manner. While mesh generation and adaptive strategies are becoming available, major computational challenges remain. One, in particular, involves moving boundaries and interfaces, such as free-surface flows and fluid-structure interactions. A 3-week program was held from July 5 to July 23, 1993 with 173 participants and 66 keynote, invited, and contributed presentations. This volume represents written versions of 21 of these lectures. These proceedings are organized roughly in order of their presentation at the workshop. Thus, the initial papers are concerned with geometry and mesh generation and discuss the representation of physical objects and surfaces on a computer and techniques to use this data to generate, principally, unstructured meshes of tetrahedral or hexahedral elements. The remainder of the papers cover adaptive strategies, error estimation, and applications. Several submissions deal with high-order p- and hp-refinement methods where mesh refinement/coarsening (h-refinement) is combined with local variation of method order (p-refinement). Combinations of mathematically verified and physically motivated approaches to error estimation are represented. Applications center on fluid mechanics. Selected papers are indexed separately for inclusion in the Energy Science and Technology Database.
Issues in adaptive mesh refinement
Dai, William Wenlong
2009-01-01
In this paper, we present an approach for a patch-based adaptive mesh refinement (AMR) for multi-physics simulations. The approach consists of clustering, symmetry preserving, mesh continuity, flux correction, communications, and management of patches. Among the special features of this patch-based AMR are symmetry preserving, efficiency of refinement, special implementation offlux correction, and patch management in parallel computing environments. Here, higher efficiency of refinement means less unnecessarily refined cells for a given set of cells to be refined. To demonstrate the capability of the AMR framework, hydrodynamics simulations with many levels of refinement are shown in both two- and three-dimensions.
toolkit computational mesh conceptual model.
Baur, David G.; Edwards, Harold Carter; Cochran, William K.; Williams, Alan B.; Sjaardema, Gregory D.
2010-03-01
The Sierra Toolkit computational mesh is a software library intended to support massively parallel multi-physics computations on dynamically changing unstructured meshes. This domain of intended use is inherently complex due to distributed memory parallelism, parallel scalability, heterogeneity of physics, heterogeneous discretization of an unstructured mesh, and runtime adaptation of the mesh. Management of this inherent complexity begins with a conceptual analysis and modeling of this domain of intended use; i.e., development of a domain model. The Sierra Toolkit computational mesh software library is designed and implemented based upon this domain model. Software developers using, maintaining, or extending the Sierra Toolkit computational mesh library must be familiar with the concepts/domain model presented in this report.
Adaptive Skin Meshes Coarsening for Biomolecular Simulation.
Shi, Xinwei; Koehl, Patrice
2011-06-01
In this paper, we present efficient algorithms for generating hierarchical molecular skin meshes with decreasing size and guaranteed quality. Our algorithms generate a sequence of coarse meshes for both the surfaces and the bounded volumes. Each coarser surface mesh is adaptive to the surface curvature and maintains the topology of the skin surface with guaranteed mesh quality. The corresponding tetrahedral mesh is conforming to the interface surface mesh and contains high quality tetrahedral that decompose both the interior of the molecule and the surrounding region (enclosed in a sphere). Our hierarchical tetrahedral meshes have a number of advantages that will facilitate fast and accurate multigrid PDE solvers. Firstly, the quality of both the surface triangulations and tetrahedral meshes is guaranteed. Secondly, the interface in the tetrahedral mesh is an accurate approximation of the molecular boundary. In particular, all the boundary points lie on the skin surface. Thirdly, our meshes are Delaunay meshes. Finally, the meshes are adaptive to the geometry. PMID:21779137
Adaptive Skin Meshes Coarsening for Biomolecular Simulation
Shi, Xinwei; Koehl, Patrice
2011-01-01
In this paper, we present efficient algorithms for generating hierarchical molecular skin meshes with decreasing size and guaranteed quality. Our algorithms generate a sequence of coarse meshes for both the surfaces and the bounded volumes. Each coarser surface mesh is adaptive to the surface curvature and maintains the topology of the skin surface with guaranteed mesh quality. The corresponding tetrahedral mesh is conforming to the interface surface mesh and contains high quality tetrahedral that decompose both the interior of the molecule and the surrounding region (enclosed in a sphere). Our hierarchical tetrahedral meshes have a number of advantages that will facilitate fast and accurate multigrid PDE solvers. Firstly, the quality of both the surface triangulations and tetrahedral meshes is guaranteed. Secondly, the interface in the tetrahedral mesh is an accurate approximation of the molecular boundary. In particular, all the boundary points lie on the skin surface. Thirdly, our meshes are Delaunay meshes. Finally, the meshes are adaptive to the geometry. PMID:21779137
NASA Astrophysics Data System (ADS)
Rastigejev, Y.; Semakin, A. N.
2012-12-01
In this work we present a multilevel Wavelet-based Adaptive Mesh Refinement (WAMR) method for numerical modeling of global atmospheric chemical transport problems. An accurate numerical simulation of such problems presents an enormous challenge. Atmospheric Chemical Transport Models (CTMs) combine chemical reactions with meteorologically predicted atmospheric advection and turbulent mixing. The resulting system of multi-scale advection-reaction-diffusion equations is extremely stiff, nonlinear and involves a large number of chemically interacting species. As a consequence, the need for enormous computational resources for solving these equations imposes severe limitations on the spatial resolution of the CTMs implemented on uniform or quasi-uniform grids. In turn, this relatively crude spatial resolution results in significant numerical diffusion introduced into the system. This numerical diffusion is shown to noticeably distort the pollutant mixing and transport dynamics for typically used grid resolutions. The developed WAMR method for numerical modeling of atmospheric chemical evolution equations presented in this work provides a significant reduction in the computational cost, without upsetting numerical accuracy, therefore it addresses the numerical difficulties described above. WAMR method introduces a fine grid in the regions where sharp transitions occur and cruder grid in the regions of smooth solution behavior. Therefore WAMR results in much more accurate solutions than conventional numerical methods implemented on uniform or quasi-uniform grids. The algorithm allows one to provide error estimates of the solution that are used in conjunction with appropriate threshold criteria to adapt the non-uniform grid. The method has been tested for a variety of problems including numerical simulation of traveling pollution plumes. It was shown that pollution plumes in the remote troposphere can propagate as well-defined layered structures for two weeks or more as
Anderson, R W; Pember, R B; Elliott, N S
2001-10-22
A new method that combines staggered grid Arbitrary Lagrangian-Eulerian (ALE) techniques with structured local adaptive mesh refinement (AMR) has been developed for solution of the Euler equations. This method facilitates the solution of problems currently at and beyond the boundary of soluble problems by traditional ALE methods by focusing computational resources where they are required through dynamic adaption. Many of the core issues involved in the development of the combined ALEAMR method hinge upon the integration of AMR with a staggered grid Lagrangian integration method. The novel components of the method are mainly driven by the need to reconcile traditional AMR techniques, which are typically employed on stationary meshes with cell-centered quantities, with the staggered grids and grid motion employed by Lagrangian methods. Numerical examples are presented which demonstrate the accuracy and efficiency of the method.
NASA Astrophysics Data System (ADS)
Rastigejev, Y.; Semakin, A. N.
2013-12-01
Accurate numerical simulations of global scale three-dimensional atmospheric chemical transport models (CTMs) are essential for studies of many important atmospheric chemistry problems such as adverse effect of air pollutants on human health, ecosystems and the Earth's climate. These simulations usually require large CPU time due to numerical difficulties associated with a wide range of spatial and temporal scales, nonlinearity and large number of reacting species. In our previous work we have shown that in order to achieve adequate convergence rate and accuracy, the mesh spacing in numerical simulation of global synoptic-scale pollution plume transport must be decreased to a few kilometers. This resolution is difficult to achieve for global CTMs on uniform or quasi-uniform grids. To address the described above difficulty we developed a three-dimensional Wavelet-based Adaptive Mesh Refinement (WAMR) algorithm. The method employs a highly non-uniform adaptive grid with fine resolution over the areas of interest without requiring small grid-spacing throughout the entire domain. The method uses multi-grid iterative solver that naturally takes advantage of a multilevel structure of the adaptive grid. In order to represent the multilevel adaptive grid efficiently, a dynamic data structure based on indirect memory addressing has been developed. The data structure allows rapid access to individual points, fast inter-grid operations and re-gridding. The WAMR method has been implemented on parallel computer architectures. The parallel algorithm is based on run-time partitioning and load-balancing scheme for the adaptive grid. The partitioning scheme maintains locality to reduce communications between computing nodes. The parallel scheme was found to be cost-effective. Specifically we obtained an order of magnitude increase in computational speed for numerical simulations performed on a twelve-core single processor workstation. We have applied the WAMR method for numerical
Operator splitting and adaptive mesh refinement for the Luo-Rudy I model
NASA Astrophysics Data System (ADS)
Trangenstein, John A.; Kim, Chisup
2004-05-01
We apply second-order operator splitting to the Luo-Rudy I model for electrical wave propagation in the heart. The purpose of the operator splitting is to separate the nonlinear but local reaction computations from the linear but globally coupled diffusion computations. This approach allows us to use local nonlinear iterations for the stiff nonlinear reactions and to solve global linear systems for the implicit treatment of diffusion. For computational efficiency, we use dynamically adaptive mesh refinement (AMR), involving hierarchies of unions of grid patches on distinct levels of refinement. The linear system for the discretization of the diffusion on the composite AMR grid is formulated via standard conforming finite elements on unions grid patches within a level of refinement and aligned mortar elements along interfaces between levels of refinement. The linear systems are solved iteratively by preconditioned conjugate gradients. Our preconditioner uses multiplicative domain decomposition between levels of refinement; the smoother involves algebraic additive domain decomposition between patches within a level of refinement, and Gauss-Seidel iteration within grid patches. Numerical results are presented in 1D and 2D, including spiral waves.
NASA Technical Reports Server (NTRS)
Savage, M.; Caldwell, R. J.; Wisor, G. D.; Lewicki, D. G.
1986-01-01
A computer model has been constructed to simulate the compliance and load sharing in a spur gear mesh. The model adds the effect of rim deflections to previously developed state-of-the-art gear tooth deflection models. The effects of deflections on mesh compliance and load sharing are examined. The model can treat gear meshes composed to two external gears or an external gear driving an internal gear. The model includes deflection contributions from the bending and shear in the teeth, the Hertzian contact deformations, and primary and secondary rotations of the gear rims. The model shows that rimmed gears increase mesh compliance and, in some cases, improve load sharing.
NASA Technical Reports Server (NTRS)
Savage, M.; Caldwell, R. J.; Wisor, G. D.; Lewicki, D. G.
1987-01-01
A computer model has been constructed to simulate the compliance and load sharing in a spur gear mesh. The model adds the effect of rim deflections to previously developed state-of-the-art gear tooth deflection models. The effects of deflections on mesh compliance and load sharing are examined. The model can treat gear meshes composed of two external gears or an external gear driving an internal gear. The model includes deflection contributions from the bending and shear in the teeth, the Hertzian contact deformations, and primary and secondary rotations of the gear rims. The model shows that rimmed gears increase mesh compliance and, in some cases, improve load sharing.
Projections of grounding line retreat in West Antarctica carried out with an adaptive mesh model
NASA Astrophysics Data System (ADS)
Cornford, Stephen; Payne, Antony; Martin, Daniel; Le Brocq, Anne
2013-04-01
Present and future sea level rise associated with mass loss from West Antarctica is typically attributed to marine glaciers retreating in response to a warming ocean. Warmer waters melt the floating ice shelves that restrain some, if not all, marine glaciers, and the glaciers themselves respond by speeding up. That leads to thinning and in turn grounding line retreat. Satellite observations indicate that Amundsen Sea Embayment and, in particular, Pine Island Glacier, are undergoing this kind of dynamic change today. Numerical models, however, struggle to reproduce the observed behavior because either high resolution or some other kind special treatment is required at the grounding line. We present 200-year projections of three major glacier systems of West Antarctica: those that drain into the Amundsen Sea , the Filchner-Ronne Ice Shelf and the Ross Ice shelf. We do so using the newly developed BISICLES ice sheet model, which employs adaptive mesh refinement to maintain sub-kilometer resolution close to the grounding line and coarser resolution elsewhere. Ice accumulation and ice shelf melt-rate are derived from a range of models of the Antarctic atmosphere and ocean forced by the SRES A1B and E1 scenarios. We find that a substantial proportion of the grounding line in West Antarctica retreats, however the total sea level rise is less than 50 mm by 2100, and less than 100 mm by 2200. The lion's share of the mass loss is attributed to Pine Island Glacier, while its immediate neighbor Thwaites Glacier does not retreat until the end of the simulations.
NASA Astrophysics Data System (ADS)
Zheng, J.; Zhu, J.; Wang, Z.; Fang, F.; Pain, C. C.; Xiang, J.
2015-10-01
An integrated method of advanced anisotropic hr-adaptive mesh and discretization numerical techniques has been, for first time, applied to modelling of multiscale advection-diffusion problems, which is based on a discontinuous Galerkin/control volume discretization on unstructured meshes. Over existing air quality models typically based on static-structured grids using a locally nesting technique, the advantage of the anisotropic hr-adaptive model has the ability to adapt the mesh according to the evolving pollutant distribution and flow features. That is, the mesh resolution can be adjusted dynamically to simulate the pollutant transport process accurately and effectively. To illustrate the capability of the anisotropic adaptive unstructured mesh model, three benchmark numerical experiments have been set up for two-dimensional (2-D) advection phenomena. Comparisons have been made between the results obtained using uniform resolution meshes and anisotropic adaptive resolution meshes. Performance achieved in 3-D simulation of power plant plumes indicates that this new adaptive multiscale model has the potential to provide accurate air quality modelling solutions effectively.
Adaptive mesh refinement in titanium
Colella, Phillip; Wen, Tong
2005-01-21
In this paper, we evaluate Titanium's usability as a high-level parallel programming language through a case study, where we implement a subset of Chombo's functionality in Titanium. Chombo is a software package applying the Adaptive Mesh Refinement methodology to numerical Partial Differential Equations at the production level. In Chombo, the library approach is used to parallel programming (C++ and Fortran, with MPI), whereas Titanium is a Java dialect designed for high-performance scientific computing. The performance of our implementation is studied and compared with that of Chombo in solving Poisson's equation based on two grid configurations from a real application. Also provided are the counts of lines of code from both sides.
Auto-adaptive finite element meshes
NASA Technical Reports Server (NTRS)
Richter, Roland; Leyland, Penelope
1995-01-01
Accurate capturing of discontinuities within compressible flow computations is achieved by coupling a suitable solver with an automatic adaptive mesh algorithm for unstructured triangular meshes. The mesh adaptation procedures developed rely on non-hierarchical dynamical local refinement/derefinement techniques, which hence enable structural optimization as well as geometrical optimization. The methods described are applied for a number of the ICASE test cases are particularly interesting for unsteady flow simulations.
Fully implicit adaptive mesh refinement MHD algorithm
NASA Astrophysics Data System (ADS)
Philip, Bobby
2005-10-01
In the macroscopic simulation of plasmas, the numerical modeler is faced with the challenge of dealing with multiple time and length scales. The former results in stiffness due to the presence of very fast waves. The latter requires one to resolve the localized features that the system develops. Traditional approaches based on explicit time integration techniques and fixed meshes are not suitable for this challenge, as such approaches prevent the modeler from using realistic plasma parameters to keep the computation feasible. We propose here a novel approach, based on implicit methods and structured adaptive mesh refinement (SAMR). Our emphasis is on both accuracy and scalability with the number of degrees of freedom. To our knowledge, a scalable, fully implicit AMR algorithm has not been accomplished before for MHD. As a proof-of-principle, we focus on the reduced resistive MHD model as a basic MHD model paradigm, which is truly multiscale. The approach taken here is to adapt mature physics-based technologyootnotetextL. Chac'on et al., J. Comput. Phys. 178 (1), 15- 36 (2002) to AMR grids, and employ AMR-aware multilevel techniques (such as fast adaptive composite --FAC-- algorithms) for scalability. We will demonstrate that the concept is indeed feasible, featuring optimal scalability under grid refinement. Results of fully-implicit, dynamically-adaptive AMR simulations will be presented on a variety of problems.
Efficient triangular adaptive meshes for tsunami simulations
NASA Astrophysics Data System (ADS)
Behrens, J.
2012-04-01
With improving technology and increased sensor density for accurate determination of tsunamogenic earthquake source parameters and consecutively uplift distribution, real-time simulations of even near-field tsunami hazard appears feasible in the near future. In order to support such efforts a new generation of tsunami models is currently under development. These models comprise adaptively refined meshes, in order to save computational resources (in areas of low wave activity) and still represent the inherently multi-scale behavior of a tsunami approaching coastal waters. So far, these methods have been based on oct-tree quadrilateral refinement. The method introduced here is based on binary tree refinement on triangular grids. By utilizing the structure stemming from the refinement strategy, a very efficient method can be achieved, with a triangular mesh, able to accurately represent complex boundaries.
Floating shock fitting via Lagrangian adaptive meshes
NASA Technical Reports Server (NTRS)
Vanrosendale, John
1995-01-01
In recent work we have formulated a new approach to compressible flow simulation, combining the advantages of shock-fitting and shock-capturing. Using a cell-centered on Roe scheme discretization on unstructured meshes, we warp the mesh while marching to steady state, so that mesh edges align with shocks and other discontinuities. This new algorithm, the Shock-fitting Lagrangian Adaptive Method (SLAM), is, in effect, a reliable shock-capturing algorithm which yields shock-fitted accuracy at convergence.
Anisotropic Mesh Adaptivity for Turbulent Flows with Boundary Layers
NASA Astrophysics Data System (ADS)
Chitale, Kedar C.
Turbulent flows are found everywhere in nature and are studied, analyzed and simulated using various experimental and numerical tools. For computational analysis, a variety of turbulence models are available and the accuracy of these models in capturing the phenomenon depends largely on the mesh spacings, especially near the walls, in the boundary layer region. Special semi-structured meshes called "mesh boundary layers" are widely used in the CFD community in simulations of turbulent flows, because of their graded and orthogonal layered structure. They provide an efficient way to achieve very fine and highly anisotropic mesh spacings without introducing poorly shaped elements. Since usually the required mesh spacings to accurately resolve the flow are not known a priori to the simulations, an adaptive approach based on a posteriori error indicators is used to achieve an appropriate mesh. In this study, we apply the adaptive meshing techniques to turbulent flows with a focus on boundary layers. We construct a framework to calculate the critical wall normal mesh spacings inside the boundary layers based on the flow physics and the knowledge of the turbulence model. This approach is combined with numerical error indicators to adapt the entire flow region. We illustrate the effectiveness of this hybrid approach by applying it to three aerodynamic flows and studying their superior performance in capturing the flow structures in detail. We also demonstrate the capabilities of the current developments in parallel boundary layer mesh adaptation by applying them to two internal flow problems. We also study the application of adaptive boundary layer meshes to complex geometries like multi element wings. We highlight the advantage of using such techniques for superior wake and tip region resolution by showcasing flow results. We also outline the future direction for the adaptive meshing techniques to be useful to the large scale flow computations.
Arbitrary Lagrangian Eulerian Adaptive Mesh Refinement
Koniges, A.; Eder, D.; Masters, N.; Fisher, A.; Anderson, R.; Gunney, B.; Wang, P.; Benson, D.; Dixit, P.
2009-09-29
This is a simulation code involving an ALE (arbitrary Lagrangian-Eulerian) hydrocode with AMR (adaptive mesh refinement) and pluggable physics packages for material strength, heat conduction, radiation diffusion, and laser ray tracing developed a LLNL, UCSD, and Berkeley Lab. The code is an extension of the open source SAMRAI (Structured Adaptive Mesh Refinement Application Interface) code/library. The code can be used in laser facilities such as the National Ignition Facility. The code is alsi being applied to slurry flow (landslides).
NASA Astrophysics Data System (ADS)
Kimura, Satoshi; Candy, Adam S.; Holland, Paul R.; Piggott, Matthew D.; Jenkins, Adrian
2013-07-01
Several different classes of ocean model are capable of representing floating glacial ice shelves. We describe the incorporation of ice shelves into Fluidity-ICOM, a nonhydrostatic finite-element ocean model with the capacity to utilize meshes that are unstructured and adaptive in three dimensions. This geometric flexibility offers several advantages over previous approaches. The model represents melting and freezing on all ice-shelf surfaces including vertical faces, treats the ice shelf topography as continuous rather than stepped, and does not require any smoothing of the ice topography or any of the additional parameterisations of the ocean mixed layer used in isopycnal or z-coordinate models. The model can also represent a water column that decreases to zero thickness at the 'grounding line', where the floating ice shelf is joined to its tributary ice streams. The model is applied to idealised ice-shelf geometries in order to demonstrate these capabilities. In these simple experiments, arbitrarily coarsening the mesh outside the ice-shelf cavity has little effect on the ice-shelf melt rate, while the mesh resolution within the cavity is found to be highly influential. Smoothing the vertical ice front results in faster flow along the smoothed ice front, allowing greater exchange with the ocean than in simulations with a realistic ice front. A vanishing water-column thickness at the grounding line has little effect in the simulations studied. We also investigate the response of ice shelf basal melting to variations in deep water temperature in the presence of salt stratification.
Anisotropic adaptive mesh generation in two dimensions for CFD
Borouchaki, H.; Castro-Diaz, M.J.; George, P.L.; Hecht, F.; Mohammadi, B.
1996-12-31
This paper describes the extension of the classical Delaunay method in the case where anisotropic meshes are required such as in CFD when the modelized physic is strongly directional. The way in which such a mesh generation method can be incorporated in an adaptative loop of CFD as well as the case of multicriterium adaptation are discussed. Several concrete application examples are provided to illustrate the capabilities of the proposed method.
Turbulent flow calculations using unstructured and adaptive meshes
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1990-01-01
A method of efficiently computing turbulent compressible flow over complex two dimensional configurations is presented. The method makes use of fully unstructured meshes throughout the entire flow-field, thus enabling the treatment of arbitrarily complex geometries and the use of adaptive meshing techniques throughout both viscous and inviscid regions of flow-field. Mesh generation is based on a locally mapped Delaunay technique in order to generate unstructured meshes with highly-stretched elements in the viscous regions. The flow equations are discretized using a finite element Navier-Stokes solver, and rapid convergence to steady-state is achieved using an unstructured multigrid algorithm. Turbulence modeling is performed using an inexpensive algebraic model, implemented for use on unstructured and adaptive meshes. Compressible turbulent flow solutions about multiple-element airfoil geometries are computed and compared with experimental data.
PARAMESH: A Parallel Adaptive Mesh Refinement Community Toolkit
NASA Technical Reports Server (NTRS)
MacNeice, Peter; Olson, Kevin M.; Mobarry, Clark; deFainchtein, Rosalinda; Packer, Charles
1999-01-01
In this paper, we describe a community toolkit which is designed to provide parallel support with adaptive mesh capability for a large and important class of computational models, those using structured, logically cartesian meshes. The package of Fortran 90 subroutines, called PARAMESH, is designed to provide an application developer with an easy route to extend an existing serial code which uses a logically cartesian structured mesh into a parallel code with adaptive mesh refinement. Alternatively, in its simplest use, and with minimal effort, it can operate as a domain decomposition tool for users who want to parallelize their serial codes, but who do not wish to use adaptivity. The package can provide them with an incremental evolutionary path for their code, converting it first to uniformly refined parallel code, and then later if they so desire, adding adaptivity.
Gravitational Collapse With Distributed Adaptive Mesh Refinement
NASA Astrophysics Data System (ADS)
Liebling, Steven; Lehner, Luis; Motl, Patrick; Neilsen, David; Rahman, Tanvir; Reula, Oscar
2006-04-01
Gravitational collapse is studied using distributed adaptive mesh refinement (AMR). The AMR infrastructure includes a novel treatment of adaptive boundaries which allows for high orders of accuracy. Results of the collapse of Brill waves to black holes are presented. Combining both vertex centered and cell centered fields in the same evolution is discussed.
A fast approach for accurate content-adaptive mesh generation.
Yang, Yongyi; Wernick, Miles N; Brankov, Jovan G
2003-01-01
Mesh modeling is an important problem with many applications in image processing. A key issue in mesh modeling is how to generate a mesh structure that well represents an image by adapting to its content. We propose a new approach to mesh generation, which is based on a theoretical result derived on the error bound of a mesh representation. In the proposed method, the classical Floyd-Steinberg error-diffusion algorithm is employed to place mesh nodes in the image domain so that their spatial density varies according to the local image content. Delaunay triangulation is next applied to connect the mesh nodes. The result of this approach is that fine mesh elements are placed automatically in regions of the image containing high-frequency features while coarse mesh elements are used to represent smooth areas. The proposed algorithm is noniterative, fast, and easy to implement. Numerical results demonstrate that, at very low computational cost, the proposed approach can produce mesh representations that are more accurate than those produced by several existing methods. Moreover, it is demonstrated that the proposed algorithm performs well with images of various kinds, even in the presence of noise. PMID:18237961
Grid adaptation using chimera composite overlapping meshes
NASA Technical Reports Server (NTRS)
Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen
1994-01-01
The objective of this paper is to perform grid adaptation using composite overlapping meshes in regions of large gradient to accurately capture the salient features during computation. The chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using trilinear interpolation. Application to the Euler equations for shock reflections and to shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well-resolved.
Grid adaptation using Chimera composite overlapping meshes
NASA Technical Reports Server (NTRS)
Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen
1993-01-01
The objective of this paper is to perform grid adaptation using composite over-lapping meshes in regions of large gradient to capture the salient features accurately during computation. The Chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using tri-linear interpolation. Applications to the Euler equations for shock reflections and to a shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well resolved.
Grid adaption using Chimera composite overlapping meshes
NASA Technical Reports Server (NTRS)
Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen
1993-01-01
The objective of this paper is to perform grid adaptation using composite over-lapping meshes in regions of large gradient to capture the salient features accurately during computation. The Chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using tri-linear interpolation. Applications to the Euler equations for shock reflections and to a shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well resolved.
NASA Astrophysics Data System (ADS)
Guo, Zhikui; Chen, Chao; Tao, Chunhui
2016-04-01
Since 2007, there are four China Da yang cruises (CDCs), which have been carried out to investigate polymetallic sulfides in the southwest Indian ridge (SWIR) and have acquired both gravity data and bathymetry data on the corresponding survey lines(Tao et al., 2014). Sandwell et al. (2014) published a new global marine gravity model including the free air gravity data and its first order vertical gradient (Vzz). Gravity data and its gradient can be used to extract unknown density structure information(e.g. crust thickness) under surface of the earth, but they contain all the mass effect under the observation point. Therefore, how to get accurate gravity and its gradient effect of the existing density structure (e.g. terrain) has been a key issue. Using the bathymetry data or ETOPO1 (http://www.ngdc.noaa.gov/mgg/global/global.html) model at a full resolution to calculate the terrain effect could spend too much computation time. We expect to develop an effective method that takes less time but can still yield the desired accuracy. In this study, a constant-density polyhedral model is used to calculate the gravity field and its vertical gradient, which is based on the work of Tsoulis (2012). According to gravity field attenuation with distance and variance of bathymetry, we present an adaptive mesh refinement and coarsening strategies to merge both global topography data and multi-beam bathymetry data. The local coarsening or size of mesh depends on user-defined accuracy and terrain variation (Davis et al., 2011). To depict terrain better, triangular surface element and rectangular surface element are used in fine and coarse mesh respectively. This strategy can also be applied to spherical coordinate in large region and global scale. Finally, we applied this method to calculate Bouguer gravity anomaly (BGA), mantle Bouguer anomaly(MBA) and their vertical gradient in SWIR. Further, we compared the result with previous results in the literature. Both synthetic model
Arbitrary Lagrangian Eulerian Adaptive Mesh Refinement
2009-09-29
This is a simulation code involving an ALE (arbitrary Lagrangian-Eulerian) hydrocode with AMR (adaptive mesh refinement) and pluggable physics packages for material strength, heat conduction, radiation diffusion, and laser ray tracing developed a LLNL, UCSD, and Berkeley Lab. The code is an extension of the open source SAMRAI (Structured Adaptive Mesh Refinement Application Interface) code/library. The code can be used in laser facilities such as the National Ignition Facility. The code is alsi being appliedmore » to slurry flow (landslides).« less
Vertical Scan (V-SCAN) for 3-D Grid Adaptive Mesh Refinement for an atmospheric Model Dynamical Core
NASA Astrophysics Data System (ADS)
Andronova, N. G.; Vandenberg, D.; Oehmke, R.; Stout, Q. F.; Penner, J. E.
2009-12-01
One of the major building blocks of a rigorous representation of cloud evolution in global atmospheric models is a parallel adaptive grid MPI-based communication library (an Adaptive Blocks for Locally Cartesian Topologies library -- ABLCarT), which manages the block-structured data layout, handles ghost cell updates among neighboring blocks and splits a block as refinements occur. The library has several modules that provide a layer of abstraction for adaptive refinement: blocks, which contain individual cells of user data; shells - the global geometry for the problem, including a sphere, reduced sphere, and now a 3D sphere; a load balancer for placement of blocks onto processors; and a communication support layer which encapsulates all data movement. A major performance concern with adaptive mesh refinement is how to represent calculations that have need to be sequenced in a particular order in a direction, such as calculating integrals along a specific path (e.g. atmospheric pressure or geopotential in the vertical dimension). This concern is compounded if the blocks have varying levels of refinement, or are scattered across different processors, as can be the case in parallel computing. In this paper we describe an implementation in ABLCarT of a vertical scan operation, which allows computing along vertical paths in the correct order across blocks transparent to their resolution and processor location. We test this functionality on a 2D and a 3D advection problem, which tests the performance of the model’s dynamics (transport) and physics (sources and sinks) for different model resolutions needed for inclusion of cloud formation.
Progress in integrated analysis with adaptive unstructured meshing
NASA Technical Reports Server (NTRS)
Dechaumphai, Pramote
1992-01-01
Design of lightweight structures and thermal protection systems for hypersonic vehicles depend on accurate prediction of aerothermal loads, structural temperatures and their gradients, and structural deformations and stresses. Concentration is on an alternative meshing technique which generates an entirely new adaptive unstructured mesh based on the solution obtained from the earlier mesh. The technique combined with the finite element method has been shown to significantly improve the efficiency and accuracy of the fluid, thermal, and structural analyses. Current capability of the adaptive unstructured meshing technique for the integrated fluid-thermal-structural analysis is described first. The technique was extended to transient thermal analysis of structures with time-dependent adaptive meshing to capture the detailed temperature response with a minimum number of unknowns and computational cost. Both linear and higher-order finite elements are implemented to demonstrate the generality of the technique and to investigate their solution accuracy. Currently, the adaptive meshing technique is being developed for plane structures that can be modeled with membrane elements and built-up structures modeled with membrane and bending elements. The capability of the technique to these different disciplinary problems is demonstrated by several examples.
Parallel object-oriented adaptive mesh refinement
Balsara, D.; Quinlan, D.J.
1997-04-01
In this paper we study adaptive mesh refinement (AMR) for elliptic and hyperbolic systems. We use the Asynchronous Fast Adaptive Composite Grid Method (AFACX), a parallel algorithm based upon the of Fast Adaptive Composite Grid Method (FAC) as a test case of an adaptive elliptic solver. For our hyperbolic system example we use TVD and ENO schemes for solving the Euler and MHD equations. We use the structured grid load balancer MLB as a tool for obtaining a load balanced distribution in a parallel environment. Parallel adaptive mesh refinement poses difficulties in expressing both the basic single grid solver, whether elliptic or hyperbolic, in a fashion that parallelizes seamlessly. It also requires that these basic solvers work together within the adaptive mesh refinement algorithm which uses the single grid solvers as one part of its adaptive solution process. We show that use of AMR++, an object-oriented library within the OVERTURE Framework, simplifies the development of AMR applications. Parallel support is provided and abstracted through the use of the P++ parallel array class.
Adaption of unstructured meshes using node movement
Carpenter, J.G.; McRae, V.D.S.
1996-12-31
The adaption algorithm of Benson and McRae is modified for application to unstructured grids. The weight function generation was modified for application to unstructured grids and movement was limited to prevent cross over. A NACA 0012 airfoil is used as a test case to evaluate the modified algorithm when applied to unstructured grids and compared to results obtained by Warren. An adaptive mesh solution for the Sudhoo and Hall four element airfoil is included as a demonstration case.
A hierarchical structure for automatic meshing and adaptive FEM analysis
NASA Technical Reports Server (NTRS)
Kela, Ajay; Saxena, Mukul; Perucchio, Renato
1987-01-01
A new algorithm for generating automatically, from solid models of mechanical parts, finite element meshes that are organized as spatially addressable quaternary trees (for 2-D work) or octal trees (for 3-D work) is discussed. Because such meshes are inherently hierarchical as well as spatially addressable, they permit efficient substructuring techniques to be used for both global analysis and incremental remeshing and reanalysis. The global and incremental techniques are summarized and some results from an experimental closed loop 2-D system in which meshing, analysis, error evaluation, and remeshing and reanalysis are done automatically and adaptively are presented. The implementation of 3-D work is briefly discussed.
Multigrid solution strategies for adaptive meshing problems
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1995-01-01
This paper discusses the issues which arise when combining multigrid strategies with adaptive meshing techniques for solving steady-state problems on unstructured meshes. A basic strategy is described, and demonstrated by solving several inviscid and viscous flow cases. Potential inefficiencies in this basic strategy are exposed, and various alternate approaches are discussed, some of which are demonstrated with an example. Although each particular approach exhibits certain advantages, all methods have particular drawbacks, and the formulation of a completely optimal strategy is considered to be an open problem.
Unstructured Adaptive Meshes: Bad for Your Memory?
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Feng, Hui-Yu; VanderWijngaart, Rob
2003-01-01
This viewgraph presentation explores the need for a NASA Advanced Supercomputing (NAS) parallel benchmark for problems with irregular dynamical memory access. This benchmark is important and necessary because: 1) Problems with localized error source benefit from adaptive nonuniform meshes; 2) Certain machines perform poorly on such problems; 3) Parallel implementation may provide further performance improvement but is difficult. Some examples of problems which use irregular dynamical memory access include: 1) Heat transfer problem; 2) Heat source term; 3) Spectral element method; 4) Base functions; 5) Elemental discrete equations; 6) Global discrete equations. Nonconforming Mesh and Mortar Element Method are covered in greater detail in this presentation.
GRChombo: Numerical relativity with adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Clough, Katy; Figueras, Pau; Finkel, Hal; Kunesch, Markus; Lim, Eugene A.; Tunyasuvunakool, Saran
2015-12-01
In this work, we introduce {\\mathtt{GRChombo}}: a new numerical relativity code which incorporates full adaptive mesh refinement (AMR) using block structured Berger-Rigoutsos grid generation. The code supports non-trivial 'many-boxes-in-many-boxes' mesh hierarchies and massive parallelism through the message passing interface. {\\mathtt{GRChombo}} evolves the Einstein equation using the standard BSSN formalism, with an option to turn on CCZ4 constraint damping if required. The AMR capability permits the study of a range of new physics which has previously been computationally infeasible in a full 3 + 1 setting, while also significantly simplifying the process of setting up the mesh for these problems. We show that {\\mathtt{GRChombo}} can stably and accurately evolve standard spacetimes such as binary black hole mergers and scalar collapses into black holes, demonstrate the performance characteristics of our code, and discuss various physics problems which stand to benefit from the AMR technique.
Floating shock fitting via Lagrangian adaptive meshes
NASA Technical Reports Server (NTRS)
Vanrosendale, John
1994-01-01
In recent works we have formulated a new approach to compressible flow simulation, combining the advantages of shock-fitting and shock-capturing. Using a cell-centered Roe scheme discretization on unstructured meshes, we warp the mesh while marching to steady state, so that mesh edges align with shocks and other discontinuities. This new algorithm, the Shock-fitting Lagrangian Adaptive Method (SLAM) is, in effect, a reliable shock-capturing algorithm which yields shock-fitted accuracy at convergence. Shock-capturing algorithms like this, which warp the mesh to yield shock-fitted accuracy, are new and relatively untried. However, their potential is clear. In the context of sonic booms, accurate calculation of near-field sonic boom signatures is critical to the design of the High Speed Civil Transport (HSCT). SLAM should allow computation of accurate N-wave pressure signatures on comparatively coarse meshes, significantly enhancing our ability to design low-boom configurations for high-speed aircraft.
Details of tetrahedral anisotropic mesh adaptation
NASA Astrophysics Data System (ADS)
Jensen, Kristian Ejlebjerg; Gorman, Gerard
2016-04-01
We have implemented tetrahedral anisotropic mesh adaptation using the local operations of coarsening, swapping, refinement and smoothing in MATLAB without the use of any for- N loops, i.e. the script is fully vectorised. In the process of doing so, we have made three observations related to details of the implementation: 1. restricting refinement to a single edge split per element not only simplifies the code, it also improves mesh quality, 2. face to edge swapping is unnecessary, and 3. optimising for the Vassilevski functional tends to give a little higher value for the mean condition number functional than optimising for the condition number functional directly. These observations have been made for a uniform and a radial shock metric field, both starting from a structured mesh in a cube. Finally, we compare two coarsening techniques and demonstrate the importance of applying smoothing in the mesh adaptation loop. The results pertain to a unit cube geometry, but we also show the effect of corners and edges by applying the implementation in a spherical geometry.
Electrostatic PIC with adaptive Cartesian mesh
NASA Astrophysics Data System (ADS)
Kolobov, Vladimir; Arslanbekov, Robert
2016-05-01
We describe an initial implementation of an electrostatic Particle-in-Cell (ES-PIC) module with adaptive Cartesian mesh in our Unified Flow Solver framework. Challenges of PIC method with cell-based adaptive mesh refinement (AMR) are related to a decrease of the particle-per-cell number in the refined cells with a corresponding increase of the numerical noise. The developed ES-PIC solver is validated for capacitively coupled plasma, its AMR capabilities are demonstrated for simulations of streamer development during high-pressure gas breakdown. It is shown that cell-based AMR provides a convenient particle management algorithm for exponential multiplications of electrons and ions in the ionization events.
Dynamic mesh adaption for triangular and tetrahedral grids
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Strawn, Roger
1993-01-01
The following topics are discussed: requirements for dynamic mesh adaption; linked-list data structure; edge-based data structure; adaptive-grid data structure; three types of element subdivision; mesh refinement; mesh coarsening; additional constraints for coarsening; anisotropic error indicator for edges; unstructured-grid Euler solver; inviscid 3-D wing; and mesh quality for solution-adaptive grids. The discussion is presented in viewgraph form.
PARAMESH V4.1: Parallel Adaptive Mesh Refinement
NASA Astrophysics Data System (ADS)
MacNeice, Peter; Olson, Kevin M.; Mobarry, Clark; de Fainchtein, Rosalinda; Packer, Charles
2011-06-01
PARAMESH is a package of Fortran 90 subroutines designed to provide an application developer with an easy route to extend an existing serial code which uses a logically cartesian structured mesh into a parallel code with adaptive mesh refinement (AMR). Alternatively, in its simplest use, and with minimal effort, it can operate as a domain decomposition tool for users who want to parallelize their serial codes, but who do not wish to use adaptivity. The package builds a hierarchy of sub-grids to cover the computational domain, with spatial resolution varying to satisfy the demands of the application. These sub-grid blocks form the nodes of a tree data-structure (quad-tree in 2D or oct-tree in 3D). Each grid block has a logically cartesian mesh. The package supports 1, 2 and 3D models. PARAMESH is released under the NASA-wide Open-Source software license.
PLUM: Parallel Load Balancing for Adaptive Unstructured Meshes
NASA Technical Reports Server (NTRS)
Oliker, Leonid; Biswas, Rupak; Saini, Subhash (Technical Monitor)
1998-01-01
Mesh adaption is a powerful tool for efficient unstructured-grid computations but causes load imbalance among processors on a parallel machine. We present a novel method called PLUM to dynamically balance the processor workloads with a global view. This paper presents the implementation and integration of all major components within our dynamic load balancing strategy for adaptive grid calculations. Mesh adaption, repartitioning, processor assignment, and remapping are critical components of the framework that must be accomplished rapidly and efficiently so as not to cause a significant overhead to the numerical simulation. A data redistribution model is also presented that predicts the remapping cost on the SP2. This model is required to determine whether the gain from a balanced workload distribution offsets the cost of data movement. Results presented in this paper demonstrate that PLUM is an effective dynamic load balancing strategy which remains viable on a large number of processors.
Octree based automatic meshing from CSG models
NASA Technical Reports Server (NTRS)
Perucchio, Renato
1987-01-01
Finite element meshes derived automatically from solid models through recursive spatial subdivision schemes (octrees) can be made to inherit the hierarchical structure and the spatial addressability intrinsic to the underlying grid. These two properties, together with the geometric regularity that can also be built into the mesh, make octree based meshes ideally suited for efficient analysis and self-adaptive remeshing and reanalysis. The element decomposition of the octal cells that intersect the boundary of the domain is emphasized. The problem, central to octree based meshing, is solved by combining template mapping and element extraction into a procedure that utilizes both constructive solid geometry and boundary respresentation techniques. Boundary cells that are not intersected by the edge of the domain boundary are easily mapped to predefined element topology. Cells containing edges (and vertices) are first transformed into a planar polyhedron and then triangulated via element extractors. The modeling environments required for the derivation of planar polyhedra and for element extraction are analyzed.
Structured Adaptive Mesh Refinement Application Infrastructure
2010-07-15
SAMRAI is an object-oriented support library for structured adaptice mesh refinement (SAMR) simulation of computational science problems, modeled by systems of partial differential equations (PDEs). SAMRAI is developed and maintained in the Center for Applied Scientific Computing (CASC) under ASCI ITS and PSE support. SAMRAI is used in a variety of application research efforts at LLNL and in academia. These applications are developed in collaboration with SAMRAI development team members.
Adaptive Meshing Techniques for Viscous Flow Calculations on Mixed Element Unstructured Meshes
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1997-01-01
An adaptive refinement strategy based on hierarchical element subdivision is formulated and implemented for meshes containing arbitrary mixtures of tetrahendra, hexahendra, prisms and pyramids. Special attention is given to keeping memory overheads as low as possible. This procedure is coupled with an algebraic multigrid flow solver which operates on mixed-element meshes. Inviscid flows as well as viscous flows are computed an adaptively refined tetrahedral, hexahedral, and hybrid meshes. The efficiency of the method is demonstrated by generating an adapted hexahedral mesh containing 3 million vertices on a relatively inexpensive workstation.
An adaptive mesh-moving and refinement procedure for one-dimensional conservation laws
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Flaherty, Joseph E.; Arney, David C.
1993-01-01
We examine the performance of an adaptive mesh-moving and /or local mesh refinement procedure for the finite difference solution of one-dimensional hyperbolic systems of conservation laws. Adaptive motion of a base mesh is designed to isolate spatially distinct phenomena, and recursive local refinement of the time step and cells of the stationary or moving base mesh is performed in regions where a refinement indicator exceeds a prescribed tolerance. These adaptive procedures are incorporated into a computer code that includes a MacCormack finite difference scheme wih Davis' artificial viscosity model and a discretization error estimate based on Richardson's extrapolation. Experiments are conducted on three problems in order to qualify the advantages of adaptive techniques relative to uniform mesh computations and the relative benefits of mesh moving and refinement. Key results indicate that local mesh refinement, with and without mesh moving, can provide reliable solutions at much lower computational cost than possible on uniform meshes; that mesh motion can be used to improve the results of uniform mesh solutions for a modest computational effort; that the cost of managing the tree data structure associated with refinement is small; and that a combination of mesh motion and refinement reliably produces solutions for the least cost per unit accuracy.
Adaptive Mesh Refinement Simulations of Relativistic Binaries
NASA Astrophysics Data System (ADS)
Motl, Patrick M.; Anderson, M.; Lehner, L.; Olabarrieta, I.; Tohline, J. E.; Liebling, S. L.; Rahman, T.; Hirschman, E.; Neilsen, D.
2006-09-01
We present recent results from our efforts to evolve relativistic binaries composed of compact objects. We simultaneously solve the general relativistic hydrodynamics equations to evolve the material components of the binary and Einstein's equations to evolve the space-time. These two codes are coupled through an adaptive mesh refinement driver (had). One of the ultimate goals of this project is to address the merger of a neutron star and black hole and assess the possible observational signature of such systems as gamma ray bursts. This work has been supported in part by NSF grants AST 04-07070 and PHY 03-26311 and in part through NASA's ATP program grant NAG5-13430. The computations were performed primarily at NCSA through grant MCA98N043 and at LSU's Center for Computation & Technology.
Visualization Tools for Adaptive Mesh Refinement Data
Weber, Gunther H.; Beckner, Vincent E.; Childs, Hank; Ligocki,Terry J.; Miller, Mark C.; Van Straalen, Brian; Bethel, E. Wes
2007-05-09
Adaptive Mesh Refinement (AMR) is a highly effective method for simulations that span a large range of spatiotemporal scales, such as astrophysical simulations that must accommodate ranges from interstellar to sub-planetary. Most mainstream visualization tools still lack support for AMR as a first class data type and AMR code teams use custom built applications for AMR visualization. The Department of Energy's (DOE's) Science Discovery through Advanced Computing (SciDAC) Visualization and Analytics Center for Enabling Technologies (VACET) is currently working on extending VisIt, which is an open source visualization tool that accommodates AMR as a first-class data type. These efforts will bridge the gap between general-purpose visualization applications and highly specialized AMR visual analysis applications. Here, we give an overview of the state of the art in AMR visualization research and tools and describe how VisIt currently handles AMR data.
Visualization of Scalar Adaptive Mesh Refinement Data
VACET; Weber, Gunther; Weber, Gunther H.; Beckner, Vince E.; Childs, Hank; Ligocki, Terry J.; Miller, Mark C.; Van Straalen, Brian; Bethel, E. Wes
2007-12-06
Adaptive Mesh Refinement (AMR) is a highly effective computation method for simulations that span a large range of spatiotemporal scales, such as astrophysical simulations, which must accommodate ranges from interstellar to sub-planetary. Most mainstream visualization tools still lack support for AMR grids as a first class data type and AMR code teams use custom built applications for AMR visualization. The Department of Energy's (DOE's) Science Discovery through Advanced Computing (SciDAC) Visualization and Analytics Center for Enabling Technologies (VACET) is currently working on extending VisIt, which is an open source visualization tool that accommodates AMR as a first-class data type. These efforts will bridge the gap between general-purpose visualization applications and highly specialized AMR visual analysis applications. Here, we give an overview of the state of the art in AMR scalar data visualization research.
Elliptic Solvers for Adaptive Mesh Refinement Grids
Quinlan, D.J.; Dendy, J.E., Jr.; Shapira, Y.
1999-06-03
We are developing multigrid methods that will efficiently solve elliptic problems with anisotropic and discontinuous coefficients on adaptive grids. The final product will be a library that provides for the simplified solution of such problems. This library will directly benefit the efforts of other Laboratory groups. The focus of this work is research on serial and parallel elliptic algorithms and the inclusion of our black-box multigrid techniques into this new setting. The approach applies the Los Alamos object-oriented class libraries that greatly simplify the development of serial and parallel adaptive mesh refinement applications. In the final year of this LDRD, we focused on putting the software together; in particular we completed the final AMR++ library, we wrote tutorials and manuals, and we built example applications. We implemented the Fast Adaptive Composite Grid method as the principal elliptic solver. We presented results at the Overset Grid Conference and other more AMR specific conferences. We worked on optimization of serial and parallel performance and published several papers on the details of this work. Performance remains an important issue and is the subject of continuing research work.
Sampling and surface reconstruction with adaptive-size meshes
NASA Astrophysics Data System (ADS)
Huang, Wen-Chen; Goldgof, Dmitry B.
1992-03-01
This paper presents a new approach to sampling and surface reconstruction which uses the physically based models. We introduce adaptive-size meshes which automatically update the size of the meshes as the distance between the nodes changes. We have implemented the adaptive-size algorithm to the following three applications: (1) Sampling of the intensity data. (2) Surface reconstruction of the range data. (3) Surface reconstruction of the 3-D computed tomography left ventricle data. The LV data was acquired by the 3-D computed tomography (CT) scanner. It was provided by Dr. Eric Hoffman at University of Pennsylvania Medical school and consists of 16 volumetric (128 X 128 X 118) images taken through the heart cycle.
A parallel adaptive mesh refinement algorithm
NASA Technical Reports Server (NTRS)
Quirk, James J.; Hanebutte, Ulf R.
1993-01-01
Over recent years, Adaptive Mesh Refinement (AMR) algorithms which dynamically match the local resolution of the computational grid to the numerical solution being sought have emerged as powerful tools for solving problems that contain disparate length and time scales. In particular, several workers have demonstrated the effectiveness of employing an adaptive, block-structured hierarchical grid system for simulations of complex shock wave phenomena. Unfortunately, from the parallel algorithm developer's viewpoint, this class of scheme is quite involved; these schemes cannot be distilled down to a small kernel upon which various parallelizing strategies may be tested. However, because of their block-structured nature such schemes are inherently parallel, so all is not lost. In this paper we describe the method by which Quirk's AMR algorithm has been parallelized. This method is built upon just a few simple message passing routines and so it may be implemented across a broad class of MIMD machines. Moreover, the method of parallelization is such that the original serial code is left virtually intact, and so we are left with just a single product to support. The importance of this fact should not be underestimated given the size and complexity of the original algorithm.
Zhang, S.; Yuen, D.A.; Zhu, A.; Song, S.; George, D.L.
2011-01-01
We parallelized the GeoClaw code on one-level grid using OpenMP in March, 2011 to meet the urgent need of simulating tsunami waves at near-shore from Tohoku 2011 and achieved over 75% of the potential speed-up on an eight core Dell Precision T7500 workstation [1]. After submitting that work to SC11 - the International Conference for High Performance Computing, we obtained an unreleased OpenMP version of GeoClaw from David George, who developed the GeoClaw code as part of his PH.D thesis. In this paper, we will show the complementary characteristics of the two approaches used in parallelizing GeoClaw and the speed-up obtained by combining the advantage of each of the two individual approaches with adaptive mesh refinement (AMR), demonstrating the capabilities of running GeoClaw efficiently on many-core systems. We will also show a novel simulation of the Tohoku 2011 Tsunami waves inundating the Sendai airport and Fukushima Nuclear Power Plants, over which the finest grid distance of 20 meters is achieved through a 4-level AMR. This simulation yields quite good predictions about the wave-heights and travel time of the tsunami waves. ?? 2011 IEEE.
Fully implicit adaptive mesh refinement algorithm for reduced MHD
NASA Astrophysics Data System (ADS)
Philip, Bobby; Pernice, Michael; Chacon, Luis
2006-10-01
In the macroscopic simulation of plasmas, the numerical modeler is faced with the challenge of dealing with multiple time and length scales. Traditional approaches based on explicit time integration techniques and fixed meshes are not suitable for this challenge, as such approaches prevent the modeler from using realistic plasma parameters to keep the computation feasible. We propose here a novel approach, based on implicit methods and structured adaptive mesh refinement (SAMR). Our emphasis is on both accuracy and scalability with the number of degrees of freedom. As a proof-of-principle, we focus on the reduced resistive MHD model as a basic MHD model paradigm, which is truly multiscale. The approach taken here is to adapt mature physics-based technology to AMR grids, and employ AMR-aware multilevel techniques (such as fast adaptive composite grid --FAC-- algorithms) for scalability. We demonstrate that the concept is indeed feasible, featuring near-optimal scalability under grid refinement. Results of fully-implicit, dynamically-adaptive AMR simulations in challenging dissipation regimes will be presented on a variety of problems that benefit from this capability, including tearing modes, the island coalescence instability, and the tilt mode instability. L. Chac'on et al., J. Comput. Phys. 178 (1), 15- 36 (2002) B. Philip, M. Pernice, and L. Chac'on, Lecture Notes in Computational Science and Engineering, accepted (2006)
Parallel adaptive mesh refinement within the PUMAA3D Project
NASA Technical Reports Server (NTRS)
Freitag, Lori; Jones, Mark; Plassmann, Paul
1995-01-01
To enable the solution of large-scale applications on distributed memory architectures, we are designing and implementing parallel algorithms for the fundamental tasks of unstructured mesh computation. In this paper, we discuss efficient algorithms developed for two of these tasks: parallel adaptive mesh refinement and mesh partitioning. The algorithms are discussed in the context of two-dimensional finite element solution on triangular meshes, but are suitable for use with a variety of element types and with h- or p-refinement. Results demonstrating the scalability and efficiency of the refinement algorithm and the quality of the mesh partitioning are presented for several test problems on the Intel DELTA.
Elliptic Solvers with Adaptive Mesh Refinement on Complex Geometries
Phillip, B.
2000-07-24
Adaptive Mesh Refinement (AMR) is a numerical technique for locally tailoring the resolution computational grids. Multilevel algorithms for solving elliptic problems on adaptive grids include the Fast Adaptive Composite grid method (FAC) and its parallel variants (AFAC and AFACx). Theory that confirms the independence of the convergence rates of FAC and AFAC on the number of refinement levels exists under certain ellipticity and approximation property conditions. Similar theory needs to be developed for AFACx. The effectiveness of multigrid-based elliptic solvers such as FAC, AFAC, and AFACx on adaptively refined overlapping grids is not clearly understood. Finally, a non-trivial eye model problem will be solved by combining the power of using overlapping grids for complex moving geometries, AMR, and multilevel elliptic solvers.
Lin, Paul Tinphone; Jameson, Antony, 1934-; Baker, Timothy J.; Martinelli, Luigi
2005-01-01
An implicit multigrid-driven algorithm for two-dimensional incompressible laminar viscous flows has been coupled with a solution adaptation method and a mesh movement method for boundary movement. Time-dependent calculations are performed implicitly by regarding each time step as a steady-state problem in pseudo-time. The method of artificial compressibility is used to solve the flow equations. The solution mesh adaptation method performs local mesh refinement using an incremental Delaunay algorithm and mesh coarsening by means of edge collapse. Mesh movement is achieved by modeling the computational domain as an elastic solid and solving the equilibrium equations for the stress field. The solution adaptation method has been validated by comparison with experimental results and other computational results for low Reynolds number flow over a shedding circular cylinder. Preliminary validation of the mesh movement method has been demonstrated by a comparison with experimental results of an oscillating airfoil and with computational results for an oscillating cylinder.
Adaptive mesh fluid simulations on GPU
NASA Astrophysics Data System (ADS)
Wang, Peng; Abel, Tom; Kaehler, Ralf
2010-10-01
We describe an implementation of compressible inviscid fluid solvers with block-structured adaptive mesh refinement on Graphics Processing Units using NVIDIA's CUDA. We show that a class of high resolution shock capturing schemes can be mapped naturally on this architecture. Using the method of lines approach with the second order total variation diminishing Runge-Kutta time integration scheme, piecewise linear reconstruction, and a Harten-Lax-van Leer Riemann solver, we achieve an overall speedup of approximately 10 times faster execution on one graphics card as compared to a single core on the host computer. We attain this speedup in uniform grid runs as well as in problems with deep AMR hierarchies. Our framework can readily be applied to more general systems of conservation laws and extended to higher order shock capturing schemes. This is shown directly by an implementation of a magneto-hydrodynamic solver and comparing its performance to the pure hydrodynamic case. Finally, we also combined our CUDA parallel scheme with MPI to make the code run on GPU clusters. Close to ideal speedup is observed on up to four GPUs.
Toward parallel, adaptive mesh refinement for chemically reacting flow simulations
Devine, K.D.; Shadid, J.N.; Salinger, A.G. Hutchinson, S.A.; Hennigan, G.L.
1997-12-01
Adaptive numerical methods offer greater efficiency than traditional numerical methods by concentrating computational effort in regions of the problem domain where the solution is difficult to obtain. In this paper, the authors describe progress toward adding mesh refinement to MPSalsa, a computer program developed at Sandia National laboratories to solve coupled three-dimensional fluid flow and detailed reaction chemistry systems for modeling chemically reacting flow on large-scale parallel computers. Data structures that support refinement and dynamic load-balancing are discussed. Results using uniform refinement with mesh sequencing to improve convergence to steady-state solutions are also presented. Three examples are presented: a lid driven cavity, a thermal convection flow, and a tilted chemical vapor deposition reactor.
Adaptive mesh refinement for stochastic reaction-diffusion processes
Bayati, Basil; Chatelain, Philippe; Koumoutsakos, Petros
2011-01-01
We present an algorithm for adaptive mesh refinement applied to mesoscopic stochastic simulations of spatially evolving reaction-diffusion processes. The transition rates for the diffusion process are derived on adaptive, locally refined structured meshes. Convergence of the diffusion process is presented and the fluctuations of the stochastic process are verified. Furthermore, a refinement criterion is proposed for the evolution of the adaptive mesh. The method is validated in simulations of reaction-diffusion processes as described by the Fisher-Kolmogorov and Gray-Scott equations.
Adaptive and Quality Quadrilateral/Hexahedral Meshing from Volumetric Data⋆
Zhang, Yongjie; Bajaj, Chandrajit
2009-01-01
This paper describes an algorithm to extract adaptive and quality quadrilateral/hexahedral meshes directly from volumetric data. First, a bottom-up surface topology preserving octree-based algorithm is applied to select a starting octree level. Then the dual contouring method is used to extract a preliminary uniform quad/hex mesh, which is decomposed into finer quads/hexes adaptively without introducing any hanging nodes. The positions of all boundary vertices are recalculated to approximate the boundary surface more accurately. Mesh adaptivity can be controlled by a feature sensitive error function, the regions that users are interested in, or finite element calculation results. Finally, a relaxation based technique is deployed to improve mesh quality. Several demonstration examples are provided from a wide variety of application domains. Some extracted meshes have been extensively used in finite element simulations. PMID:19750180
Adaptive-mesh algorithms for computational fluid dynamics
NASA Technical Reports Server (NTRS)
Powell, Kenneth G.; Roe, Philip L.; Quirk, James
1993-01-01
The basic goal of adaptive-mesh algorithms is to distribute computational resources wisely by increasing the resolution of 'important' regions of the flow and decreasing the resolution of regions that are less important. While this goal is one that is worthwhile, implementing schemes that have this degree of sophistication remains more of an art than a science. In this paper, the basic pieces of adaptive-mesh algorithms are described and some of the possible ways to implement them are discussed and compared. These basic pieces are the data structure to be used, the generation of an initial mesh, the criterion to be used to adapt the mesh to the solution, and the flow-solver algorithm on the resulting mesh. Each of these is discussed, with particular emphasis on methods suitable for the computation of compressible flows.
Serial and parallel dynamic adaptation of general hybrid meshes
NASA Astrophysics Data System (ADS)
Kavouklis, Christos
The Navier-Stokes equations are a standard mathematical representation of viscous fluid flow. Their numerical solution in three dimensions remains a computationally intensive and challenging task, despite recent advances in computer speed and memory. A strategy to increase accuracy of Navier-Stokes simulations, while maintaining computing resources to a minimum, is local refinement of the associated computational mesh in regions of large solution gradients and coarsening in regions where the solution does not vary appreciably. In this work we consider adaptation of general hybrid meshes for Computational Fluid Dynamics (CFD) applications. Hybrid meshes are composed of four types of elements; hexahedra, prisms, pyramids and tetrahedra, and have been proven a promising technology in accurately resolving fluid flow for complex geometries. The first part of this dissertation is concerned with the design and implementation of a serial scheme for the adaptation of general three dimensional hybrid meshes. We have defined 29 refinement types, for all four kinds of elements. The core of the present adaptation scheme is an iterative algorithm that flags mesh edges for refinement, so that the adapted mesh is conformal. Of primary importance is considered the design of a suitable dynamic data structure that facilitates refinement and coarsening operations and furthermore minimizes memory requirements. A special dynamic list is defined for mesh elements, in contrast with the usual tree structures. It contains only elements of the current adaptation step and minimal information that is utilized to reconstruct parent elements when the mesh is coarsened. In the second part of this work, a new parallel dynamic mesh adaptation and load balancing algorithm for general hybrid meshes is presented. Partitioning of a hybrid mesh reduces to partitioning of the corresponding dual graph. Communication among processors is based on the faces of the interpartition boundary. The distributed
Adaptive mesh and algorithm refinement using direct simulation Monte Carlo
Garcia, A.L.; Bell, J.B.; Crutchfield, W.Y.; Alder, B.J.
1999-09-01
Adaptive mesh and algorithm refinement (AMAR) embeds a particle method within a continuum method at the finest level of an adaptive mesh refinement (AMR) hierarchy. The coupling between the particle region and the overlaying continuum grid is algorithmically equivalent to that between the fine and coarse levels of AMR. Direct simulation Monte Carlo (DSMC) is used as the particle algorithm embedded within a Godunov-type compressible Navier-Stokes solver. Several examples are presented and compared with purely continuum calculations.
Parallel adaptive mesh refinement for electronic structure calculations
Kohn, S.; Weare, J.; Ong, E.; Baden, S.
1996-12-01
We have applied structured adaptive mesh refinement techniques to the solution of the LDA equations for electronic structure calculations. Local spatial refinement concentrates memory resources and numerical effort where it is most needed, near the atomic centers and in regions of rapidly varying charge density. The structured grid representation enables us to employ efficient iterative solver techniques such as conjugate gradients with multigrid preconditioning. We have parallelized our solver using an object-oriented adaptive mesh refinement framework.
Friedman, A.; Miller, W. Jr.
1993-12-31
The program was divided into segments: (Week 1) geometric modeling and mesh generation (Weeks 2 and 3) error estimation and adaptive strategies. Participants in the program came from a wide variety of disciplines dealing with remarkably analogous problems in this area. Ideas were exchanged and interdisciplinary collaboration was initiated in informal contexts as well as in the talks and question periods. In the talks, a number of algorithms were described along with specific applications to problems of great current interest in various scientific and engineering disciplines. In this emerging field, participants developed criteria for evaluation of algorithms and established guidelines for selection of algorithms appropriate to any specific problem. Special features of a problem may include curved surfaces, complicated boundaries, evolving interfaces (such as occur in coating flows), and/or criticality of error estimation.
3D Finite Element Trajectory Code with Adaptive Meshing
NASA Astrophysics Data System (ADS)
Ives, Lawrence; Bui, Thuc; Vogler, William; Bauer, Andy; Shephard, Mark; Beal, Mark; Tran, Hien
2004-11-01
Beam Optics Analysis, a new, 3D charged particle program is available and in use for the design of complex, 3D electron guns and charged particle devices. The code reads files directly from most CAD and solid modeling programs, includes an intuitive Graphical User Interface (GUI), and a robust mesh generator that is fully automatic. Complex problems can be set up, and analysis initiated in minutes. The program includes a user-friendly post processor for displaying field and trajectory data using 3D plots and images. The electrostatic solver is based on the standard nodal finite element method. The magnetostatic field solver is based on the vector finite element method and is also called during the trajectory simulation process to solve for self magnetic fields. The user imports the geometry from essentially any commercial CAD program and uses the GUI to assign parameters (voltages, currents, dielectric constant) and designate emitters (including work function, emitter temperature, and number of trajectories). The the mesh is generated automatically and analysis is performed, including mesh adaptation to improve accuracy and optimize computational resources. This presentation will provide information on the basic structure of the code, its operation, and it's capabilities.
NASA Astrophysics Data System (ADS)
Lee, W. H.; Kim, T.-S.; Cho, M. H.; Ahn, Y. B.; Lee, S. Y.
2006-12-01
In studying bioelectromagnetic problems, finite element analysis (FEA) offers several advantages over conventional methods such as the boundary element method. It allows truly volumetric analysis and incorporation of material properties such as anisotropic conductivity. For FEA, mesh generation is the first critical requirement and there exist many different approaches. However, conventional approaches offered by commercial packages and various algorithms do not generate content-adaptive meshes (cMeshes), resulting in numerous nodes and elements in modelling the conducting domain, and thereby increasing computational load and demand. In this work, we present efficient content-adaptive mesh generation schemes for complex biological volumes of MR images. The presented methodology is fully automatic and generates FE meshes that are adaptive to the geometrical contents of MR images, allowing optimal representation of conducting domain for FEA. We have also evaluated the effect of cMeshes on FEA in three dimensions by comparing the forward solutions from various cMesh head models to the solutions from the reference FE head model in which fine and equidistant FEs constitute the model. The results show that there is a significant gain in computation time with minor loss in numerical accuracy. We believe that cMeshes should be useful in the FEA of bioelectromagnetic problems.
Simulation of nonpoint source contamination based on adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Kourakos, G.; Harter, T.
2014-12-01
Contamination of groundwater aquifers from nonpoint sources is a worldwide problem. Typical agricultural groundwater basins receive contamination from a large array (in the order of ~10^5-6) of spatially and temporally heterogeneous sources such as fields, crops, dairies etc, while the received contaminants emerge at significantly uncertain time lags to a large array of discharge surfaces such as public supply, domestic and irrigation wells and streams. To support decision making in such complex regimes several approaches have been developed, which can be grouped into 3 categories: i) Index methods, ii)regression methods and iii) physically based methods. Among the three, physically based methods are considered more accurate, but at the cost of computational demand. In this work we present a physically based simulation framework which exploits the latest hardware and software developments to simulate large (>>1,000 km2) groundwater basins. First we simulate groundwater flow using a sufficiently detailed mesh to capture the spatial heterogeneity. To achieve optimal mesh quality we combine adaptive mesh refinement with the nonlinear solution for unconfined flow. Starting from a coarse grid the mesh is refined iteratively in the parts of the domain where the flow heterogeneity appears higher resulting in optimal grid. Secondly we simulate the nonpoint source pollution based on the detailed velocity field computed from the previous step. In our approach we use the streamline model where the 3D transport problem is decomposed into multiple 1D transport problems. The proposed framework is applied to simulate nonpoint source pollution in the Central Valley aquifer system, California.
PLUM: Parallel Load Balancing for Unstructured Adaptive Meshes
NASA Technical Reports Server (NTRS)
Oliker, Leonid
1998-01-01
Dynamic mesh adaption on unstructured grids is a powerful tool for computing large-scale problems that require grid modifications to efficiently resolve solution features. Unfortunately, an efficient parallel implementation is difficult to achieve, primarily due to the load imbalance created by the dynamically-changing nonuniform grid. To address this problem, we have developed PLUM, an automatic portable framework for performing adaptive large-scale numerical computations in a message-passing environment. First, we present an efficient parallel implementation of a tetrahedral mesh adaption scheme. Extremely promising parallel performance is achieved for various refinement and coarsening strategies on a realistic-sized domain. Next we describe PLUM, a novel method for dynamically balancing the processor workloads in adaptive grid computations. This research includes interfacing the parallel mesh adaption procedure based on actual flow solutions to a data remapping module, and incorporating an efficient parallel mesh repartitioner. A significant runtime improvement is achieved by observing that data movement for a refinement step should be performed after the edge-marking phase but before the actual subdivision. We also present optimal and heuristic remapping cost metrics that can accurately predict the total overhead for data redistribution. Several experiments are performed to verify the effectiveness of PLUM on sequences of dynamically adapted unstructured grids. Portability is demonstrated by presenting results on the two vastly different architectures of the SP2 and the Origin2OOO. Additionally, we evaluate the performance of five state-of-the-art partitioning algorithms that can be used within PLUM. It is shown that for certain classes of unsteady adaption, globally repartitioning the computational mesh produces higher quality results than diffusive repartitioning schemes. We also demonstrate that a coarse starting mesh produces high quality load balancing, at
Adaptive upscaling with the dual mesh method
Guerillot, D.; Verdiere, S.
1997-08-01
The objective of this paper is to demonstrate that upscaling should be calculated during the flow simulation instead of trying to enhance the a priori upscaling methods. Hence, counter-examples are given to motivate our approach, the so-called Dual Mesh Method. The main steps of this numerical algorithm are recalled. Applications illustrate the necessity to consider different average relative permeability values depending on the direction in space. Moreover, these values could be different for the same average saturation. This proves that an a priori upscaling cannot be the answer even in homogeneous cases because of the {open_quotes}dynamical heterogeneity{close_quotes} created by the saturation profile. Other examples show the efficiency of the Dual Mesh Method applied to heterogeneous medium and to an actual field case in South America.
Adaptive mesh generation for viscous flows using Delaunay triangulation
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1988-01-01
A method for generating an unstructured triangular mesh in two dimensions, suitable for computing high Reynolds number flows over arbitrary configurations is presented. The method is based on a Delaunay triangulation, which is performed in a locally stretched space, in order to obtain very high aspect ratio triangles in the boundary layer and the wake regions. It is shown how the method can be coupled with an unstructured Navier-Stokes solver to produce a solution adaptive mesh generation procedure for viscous flows.
Parallel tetrahedral mesh adaptation with dynamic load balancing
Oliker, Leonid; Biswas, Rupak; Gabow, Harold N.
2000-06-28
The ability to dynamically adapt an unstructured grid is a powerful tool for efficiently solving computational problems with evolving physical features. In this paper, we report on our experience parallelizing an edge-based adaptation scheme, called 3D-TAG, using message passing. Results show excellent speedup when a realistic helicopter rotor mesh is randomly refined. However, performance deteriorates when the mesh is refined using a solution-based error indicator since mesh adaptation for practical problems occurs in a localized region, creating a severe load imbalance. To address this problem, we have developed PLUM, a global dynamic load balancing framework for adaptive numerical computations. Even though PLUM primarily balances processor workloads for the solution phase, it reduces the load imbalance problem within mesh adaptation by repartitioning the mesh after targeting edges for refinement but before the actual subdivision. This dramatically improves the performance of parallel 3D-TAG since refinement occurs in a more load balanced fashion. We also present optimal and heuristic algorithms that, when applied to the default mapping of a parallel repartitioner, significantly reduce the data redistribution overhead. Finally, portability is examined by comparing performance on three state-of-the-art parallel machines.
Parallel Tetrahedral Mesh Adaptation with Dynamic Load Balancing
NASA Technical Reports Server (NTRS)
Oliker, Leonid; Biswas, Rupak; Gabow, Harold N.
1999-01-01
The ability to dynamically adapt an unstructured grid is a powerful tool for efficiently solving computational problems with evolving physical features. In this paper, we report on our experience parallelizing an edge-based adaptation scheme, called 3D_TAG. using message passing. Results show excellent speedup when a realistic helicopter rotor mesh is randomly refined. However. performance deteriorates when the mesh is refined using a solution-based error indicator since mesh adaptation for practical problems occurs in a localized region., creating a severe load imbalance. To address this problem, we have developed PLUM, a global dynamic load balancing framework for adaptive numerical computations. Even though PLUM primarily balances processor workloads for the solution phase, it reduces the load imbalance problem within mesh adaptation by repartitioning the mesh after targeting edges for refinement but before the actual subdivision. This dramatically improves the performance of parallel 3D_TAG since refinement occurs in a more load balanced fashion. We also present optimal and heuristic algorithms that, when applied to the default mapping of a parallel repartitioner, significantly reduce the data redistribution overhead. Finally, portability is examined by comparing performance on three state-of-the-art parallel machines.
Lober, R.R.; Tautges, T.J.; Vaughan, C.T.
1997-03-01
Paving is an automated mesh generation algorithm which produces all-quadrilateral elements. It can additionally generate these elements in varying sizes such that the resulting mesh adapts to a function distribution, such as an error function. While powerful, conventional paving is a very serial algorithm in its operation. Parallel paving is the extension of serial paving into parallel environments to perform the same meshing functions as conventional paving only on distributed, discretized models. This extension allows large, adaptive, parallel finite element simulations to take advantage of paving`s meshing capabilities for h-remap remeshing. A significantly modified version of the CUBIT mesh generation code has been developed to host the parallel paving algorithm and demonstrate its capabilities on both two dimensional and three dimensional surface geometries and compare the resulting parallel produced meshes to conventionally paved meshes for mesh quality and algorithm performance. Sandia`s {open_quotes}tiling{close_quotes} dynamic load balancing code has also been extended to work with the paving algorithm to retain parallel efficiency as subdomains undergo iterative mesh refinement.
3D Compressible Melt Transport with Adaptive Mesh Refinement
NASA Astrophysics Data System (ADS)
Dannberg, Juliane; Heister, Timo
2015-04-01
Melt generation and migration have been the subject of numerous investigations, but their typical time and length-scales are vastly different from mantle convection, which makes it difficult to study these processes in a unified framework. The equations that describe coupled Stokes-Darcy flow have been derived a long time ago and they have been successfully implemented and applied in numerical models (Keller et al., 2013). However, modelling magma dynamics poses the challenge of highly non-linear and spatially variable material properties, in particular the viscosity. Applying adaptive mesh refinement to this type of problems is particularly advantageous, as the resolution can be increased in mesh cells where melt is present and viscosity gradients are high, whereas a lower resolution is sufficient in regions without melt. In addition, previous models neglect the compressibility of both the solid and the fluid phase. However, experiments have shown that the melt density change from the depth of melt generation to the surface leads to a volume increase of up to 20%. Considering these volume changes in both phases also ensures self-consistency of models that strive to link melt generation to processes in the deeper mantle, where the compressibility of the solid phase becomes more important. We describe our extension of the finite-element mantle convection code ASPECT (Kronbichler et al., 2012) that allows for solving additional equations describing the behaviour of silicate melt percolating through and interacting with a viscously deforming host rock. We use the original compressible formulation of the McKenzie equations, augmented by an equation for the conservation of energy. This approach includes both melt migration and melt generation with the accompanying latent heat effects. We evaluate the functionality and potential of this method using a series of simple model setups and benchmarks, comparing results of the compressible and incompressible formulation and
Parallel adaptation of general three-dimensional hybrid meshes
NASA Astrophysics Data System (ADS)
Kavouklis, Christos; Kallinderis, Yannis
2010-05-01
A new parallel dynamic mesh adaptation and load balancing algorithm for general hybrid grids has been developed. The meshes considered in this work are composed of four kinds of elements; tetrahedra, prisms, hexahedra and pyramids, which poses a challenge to parallel mesh adaptation. Additional complexity imposed by the presence of multiple types of elements affects especially data migration, updates of local data structures and interpartition data structures. Efficient partition of hybrid meshes has been accomplished by transforming them to suitable graphs and using serial graph partitioning algorithms. Communication among processors is based on the faces of the interpartition boundary and the termination detection algorithm of Dijkstra is employed to ensure proper flagging of edges for refinement. An inexpensive dynamic load balancing strategy is introduced to redistribute work load among processors after adaptation. In particular, only the initial coarse mesh, with proper weighting, is balanced which yields savings in computation time and relatively simple implementation of mesh quality preservation rules, while facilitating coarsening of refined elements. Special algorithms are employed for (i) data migration and dynamic updates of the local data structures, (ii) determination of the resulting interpartition boundary and (iii) identification of the communication pattern of processors. Several representative applications are included to evaluate the method.
Huang, W.; Zheng, Lingyun; Zhan, X.
2002-01-01
Accurate modelling of groundwater flow and transport with sharp moving fronts often involves high computational cost, when a fixed/uniform mesh is used. In this paper, we investigate the modelling of groundwater problems using a particular adaptive mesh method called the moving mesh partial differential equation approach. With this approach, the mesh is dynamically relocated through a partial differential equation to capture the evolving sharp fronts with a relatively small number of grid points. The mesh movement and physical system modelling are realized by solving the mesh movement and physical partial differential equations alternately. The method is applied to the modelling of a range of groundwater problems, including advection dominated chemical transport and reaction, non-linear infiltration in soil, and the coupling of density dependent flow and transport. Numerical results demonstrate that sharp moving fronts can be accurately and efficiently captured by the moving mesh approach. Also addressed are important implementation strategies, e.g. the construction of the monitor function based on the interpolation error, control of mesh concentration, and two-layer mesh movement. Copyright ?? 2002 John Wiley and Sons, Ltd.
Adaptive mesh refinement and adjoint methods in geophysics simulations
NASA Astrophysics Data System (ADS)
Burstedde, Carsten
2013-04-01
It is an ongoing challenge to increase the resolution that can be achieved by numerical geophysics simulations. This applies to considering sub-kilometer mesh spacings in global-scale mantle convection simulations as well as to using frequencies up to 1 Hz in seismic wave propagation simulations. One central issue is the numerical cost, since for three-dimensional space discretizations, possibly combined with time stepping schemes, a doubling of resolution can lead to an increase in storage requirements and run time by factors between 8 and 16. A related challenge lies in the fact that an increase in resolution also increases the dimensionality of the model space that is needed to fully parametrize the physical properties of the simulated object (a.k.a. earth). Systems that exhibit a multiscale structure in space are candidates for employing adaptive mesh refinement, which varies the resolution locally. An example that we found well suited is the mantle, where plate boundaries and fault zones require a resolution on the km scale, while deeper area can be treated with 50 or 100 km mesh spacings. This approach effectively reduces the number of computational variables by several orders of magnitude. While in this case it is possible to derive the local adaptation pattern from known physical parameters, it is often unclear what are the most suitable criteria for adaptation. We will present the goal-oriented error estimation procedure, where such criteria are derived from an objective functional that represents the observables to be computed most accurately. Even though this approach is well studied, it is rarely used in the geophysics community. A related strategy to make finer resolution manageable is to design methods that automate the inference of model parameters. Tweaking more than a handful of numbers and judging the quality of the simulation by adhoc comparisons to known facts and observations is a tedious task and fundamentally limited by the turnaround times
A structured multi-block solution-adaptive mesh algorithm with mesh quality assessment
NASA Technical Reports Server (NTRS)
Ingram, Clint L.; Laflin, Kelly R.; Mcrae, D. Scott
1995-01-01
The dynamic solution adaptive grid algorithm, DSAGA3D, is extended to automatically adapt 2-D structured multi-block grids, including adaption of the block boundaries. The extension is general, requiring only input data concerning block structure, connectivity, and boundary conditions. Imbedded grid singular points are permitted, but must be prevented from moving in space. Solutions for workshop cases 1 and 2 are obtained on multi-block grids and illustrate both increased resolution of and alignment with the solution. A mesh quality assessment criteria is proposed to determine how well a given mesh resolves and aligns with the solution obtained upon it. The criteria is used to evaluate the grid quality for solutions of workshop case 6 obtained on both static and dynamically adapted grids. The results indicate that this criteria shows promise as a means of evaluating resolution.
Multigrid solution of internal flows using unstructured solution adaptive meshes
NASA Astrophysics Data System (ADS)
Smith, Wayne A.; Blake, Kenneth R.
1992-11-01
This is the final report of the NASA Lewis SBIR Phase 2 Contract Number NAS3-25785, Multigrid Solution of Internal Flows Using Unstructured Solution Adaptive Meshes. The objective of this project, as described in the Statement of Work, is to develop and deliver to NASA a general three-dimensional Navier-Stokes code using unstructured solution-adaptive meshes for accuracy and multigrid techniques for convergence acceleration. The code will primarily be applied, but not necessarily limited, to high speed internal flows in turbomachinery.
A Spectral Adaptive Mesh Refinement Method for the Burgers equation
NASA Astrophysics Data System (ADS)
Nasr Azadani, Leila; Staples, Anne
2013-03-01
Adaptive mesh refinement (AMR) is a powerful technique in computational fluid dynamics (CFD). Many CFD problems have a wide range of scales which vary with time and space. In order to resolve all the scales numerically, high grid resolutions are required. The smaller the scales the higher the resolutions should be. However, small scales are usually formed in a small portion of the domain or in a special period of time. AMR is an efficient method to solve these types of problems, allowing high grid resolutions where and when they are needed and minimizing memory and CPU time. Here we formulate a spectral version of AMR in order to accelerate simulations of a 1D model for isotropic homogenous turbulence, the Burgers equation, as a first test of this method. Using pseudo spectral methods, we applied AMR in Fourier space. The spectral AMR (SAMR) method we present here is applied to the Burgers equation and the results are compared with the results obtained using standard solution methods performed using a fine mesh.
Kinetic solvers with adaptive mesh in phase space.
Arslanbekov, Robert R; Kolobov, Vladimir I; Frolova, Anna A
2013-12-01
An adaptive mesh in phase space (AMPS) methodology has been developed for solving multidimensional kinetic equations by the discrete velocity method. A Cartesian mesh for both configuration (r) and velocity (v) spaces is produced using a "tree of trees" (ToT) data structure. The r mesh is automatically generated around embedded boundaries, and is dynamically adapted to local solution properties. The v mesh is created on-the-fly in each r cell. Mappings between neighboring v-space trees is implemented for the advection operator in r space. We have developed algorithms for solving the full Boltzmann and linear Boltzmann equations with AMPS. Several recent innovations were used to calculate the discrete Boltzmann collision integral with dynamically adaptive v mesh: the importance sampling, multipoint projection, and variance reduction methods. We have developed an efficient algorithm for calculating the linear Boltzmann collision integral for elastic and inelastic collisions of hot light particles in a Lorentz gas. Our AMPS technique has been demonstrated for simulations of hypersonic rarefied gas flows, ion and electron kinetics in weakly ionized plasma, radiation and light-particle transport through thin films, and electron streaming in semiconductors. We have shown that AMPS allows minimizing the number of cells in phase space to reduce the computational cost and memory usage for solving challenging kinetic problems. PMID:24483578
Kinetic solvers with adaptive mesh in phase space
NASA Astrophysics Data System (ADS)
Arslanbekov, Robert R.; Kolobov, Vladimir I.; Frolova, Anna A.
2013-12-01
An adaptive mesh in phase space (AMPS) methodology has been developed for solving multidimensional kinetic equations by the discrete velocity method. A Cartesian mesh for both configuration (r) and velocity (v) spaces is produced using a “tree of trees” (ToT) data structure. The r mesh is automatically generated around embedded boundaries, and is dynamically adapted to local solution properties. The v mesh is created on-the-fly in each r cell. Mappings between neighboring v-space trees is implemented for the advection operator in r space. We have developed algorithms for solving the full Boltzmann and linear Boltzmann equations with AMPS. Several recent innovations were used to calculate the discrete Boltzmann collision integral with dynamically adaptive v mesh: the importance sampling, multipoint projection, and variance reduction methods. We have developed an efficient algorithm for calculating the linear Boltzmann collision integral for elastic and inelastic collisions of hot light particles in a Lorentz gas. Our AMPS technique has been demonstrated for simulations of hypersonic rarefied gas flows, ion and electron kinetics in weakly ionized plasma, radiation and light-particle transport through thin films, and electron streaming in semiconductors. We have shown that AMPS allows minimizing the number of cells in phase space to reduce the computational cost and memory usage for solving challenging kinetic problems.
NASA Astrophysics Data System (ADS)
Sarkis, C.; Silva, L.; Gandin, Ch-A.; Plapp, M.
2016-03-01
Dendritic growth is computed with automatic adaptation of an anisotropic and unstructured finite element mesh. The energy conservation equation is formulated for solid and liquid phases considering an interface balance that includes the Gibbs-Thomson effect. An equation for a diffuse interface is also developed by considering a phase field function with constant negative value in the liquid and constant positive value in the solid. Unknowns are the phase field function and a dimensionless temperature, as proposed by [1]. Linear finite element interpolation is used for both variables, and discretization stabilization techniques ensure convergence towards a correct non-oscillating solution. In order to perform quantitative computations of dendritic growth on a large domain, two additional numerical ingredients are necessary: automatic anisotropic unstructured adaptive meshing [2,[3] and parallel implementations [4], both made available with the numerical platform used (CimLib) based on C++ developments. Mesh adaptation is found to greatly reduce the number of degrees of freedom. Results of phase field simulations for dendritic solidification of a pure material in two and three dimensions are shown and compared with reference work [1]. Discussion on algorithm details and the CPU time will be outlined.
Adaptive anisotropic meshing for steady convection-dominated problems
Nguyen, Hoa; Gunzburger, Max; Ju, Lili; Burkardt, John
2009-01-01
Obtaining accurate solutions for convection–diffusion equations is challenging due to the presence of layers when convection dominates the diffusion. To solve this problem, we design an adaptive meshing algorithm which optimizes the alignment of anisotropic meshes with the numerical solution. Three main ingredients are used. First, the streamline upwind Petrov–Galerkin method is used to produce a stabilized solution. Second, an adapted metric tensor is computed from the approximate solution. Third, optimized anisotropic meshes are generated from the computed metric tensor by an anisotropic centroidal Voronoi tessellation algorithm. Our algorithm is tested on a variety of two-dimensional examples and the results shows that the algorithm is robust in detecting layers and efficient in avoiding non-physical oscillations in the numerical approximation.
Numerical simulation of immiscible viscous fingering using adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Adam, A.; Salinas, P.; Percival, J. R.; Pavlidis, D.; Pain, C.; Muggeridge, A. H.; Jackson, M.
2015-12-01
Displacement of one fluid by another in porous media occurs in various settings including hydrocarbon recovery, CO2 storage and water purification. When the invading fluid is of lower viscosity than the resident fluid, the displacement front is subject to a Saffman-Taylor instability and is unstable to transverse perturbations. These instabilities can grow, leading to fingering of the invading fluid. Numerical simulation of viscous fingering is challenging. The physics is controlled by a complex interplay of viscous and diffusive forces and it is necessary to ensure physical diffusion dominates numerical diffusion to obtain converged solutions. This typically requires the use of high mesh resolution and high order numerical methods. This is computationally expensive. We demonstrate here the use of a novel control volume - finite element (CVFE) method along with dynamic unstructured mesh adaptivity to simulate viscous fingering with higher accuracy and lower computational cost than conventional methods. Our CVFE method employs a discontinuous representation for both pressure and velocity, allowing the use of smaller control volumes (CVs). This yields higher resolution of the saturation field which is represented CV-wise. Moreover, dynamic mesh adaptivity allows high mesh resolution to be employed where it is required to resolve the fingers and lower resolution elsewhere. We use our results to re-examine the existing criteria that have been proposed to govern the onset of instability.Mesh adaptivity requires the mapping of data from one mesh to another. Conventional methods such as consistent interpolation do not readily generalise to discontinuous fields and are non-conservative. We further contribute a general framework for interpolation of CV fields by Galerkin projection. The method is conservative, higher order and yields improved results, particularly with higher order or discontinuous elements where existing approaches are often excessively diffusive.
Adaptive mesh refinement for 1-dimensional gas dynamics
Hedstrom, G.; Rodrigue, G.; Berger, M.; Oliger, J.
1982-01-01
We consider the solution of the one-dimensional equation of gas-dynamics. Accurate numerical solutions are difficult to obtain on a given spatial mesh because of the existence of physical regions where components of the exact solution are either discontinuous or have large gradient changes. Numerical methods treat these phenomena in a variety of ways. In this paper, the method of adaptive mesh refinement is used. A thorough description of this method for general hyperbolic systems is given elsewhere and only properties of the method pertinent to the system are elaborated.
Adaptive mesh strategies for the spectral element method
NASA Technical Reports Server (NTRS)
Mavriplis, Catherine
1992-01-01
An adaptive spectral method was developed for the efficient solution of time dependent partial differential equations. Adaptive mesh strategies that include resolution refinement and coarsening by three different methods are illustrated on solutions to the 1-D viscous Burger equation and the 2-D Navier-Stokes equations for driven flow in a cavity. Sharp gradients, singularities, and regions of poor resolution are resolved optimally as they develop in time using error estimators which indicate the choice of refinement to be used. The adaptive formulation presents significant increases in efficiency, flexibility, and general capabilities for high order spectral methods.
Parallel 3D Mortar Element Method for Adaptive Nonconforming Meshes
NASA Technical Reports Server (NTRS)
Feng, Huiyu; Mavriplis, Catherine; VanderWijngaart, Rob; Biswas, Rupak
2004-01-01
High order methods are frequently used in computational simulation for their high accuracy. An efficient way to avoid unnecessary computation in smooth regions of the solution is to use adaptive meshes which employ fine grids only in areas where they are needed. Nonconforming spectral elements allow the grid to be flexibly adjusted to satisfy the computational accuracy requirements. The method is suitable for computational simulations of unsteady problems with very disparate length scales or unsteady moving features, such as heat transfer, fluid dynamics or flame combustion. In this work, we select the Mark Element Method (MEM) to handle the non-conforming interfaces between elements. A new technique is introduced to efficiently implement MEM in 3-D nonconforming meshes. By introducing an "intermediate mortar", the proposed method decomposes the projection between 3-D elements and mortars into two steps. In each step, projection matrices derived in 2-D are used. The two-step method avoids explicitly forming/deriving large projection matrices for 3-D meshes, and also helps to simplify the implementation. This new technique can be used for both h- and p-type adaptation. This method is applied to an unsteady 3-D moving heat source problem. With our new MEM implementation, mesh adaptation is able to efficiently refine the grid near the heat source and coarsen the grid once the heat source passes. The savings in computational work resulting from the dynamic mesh adaptation is demonstrated by the reduction of the the number of elements used and CPU time spent. MEM and mesh adaptation, respectively, bring irregularity and dynamics to the computer memory access pattern. Hence, they provide a good way to gauge the performance of computer systems when running scientific applications whose memory access patterns are irregular and unpredictable. We select a 3-D moving heat source problem as the Unstructured Adaptive (UA) grid benchmark, a new component of the NAS Parallel
Parallel Block Structured Adaptive Mesh Refinement on Graphics Processing Units
Beckingsale, D. A.; Gaudin, W. P.; Hornung, R. D.; Gunney, B. T.; Gamblin, T.; Herdman, J. A.; Jarvis, S. A.
2014-11-17
Block-structured adaptive mesh refinement is a technique that can be used when solving partial differential equations to reduce the number of zones necessary to achieve the required accuracy in areas of interest. These areas (shock fronts, material interfaces, etc.) are recursively covered with finer mesh patches that are grouped into a hierarchy of refinement levels. Despite the potential for large savings in computational requirements and memory usage without a corresponding reduction in accuracy, AMR adds overhead in managing the mesh hierarchy, adding complex communication and data movement requirements to a simulation. In this paper, we describe the design and implementation of a native GPU-based AMR library, including: the classes used to manage data on a mesh patch, the routines used for transferring data between GPUs on different nodes, and the data-parallel operators developed to coarsen and refine mesh data. We validate the performance and accuracy of our implementation using three test problems and two architectures: an eight-node cluster, and over four thousand nodes of Oak Ridge National Laboratory’s Titan supercomputer. Our GPU-based AMR hydrodynamics code performs up to 4.87× faster than the CPU-based implementation, and has been scaled to over four thousand GPUs using a combination of MPI and CUDA.
AN ADAPTIVE PARTICLE-MESH GRAVITY SOLVER FOR ENZO
Passy, Jean-Claude; Bryan, Greg L.
2014-11-01
We describe and implement an adaptive particle-mesh algorithm to solve the Poisson equation for grid-based hydrodynamics codes with nested grids. The algorithm is implemented and extensively tested within the astrophysical code Enzo against the multigrid solver available by default. We find that while both algorithms show similar accuracy for smooth mass distributions, the adaptive particle-mesh algorithm is more accurate for the case of point masses, and is generally less noisy. We also demonstrate that the two-body problem can be solved accurately in a configuration with nested grids. In addition, we discuss the effect of subcycling, and demonstrate that evolving all the levels with the same timestep yields even greater precision.
Boltzmann Solver with Adaptive Mesh in Velocity Space
Kolobov, Vladimir I.; Arslanbekov, Robert R.; Frolova, Anna A.
2011-05-20
We describe the implementation of direct Boltzmann solver with Adaptive Mesh in Velocity Space (AMVS) using quad/octree data structure. The benefits of the AMVS technique are demonstrated for the charged particle transport in weakly ionized plasmas where the collision integral is linear. We also describe the implementation of AMVS for the nonlinear Boltzmann collision integral. Test computations demonstrate both advantages and deficiencies of the current method for calculations of narrow-kernel distributions.
An adaptive mesh finite volume method for the Euler equations of gas dynamics
NASA Astrophysics Data System (ADS)
Mungkasi, Sudi
2016-06-01
The Euler equations have been used to model gas dynamics for decades. They consist of mathematical equations for the conservation of mass, momentum, and energy of the gas. For a large time value, the solution may contain discontinuities, even when the initial condition is smooth. A standard finite volume numerical method is not able to give accurate solutions to the Euler equations around discontinuities. Therefore we solve the Euler equations using an adaptive mesh finite volume method. In this paper, we present a new construction of the adaptive mesh finite volume method with an efficient computation of the refinement indicator. The adaptive method takes action automatically at around places having inaccurate solutions. Inaccurate solutions are reconstructed to reduce the error by refining the mesh locally up to a certain level. On the other hand, if the solution is already accurate, then the mesh is coarsened up to another certain level to minimize computational efforts. We implement the numerical entropy production as the mesh refinement indicator. As a test problem, we take the Sod shock tube problem. Numerical results show that the adaptive method is more promising than the standard one in solving the Euler equations of gas dynamics.
Advances in Patch-Based Adaptive Mesh Refinement Scalability
Gunney, Brian T.N.; Anderson, Robert W.
2015-12-18
Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extension of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.
Advances in Patch-Based Adaptive Mesh Refinement Scalability
Gunney, Brian T.N.; Anderson, Robert W.
2015-12-18
Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extensionmore » of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.« less
Adaptive radial basis function mesh deformation using data reduction
NASA Astrophysics Data System (ADS)
Gillebaart, T.; Blom, D. S.; van Zuijlen, A. H.; Bijl, H.
2016-09-01
Radial Basis Function (RBF) mesh deformation is one of the most robust mesh deformation methods available. Using the greedy (data reduction) method in combination with an explicit boundary correction, results in an efficient method as shown in literature. However, to ensure the method remains robust, two issues are addressed: 1) how to ensure that the set of control points remains an accurate representation of the geometry in time and 2) how to use/automate the explicit boundary correction, while ensuring a high mesh quality. In this paper, we propose an adaptive RBF mesh deformation method, which ensures the set of control points always represents the geometry/displacement up to a certain (user-specified) criteria, by keeping track of the boundary error throughout the simulation and re-selecting when needed. Opposed to the unit displacement and prescribed displacement selection methods, the adaptive method is more robust, user-independent and efficient, for the cases considered. Secondly, the analysis of a single high aspect ratio cell is used to formulate an equation for the correction radius needed, depending on the characteristics of the correction function used, maximum aspect ratio, minimum first cell height and boundary error. Based on the analysis two new radial basis correction functions are derived and proposed. This proposed automated procedure is verified while varying the correction function, Reynolds number (and thus first cell height and aspect ratio) and boundary error. Finally, the parallel efficiency is studied for the two adaptive methods, unit displacement and prescribed displacement for both the CPU as well as the memory formulation with a 2D oscillating and translating airfoil with oscillating flap, a 3D flexible locally deforming tube and deforming wind turbine blade. Generally, the memory formulation requires less work (due to the large amount of work required for evaluating RBF's), but the parallel efficiency reduces due to the limited
Block-structured adaptive mesh refinement - theory, implementation and application
Deiterding, Ralf
2011-01-01
Structured adaptive mesh refinement (SAMR) techniques can enable cutting-edge simulations of problems governed by conservation laws. Focusing on the strictly hyperbolic case, these notes explain all algorithmic and mathematical details of a technically relevant implementation tailored for distributed memory computers. An overview of the background of commonly used finite volume discretizations for gas dynamics is included and typical benchmarks to quantify accuracy and performance of the dynamically adaptive code are discussed. Large-scale simulations of shock-induced realistic combustion in non-Cartesian geometry and shock-driven fluid-structure interaction with fully coupled dynamic boundary motion demonstrate the applicability of the discussed techniques for complex scenarios.
Applications of automatic mesh generation and adaptive methods in computational medicine
Schmidt, J.A.; Macleod, R.S.; Johnson, C.R.; Eason, J.C.
1995-12-31
Important problems in Computational Medicine exist that can benefit from the implementation of adaptive mesh refinement techniques. Biological systems are so inherently complex that only efficient models running on state of the art hardware can begin to simulate reality. To tackle the complex geometries associated with medical applications we present a general purpose mesh generation scheme based upon the Delaunay tessellation algorithm and an iterative point generator. In addition, automatic, two- and three-dimensional adaptive mesh refinement methods are presented that are derived from local and global estimates of the finite element error. Mesh generation and adaptive refinement techniques are utilized to obtain accurate approximations of bioelectric fields within anatomically correct models of the heart and human thorax. Specifically, we explore the simulation of cardiac defibrillation and the general forward and inverse problems in electrocardiography (ECG). Comparisons between uniform and adaptive refinement techniques are made to highlight the computational efficiency and accuracy of adaptive methods in the solution of field problems in computational medicine.
Laser ray tracing in a parallel arbitrary Lagrangian-Eulerian adaptive mesh refinement hydrocode
NASA Astrophysics Data System (ADS)
Masters, N. D.; Kaiser, T. B.; Anderson, R. W.; Eder, D. C.; Fisher, A. C.; Koniges, A. E.
2010-08-01
ALE-AMR is a new hydrocode that we are developing as a predictive modeling tool for debris and shrapnel formation in high-energy laser experiments. In this paper we present our approach to implementing laser ray tracing in ALE-AMR. We present the basic concepts of laser ray tracing and our approach to efficiently traverse the adaptive mesh hierarchy.
Numerical study of Taylor bubbles with adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Xie, Zhihua; Pavlidis, Dimitrios; Percival, James; Pain, Chris; Matar, Omar; Hasan, Abbas; Azzopardi, Barry
2014-11-01
The Taylor bubble is a single long bubble which nearly fills the entire cross section of a liquid-filled circular tube. This type of bubble flow regime often occurs in gas-liquid slug flows in many industrial applications, including oil-and-gas production, chemical and nuclear reactors, and heat exchangers. The objective of this study is to investigate the fluid dynamics of Taylor bubbles rising in a vertical pipe filled with oils of extremely high viscosity (mimicking the ``heavy oils'' found in the oil-and-gas industry). A modelling and simulation framework is presented here which can modify and adapt anisotropic unstructured meshes to better represent the underlying physics of bubble rise and reduce the computational effort without sacrificing accuracy. The numerical framework consists of a mixed control-volume and finite-element formulation, a ``volume of fluid''-type method for the interface capturing based on a compressive control volume advection method, and a force-balanced algorithm for the surface tension implementation. Numerical examples of some benchmark tests and the dynamics of Taylor bubbles are presented to show the capability of this method. EPSRC Programme Grant, MEMPHIS, EP/K0039761/1.
ROAMing terrain (Real-time Optimally Adapting Meshes)
Duchaineau, M.; Wolinsky, M.; Sigeti, D.E.; Miller, M.C.; Aldrich, C.; Mineev, M.
1997-07-01
Terrain visualization is a difficult problem for applications requiring accurate images of large datasets at high frame rates, such as flight simulation and ground-based aircraft testing using synthetic sensor stimulation. On current graphics hardware, the problem is to maintain dynamic, view-dependent triangle meshes and texture maps that produce good images at the required frame rate. We present an algorithm for constructing triangle meshes that optimizes flexible view-dependent error metrics, produces guaranteed error bounds, achieves specified triangle counts directly, and uses frame-to-frame coherence to operate at high frame rates for thousands of triangles per frame. Our method, dubbed Real-time Optimally Adapting Meshes (ROAM), uses two priority queues to drive split and merge operations that maintain continuous triangulations built from pre-processed bintree triangles. We introduce two additional performance optimizations: incremental triangle stripping and priority-computation deferral lists. ROAM execution time is proportionate to the number of triangle changes per frame, which is typically a few percent of the output mesh size, hence ROAM performance is insensitive to the resolution and extent of the input terrain. Dynamic terrain and simple vertex morphing are supported.
Anisotropic Mesh Adaptivity for FE-simulation of cardiovascular flow
NASA Astrophysics Data System (ADS)
Mueller, Jens; Sahni, Onkar; Jansen, Kenneth E.; Shephard, Mark S.; Taylor, Charles A.
2004-11-01
In this study we present an adaptive anisotropic finite element method and demonstrate how computational efficiency can be increased when applying the method to the simulation of blood flow in the cardiovascular system. We use the weak SUPG formulation for the transient 3D incompressible Navier-Stokes equations which are discretized by linear finite elements, both for the pressure and the velocity field. Given the pulsatile nature of the flow in blood vessels we have pursued adaptavity based on the average flow over a cardiac cycle. Error indicators are derived to define an anisotropic mesh metric field. Mesh modification algorithms are used to anisotropically adapt the mesh according to the desired size field. We demonstrate the efficiency of the method by first applying it to pulsatile flow in a straight cylindrical pipe and then to a pig artery with a stenosis bypassed by a graft. The efficiency of the method is measured in terms of computational savings when we compute the wall shear stresses, a quantity identified to be important to understanding arterial disease.
Multilevel Error Estimation and Adaptive h-Refinement for Cartesian Meshes with Embedded Boundaries
NASA Technical Reports Server (NTRS)
Aftosmis, M. J.; Berger, M. J.; Kwak, Dochan (Technical Monitor)
2002-01-01
This paper presents the development of a mesh adaptation module for a multilevel Cartesian solver. While the module allows mesh refinement to be driven by a variety of different refinement parameters, a central feature in its design is the incorporation of a multilevel error estimator based upon direct estimates of the local truncation error using tau-extrapolation. This error indicator exploits the fact that in regions of uniform Cartesian mesh, the spatial operator is exactly the same on the fine and coarse grids, and local truncation error estimates can be constructed by evaluating the residual on the coarse grid of the restricted solution from the fine grid. A new strategy for adaptive h-refinement is also developed to prevent errors in smooth regions of the flow from being masked by shocks and other discontinuous features. For certain classes of error histograms, this strategy is optimal for achieving equidistribution of the refinement parameters on hierarchical meshes, and therefore ensures grid converged solutions will be achieved for appropriately chosen refinement parameters. The robustness and accuracy of the adaptation module is demonstrated using both simple model problems and complex three dimensional examples using meshes with from 10(exp 6), to 10(exp 7) cells.
Laser Ray Tracing in a Parallel Arbitrary Lagrangian-Eulerian Adaptive Mesh Refinement Hydrocode
Masters, N D; Kaiser, T B; Anderson, R W; Eder, D C; Fisher, A C; Koniges, A E
2009-09-28
ALE-AMR is a new hydrocode that we are developing as a predictive modeling tool for debris and shrapnel formation in high-energy laser experiments. In this paper we present our approach to implementing laser ray-tracing in ALE-AMR. We present the equations of laser ray tracing, our approach to efficient traversal of the adaptive mesh hierarchy in which we propagate computational rays through a virtual composite mesh consisting of the finest resolution representation of the modeled space, and anticipate simulations that will be compared to experiments for code validation.
How to model wireless mesh networks topology
NASA Astrophysics Data System (ADS)
Sanni, M. L.; Hashim, A. A.; Anwar, F.; Ahmed, G. S. M.; Ali, S.
2013-12-01
The specification of network connectivity model or topology is the beginning of design and analysis in Computer Network researches. Wireless Mesh Networks is an autonomic network that is dynamically self-organised, self-configured while the mesh nodes establish automatic connectivity with the adjacent nodes in the relay network of wireless backbone routers. Researches in Wireless Mesh Networks range from node deployment to internetworking issues with sensor, Internet and cellular networks. These researches require modelling of relationships and interactions among nodes including technical characteristics of the links while satisfying the architectural requirements of the physical network. However, the existing topology generators model geographic topologies which constitute different architectures, thus may not be suitable in Wireless Mesh Networks scenarios. The existing methods of topology generation are explored, analysed and parameters for their characterisation are identified. Furthermore, an algorithm for the design of Wireless Mesh Networks topology based on square grid model is proposed in this paper. The performance of the topology generated is also evaluated. This research is particularly important in the generation of a close-to-real topology for ensuring relevance of design to the intended network and validity of results obtained in Wireless Mesh Networks researches.
NASA Astrophysics Data System (ADS)
Danilov, A. A.; Salamatova, V. Yu; Vassilevski, Yu V.
2012-12-01
Here, a workflow for high-resolution efficient numerical modeling of bioimpedance measurements is suggested that includes 3D image segmentation, adaptive mesh generation, finite-element discretization, and the analysis of simulation results. Using the adaptive unstructured tetrahedral meshes enables to decrease significantly a number of mesh elements while keeping model accuracy. The numerical results illustrate current, potential, and sensitivity field distributions for a conventional Kubicek-like scheme of bioimpedance measurements using segmented geometric model of human torso based on Visible Human Project data. The whole body VHP man computational mesh is constructed that contains 574 thousand vertices and 3.3 million tetrahedrons.
Patch-based Adaptive Mesh Refinement for Multimaterial Hydrodynamics
Lomov, I; Pember, R; Greenough, J; Liu, B
2005-10-18
We present a patch-based direct Eulerian adaptive mesh refinement (AMR) algorithm for modeling real equation-of-state, multimaterial compressible flow with strength. Our approach to AMR uses a hierarchical, structured grid approach first developed by (Berger and Oliger 1984), (Berger and Oliger 1984). The grid structure is dynamic in time and is composed of nested uniform rectangular grids of varying resolution. The integration scheme on the grid hierarchy is a recursive procedure in which the coarse grids are advanced, then the fine grids are advanced multiple steps to reach the same time, and finally the coarse and fine grids are synchronized to remove conservation errors during the separate advances. The methodology presented here is based on a single grid algorithm developed for multimaterial gas dynamics by (Colella et al. 1993), refined by(Greenough et al. 1995), and extended to the solution of solid mechanics problems with significant strength by (Lomov and Rubin 2003). The single grid algorithm uses a second-order Godunov scheme with an approximate single fluid Riemann solver and a volume-of-fluid treatment of material interfaces. The method also uses a non-conservative treatment of the deformation tensor and an acoustic approximation for shear waves in the Riemann solver. This departure from a strict application of the higher-order Godunov methodology to the equation of solid mechanics is justified due to the fact that highly nonlinear behavior of shear stresses is rare. This algorithm is implemented in two codes, Geodyn and Raptor, the latter of which is a coupled rad-hydro code. The present discussion will be solely concerned with hydrodynamics modeling. Results from a number of simulations for flows with and without strength will be presented.
An unstructured-mesh atmospheric model for nonhydrostatic dynamics: Towards optimal mesh resolution
NASA Astrophysics Data System (ADS)
Szmelter, Joanna; Zhang, Zhao; Smolarkiewicz, Piotr K.
2015-08-01
The paper advances the limited-area anelastic model (Smolarkiewicz et al. (2013) [45]) for investigation of nonhydrostatic dynamics in mesoscale atmospheric flows. New developments include the extension to a tetrahedral-based median-dual option for unstructured meshes and a static mesh adaptivity technique using an error indicator based on inherent properties of the Multidimensional Positive Definite Advection Transport Algorithm (MPDATA). The model employs semi-implicit nonoscillatory forward-in-time integrators for soundproof PDEs, built on MPDATA and a robust non-symmetric Krylov-subspace elliptic solver. Finite-volume spatial discretisation adopts an edge-based data structure. Simulations of stratified orographic flows and the associated gravity-wave phenomena in media with uniform and variable dispersive properties verify the advancement and demonstrate the potential of heterogeneous anisotropic discretisation with large variation in spatial resolution for study of complex stratified flows that can be computationally unattainable with regular grids.
Compressible Flows on Adaptive and Unstrucured Meshes with FLUIDITY
NASA Astrophysics Data System (ADS)
Nelson, R.; Piggott, M.; Wilson, C.; Kramer, S.
2011-09-01
Fluidity is an open source, general purpose, multi-phase CFD code capable of solving numerically the Navier-Stokes and accompanying field equations on arbitrary unstructured finite element meshes in one, two and three dimensions. It uses a moving finite element/control volume method which allows arbitrary movement of the mesh in time dependent problems. It has a wide range of finite element/control volume element choices including mixed formulations. Here, continuous Galerkin (CG) and control volumes (CV) solutions of the compressible Navier-Stokes (N-S) equations are presented for the stratified tests cases of the rising thermal bubble and inertia gravity waves. Results show good agreement with previously published literature and novel result presented here is the ability to dynamically adapt the mesh to increase resolution in the region of interest, thus reducing the number of degrees of freedom in the problem without decreasing the accuracy of the result. Finally, results from the case of a fully three dimensional rising thermal bubble are presented.
The geometry of r-adaptive meshes generated using optimal transport methods
NASA Astrophysics Data System (ADS)
Budd, C. J.; Russell, R. D.; Walsh, E.
2015-02-01
The principles of mesh equidistribution and alignment play a fundamental role in the design of adaptive methods, and a metric tensor and mesh metric are useful theoretical tools for understanding a method's level of mesh alignment, or anisotropy. We consider a mesh redistribution method based on the Monge-Ampère equation which combines equidistribution of a given scalar density function with optimal transport. It does not involve explicit use of a metric tensor, although such a tensor must exist for the method, and an interesting question to ask is whether or not the alignment produced by the metric gives an anisotropic mesh. For model problems with a linear feature and with a radially symmetric feature, we derive the exact form of the metric, which involves expressions for its eigenvalues and eigenvectors. The eigenvectors are shown to be orthogonal and tangential to the feature, and the ratio of the eigenvalues (corresponding to the level of anisotropy) is shown to depend, both locally and globally, on the value of the density function and the amount of curvature. We thereby demonstrate how the optimal transport method produces an anisotropic mesh along a given feature while equidistributing a suitably chosen scalar density function. Numerical results are given to verify these results and to demonstrate how the analysis is useful for problems involving more complex features, including for a non-trivial time dependant nonlinear PDE which evolves narrow and curved reaction fronts.
NASA Astrophysics Data System (ADS)
Yang, Dikun; Oldenburg, Douglas W.; Haber, Eldad
2014-03-01
Airborne electromagnetic (AEM) methods are highly efficient tools for assessing the Earth's conductivity structures in a large area at low cost. However, the configuration of AEM measurements, which typically have widely distributed transmitter-receiver pairs, makes the rigorous modelling and interpretation extremely time-consuming in 3-D. Excessive overcomputing can occur when working on a large mesh covering the entire survey area and inverting all soundings in the data set. We propose two improvements. The first is to use a locally optimized mesh for each AEM sounding for the forward modelling and calculation of sensitivity. This dedicated local mesh is small with fine cells near the sounding location and coarse cells far away in accordance with EM diffusion and the geometric decay of the signals. Once the forward problem is solved on the local meshes, the sensitivity for the inversion on the global mesh is available through quick interpolation. Using local meshes for AEM forward modelling avoids unnecessary computing on fine cells on a global mesh that are far away from the sounding location. Since local meshes are highly independent, the forward modelling can be efficiently parallelized over an array of processors. The second improvement is random and dynamic down-sampling of the soundings. Each inversion iteration only uses a random subset of the soundings, and the subset is reselected for every iteration. The number of soundings in the random subset, determined by an adaptive algorithm, is tied to the degree of model regularization. This minimizes the overcomputing caused by working with redundant soundings. Our methods are compared against conventional methods and tested with a synthetic example. We also invert a field data set that was previously considered to be too large to be practically inverted in 3-D. These examples show that our methodology can dramatically reduce the processing time of 3-D inversion to a practical level without losing resolution
Parallel Processing of Adaptive Meshes with Load Balancing
NASA Technical Reports Server (NTRS)
Das, Sajal K.; Harvey, Daniel J.; Biswas, Rupak; Biegel, Bryan (Technical Monitor)
2001-01-01
Many scientific applications involve grids that lack a uniform underlying structure. These applications are often also dynamic in nature in that the grid structure significantly changes between successive phases of execution. In parallel computing environments, mesh adaptation of unstructured grids through selective refinement/coarsening has proven to be an effective approach. However, achieving load balance while minimizing interprocessor communication and redistribution costs is a difficult problem. Traditional dynamic load balancers are mostly inadequate because they lack a global view of system loads across processors. In this paper, we propose a novel and general-purpose load balancer that utilizes symmetric broadcast networks (SBN) as the underlying communication topology, and compare its performance with a successful global load balancing environment, called PLUM, specifically created to handle adaptive unstructured applications. Our experimental results on an IBM SP2 demonstrate that the SBN-based load balancer achieves lower redistribution costs than that under PLUM by overlapping processing and data migration.
Dynamic Load Balancing for Adaptive Meshes using Symmetric Broadcast Networks
NASA Technical Reports Server (NTRS)
Das, Sajal K.; Harvey, Daniel J.; Biswas, Rupak; Saini, Subhash (Technical Monitor)
1998-01-01
Many scientific applications involve grids that lack a uniform underlying structure. These applications are often dynamic in the sense that the grid structure significantly changes between successive phases of execution. In parallel computing environments, mesh adaptation of grids through selective refinement/coarsening has proven to be an effective approach. However, achieving load balance while minimizing inter-processor communication and redistribution costs is a difficult problem. Traditional dynamic load balancers are mostly inadequate because they lack a global view across processors. In this paper, we compare a novel load balancer that utilizes symmetric broadcast networks (SBN) to a successful global load balancing environment (PLUM) created to handle adaptive unstructured applications. Our experimental results on the IBM SP2 demonstrate that performance of the proposed SBN load balancer is comparable to results achieved under PLUM.
THE GENERATION OF TETRAHEDRAL MESH MODELS FOR NEUROANATOMICAL MRI
Lederman, Carl; Joshi, Anand; Dinov, Ivo; Vese, Luminita; Toga, Arthur; Van Horn, John Darrell
2010-01-01
In this article, we describe a detailed method for automatically generating tetrahedral meshes from 3D images having multiple region labels. An adaptively sized tetrahedral mesh modeling approach is described that is capable of producing meshes conforming precisely to the voxelized regions in the image. Efficient tetrahedral construction is performed minimizing an energy function containing three terms: a smoothing term to remove the voxelization, a fidelity term to maintain continuity with the image data, and a novel elasticity term to prevent the tetrahedra from becoming flattened or inverted as the mesh deforms while allowing the voxelization to be removed entirely. The meshing algorithm is applied to structural MR image data that has been automatically segmented into 56 neuroanatomical sub-divisions as well as on two other examples. The resulting tetrahedral representation has several desirable properties such as tetrahedra with dihedral angles away from 0 and 180 degrees, smoothness, and a high resolution. Tetrahedral modeling via the approach described here has applications in modeling brain structure in normal as well as diseased brain in human and non-human data and facilitates examination of 3D object deformations resulting from neurological illness (e.g. Alzheimer’s Disease), development, and/or aging. PMID:21073968
Spatially adaptive bases in wavelet-based coding of semi-regular meshes
NASA Astrophysics Data System (ADS)
Denis, Leon; Florea, Ruxandra; Munteanu, Adrian; Schelkens, Peter
2010-05-01
In this paper we present a wavelet-based coding approach for semi-regular meshes, which spatially adapts the employed wavelet basis in the wavelet transformation of the mesh. The spatially-adaptive nature of the transform requires additional information to be stored in the bit-stream in order to allow the reconstruction of the transformed mesh at the decoder side. In order to limit this overhead, the mesh is first segmented into regions of approximately equal size. For each spatial region, a predictor is selected in a rate-distortion optimal manner by using a Lagrangian rate-distortion optimization technique. When compared against the classical wavelet transform employing the butterfly subdivision filter, experiments reveal that the proposed spatially-adaptive wavelet transform significantly decreases the energy of the wavelet coefficients for all subbands. Preliminary results show also that employing the proposed transform for the lowest-resolution subband systematically yields improved compression performance at low-to-medium bit-rates. For the Venus and Rabbit test models the compression improvements add up to 1.47 dB and 0.95 dB, respectively.
Visualization of Octree Adaptive Mesh Refinement (AMR) in Astrophysical Simulations
NASA Astrophysics Data System (ADS)
Labadens, M.; Chapon, D.; Pomaréde, D.; Teyssier, R.
2012-09-01
Computer simulations are important in current cosmological research. Those simulations run in parallel on thousands of processors, and produce huge amount of data. Adaptive mesh refinement is used to reduce the computing cost while keeping good numerical accuracy in regions of interest. RAMSES is a cosmological code developed by the Commissariat à l'énergie atomique et aux énergies alternatives (English: Atomic Energy and Alternative Energies Commission) which uses Octree adaptive mesh refinement. Compared to grid based AMR, the Octree AMR has the advantage to fit very precisely the adaptive resolution of the grid to the local problem complexity. However, this specific octree data type need some specific software to be visualized, as generic visualization tools works on Cartesian grid data type. This is why the PYMSES software has been also developed by our team. It relies on the python scripting language to ensure a modular and easy access to explore those specific data. In order to take advantage of the High Performance Computer which runs the RAMSES simulation, it also uses MPI and multiprocessing to run some parallel code. We would like to present with more details our PYMSES software with some performance benchmarks. PYMSES has currently two visualization techniques which work directly on the AMR. The first one is a splatting technique, and the second one is a custom ray tracing technique. Both have their own advantages and drawbacks. We have also compared two parallel programming techniques with the python multiprocessing library versus the use of MPI run. The load balancing strategy has to be smartly defined in order to achieve a good speed up in our computation. Results obtained with this software are illustrated in the context of a massive, 9000-processor parallel simulation of a Milky Way-like galaxy.
TRIM: A finite-volume MHD algorithm for an unstructured adaptive mesh
Schnack, D.D.; Lottati, I.; Mikic, Z.
1995-07-01
The authors describe TRIM, a MHD code which uses finite volume discretization of the MHD equations on an unstructured adaptive grid of triangles in the poloidal plane. They apply it to problems related to modeling tokamak toroidal plasmas. The toroidal direction is treated by a pseudospectral method. Care was taken to center variables appropriately on the mesh and to construct a self adjoint diffusion operator for cell centered variables.
Ray, Jaideep; Lefantzi, Sophia; Najm, Habib N.; Kennedy, Christopher A.
2006-01-01
Block-structured adaptively refined meshes (SAMR) strive for efficient resolution of partial differential equations (PDEs) solved on large computational domains by clustering mesh points only where required by large gradients. Previous work has indicated that fourth-order convergence can be achieved on such meshes by using a suitable combination of high-order discretizations, interpolations, and filters and can deliver significant computational savings over conventional second-order methods at engineering error tolerances. In this paper, we explore the interactions between the errors introduced by discretizations, interpolations and filters. We develop general expressions for high-order discretizations, interpolations, and filters, in multiple dimensions, using a Fourier approach, facilitating the high-order SAMR implementation. We derive a formulation for the necessary interpolation order for given discretization and derivative orders. We also illustrate this order relationship empirically using one and two-dimensional model problems on refined meshes. We study the observed increase in accuracy with increasing interpolation order. We also examine the empirically observed order of convergence, as the effective resolution of the mesh is increased by successively adding levels of refinement, with different orders of discretization, interpolation, or filtering.
Production-quality Tools for Adaptive Mesh RefinementVisualization
Weber, Gunther H.; Childs, Hank; Bonnell, Kathleen; Meredith,Jeremy; Miller, Mark; Whitlock, Brad; Bethel, E. Wes
2007-10-25
Adaptive Mesh Refinement (AMR) is a highly effectivesimulation method for spanning a large range of spatiotemporal scales,such as astrophysical simulations that must accommodate ranges frominterstellar to sub-planetary. Most mainstream visualization tools stilllack support for AMR as a first class data type and AMR code teams usecustom built applications for AMR visualization. The Department ofEnergy's (DOE's) Science Discovery through Advanced Computing (SciDAC)Visualization and Analytics Center for Enabling Technologies (VACET) isextending and deploying VisIt, an open source visualization tool thataccommodates AMR as a first-class data type, for use asproduction-quality, parallel-capable AMR visual data analysisinfrastructure. This effort will help science teams that use AMR-basedsimulations and who develop their own AMR visual data analysis softwareto realize cost and labor savings.
Structured adaptive mesh refinement on the connection machine
Berger, M.J. . Courant Inst. of Mathematical Sciences); Saltzman, J.S. )
1993-01-01
Adaptive mesh refinement has proven itself to be a useful tool in a large collection of applications. By refining only a small portion of the computational domain, computational savings of up to a factor of 80 in 3 dimensional calculations have been obtained on serial machines. A natural question is, can this algorithm be used on massively parallel machines and still achieve the same efficiencies We have designed a data layout scheme for mapping grid points to processors that preserves locality and minimizes global communication for the CM-200. The effect of the data layout scheme is that at the finest level nearby grid points from adjacent grids in physical space are in adjacent memory locations. Furthermore, coarse grid points are arranged in memory to be near their associated fine grid points. We show applications of the algorithm to inviscid compressible fluid flow in two space dimensions.
Unstructured and adaptive mesh generation for high Reynolds number viscous flows
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1991-01-01
A method for generating and adaptively refining a highly stretched unstructured mesh suitable for the computation of high-Reynolds-number viscous flows about arbitrary two-dimensional geometries was developed. The method is based on the Delaunay triangulation of a predetermined set of points and employs a local mapping in order to achieve the high stretching rates required in the boundary-layer and wake regions. The initial mesh-point distribution is determined in a geometry-adaptive manner which clusters points in regions of high curvature and sharp corners. Adaptive mesh refinement is achieved by adding new points in regions of large flow gradients, and locally retriangulating; thus, obviating the need for global mesh regeneration. Initial and adapted meshes about complex multi-element airfoil geometries are shown and compressible flow solutions are computed on these meshes.
Implicit adaptive mesh refinement for 2D reduced resistive magnetohydrodynamics
NASA Astrophysics Data System (ADS)
Philip, Bobby; Chacón, Luis; Pernice, Michael
2008-10-01
An implicit structured adaptive mesh refinement (SAMR) solver for 2D reduced magnetohydrodynamics (MHD) is described. The time-implicit discretization is able to step over fast normal modes, while the spatial adaptivity resolves thin, dynamically evolving features. A Jacobian-free Newton-Krylov method is used for the nonlinear solver engine. For preconditioning, we have extended the optimal "physics-based" approach developed in [L. Chacón, D.A. Knoll, J.M. Finn, An implicit, nonlinear reduced resistive MHD solver, J. Comput. Phys. 178 (2002) 15-36] (which employed multigrid solver technology in the preconditioner for scalability) to SAMR grids using the well-known Fast Adaptive Composite grid (FAC) method [S. McCormick, Multilevel Adaptive Methods for Partial Differential Equations, SIAM, Philadelphia, PA, 1989]. A grid convergence study demonstrates that the solver performance is independent of the number of grid levels and only depends on the finest resolution considered, and that it scales well with grid refinement. The study of error generation and propagation in our SAMR implementation demonstrates that high-order (cubic) interpolation during regridding, combined with a robustly damping second-order temporal scheme such as BDF2, is required to minimize impact of grid errors at coarse-fine interfaces on the overall error of the computation for this MHD application. We also demonstrate that our implementation features the desired property that the overall numerical error is dependent only on the finest resolution level considered, and not on the base-grid resolution or on the number of refinement levels present during the simulation. We demonstrate the effectiveness of the tool on several challenging problems.
Curved Mesh Correction And Adaptation Tool to Improve COMPASS Electromagnetic Analyses
Luo, X.; Shephard, M.; Lee, L.Q.; Ng, C.; Ge, L.; /SLAC
2011-11-14
SLAC performs large-scale simulations for the next-generation accelerator design using higher-order finite elements. This method requires using valid curved meshes and adaptive mesh refinement in complex 3D curved domains to achieve its fast rate of convergence. ITAPS has developed a procedure to address those mesh requirements to enable petascale electromagnetic accelerator simulations by SLAC. The results demonstrate that those correct valid curvilinear meshes can not only make the simulation more reliable but also improve computational efficiency up to 30%. This paper presents a procedure to track moving adaptive mesh refinement in curved domains. The procedure is capable of generating suitable curvilinear meshes to enable large-scale accelerator simulations. The procedure can generate valid curved meshes with substantially fewer elements to improve the computational efficiency and reliability of the COMPASS electromagnetic analyses. Future work will focus on the scalable parallelization of all steps for petascale simulations.
Lung lobe modeling and segmentation with individualized surface meshes
NASA Astrophysics Data System (ADS)
Blaffert, Thomas; Barschdorf, Hans; von Berg, Jens; Dries, Sebastian; Franz, Astrid; Klinder, Tobias; Lorenz, Cristian; Renisch, Steffen; Wiemker, Rafael
2008-03-01
An automated segmentation of lung lobes in thoracic CT images is of interest for various diagnostic purposes like the quantification of emphysema or the localization of tumors within the lung. Although the separating lung fissures are visible in modern multi-slice CT-scanners, their contrast in the CT-image often does not separate the lobes completely. This makes it impossible to build a reliable segmentation algorithm without additional information. Our approach uses general anatomical knowledge represented in a geometrical mesh model to construct a robust lobe segmentation, which even gives reasonable estimates of lobe volumes if fissures are not visible at all. The paper describes the generation of the lung model mesh including lobes by an average volume model, its adaptation to individual patient data using a special fissure feature image, and a performance evaluation over a test data set showing an average segmentation accuracy of 1 to 3 mm.
NASA Astrophysics Data System (ADS)
Lian, Y.-Y.; Hsu, K.-H.; Shao, Y.-L.; Lee, Y.-M.; Jeng, Y.-W.; Wu, J.-S.
2006-12-01
The development of a parallel three-dimensional (3-D) adaptive mesh refinement (PAMR) scheme for an unstructured tetrahedral mesh using dynamic domain decomposition on a memory-distributed machine is presented in detail. A memory-saving cell-based data structure is designed such that the resulting mesh information can be readily utilized in both node- or cell-based numerical methods. The general procedures include isotropic refinement from one parent cell into eight child cells and then followed by anisotropic refinement which effectively removes hanging nodes. A simple but effective mesh-quality control mechanism is employed to preserve the mesh quality. The resulting parallel performance of this PAMR is found to scale approximately as N for N⩽32. Two test cases, including a particle method (parallel DSMC solver for rarefied gas dynamics) and an equation-based method (parallel Poisson-Boltzmann equation solver for electrostatic field), are used to demonstrate the generality of the PAMR module. It is argued that this PAMR scheme can be applied in any numerical method if the unstructured tetrahedral mesh is adopted.
An object-oriented approach for parallel self adaptive mesh refinement on block structured grids
NASA Technical Reports Server (NTRS)
Lemke, Max; Witsch, Kristian; Quinlan, Daniel
1993-01-01
Self-adaptive mesh refinement dynamically matches the computational demands of a solver for partial differential equations to the activity in the application's domain. In this paper we present two C++ class libraries, P++ and AMR++, which significantly simplify the development of sophisticated adaptive mesh refinement codes on (massively) parallel distributed memory architectures. The development is based on our previous research in this area. The C++ class libraries provide abstractions to separate the issues of developing parallel adaptive mesh refinement applications into those of parallelism, abstracted by P++, and adaptive mesh refinement, abstracted by AMR++. P++ is a parallel array class library to permit efficient development of architecture independent codes for structured grid applications, and AMR++ provides support for self-adaptive mesh refinement on block-structured grids of rectangular non-overlapping blocks. Using these libraries, the application programmers' work is greatly simplified to primarily specifying the serial single grid application and obtaining the parallel and self-adaptive mesh refinement code with minimal effort. Initial results for simple singular perturbation problems solved by self-adaptive multilevel techniques (FAC, AFAC), being implemented on the basis of prototypes of the P++/AMR++ environment, are presented. Singular perturbation problems frequently arise in large applications, e.g. in the area of computational fluid dynamics. They usually have solutions with layers which require adaptive mesh refinement and fast basic solvers in order to be resolved efficiently.
Xia, Kelin; Zhan, Meng; Wan, Decheng; Wei, Guo-Wei
2012-02-01
Mesh deformation methods are a versatile strategy for solving partial differential equations (PDEs) with a vast variety of practical applications. However, these methods break down for elliptic PDEs with discontinuous coefficients, namely, elliptic interface problems. For this class of problems, the additional interface jump conditions are required to maintain the well-posedness of the governing equation. Consequently, in order to achieve high accuracy and high order convergence, additional numerical algorithms are required to enforce the interface jump conditions in solving elliptic interface problems. The present work introduces an interface technique based adaptively deformed mesh strategy for resolving elliptic interface problems. We take the advantages of the high accuracy, flexibility and robustness of the matched interface and boundary (MIB) method to construct an adaptively deformed mesh based interface method for elliptic equations with discontinuous coefficients. The proposed method generates deformed meshes in the physical domain and solves the transformed governed equations in the computational domain, which maintains regular Cartesian meshes. The mesh deformation is realized by a mesh transformation PDE, which controls the mesh redistribution by a source term. The source term consists of a monitor function, which builds in mesh contraction rules. Both interface geometry based deformed meshes and solution gradient based deformed meshes are constructed to reduce the L(∞) and L(2) errors in solving elliptic interface problems. The proposed adaptively deformed mesh based interface method is extensively validated by many numerical experiments. Numerical results indicate that the adaptively deformed mesh based interface method outperforms the original MIB method for dealing with elliptic interface problems. PMID:22586356
Xia, Kelin; Zhan, Meng; Wan, Decheng; Wei, Guo-Wei
2011-01-01
Mesh deformation methods are a versatile strategy for solving partial differential equations (PDEs) with a vast variety of practical applications. However, these methods break down for elliptic PDEs with discontinuous coefficients, namely, elliptic interface problems. For this class of problems, the additional interface jump conditions are required to maintain the well-posedness of the governing equation. Consequently, in order to achieve high accuracy and high order convergence, additional numerical algorithms are required to enforce the interface jump conditions in solving elliptic interface problems. The present work introduces an interface technique based adaptively deformed mesh strategy for resolving elliptic interface problems. We take the advantages of the high accuracy, flexibility and robustness of the matched interface and boundary (MIB) method to construct an adaptively deformed mesh based interface method for elliptic equations with discontinuous coefficients. The proposed method generates deformed meshes in the physical domain and solves the transformed governed equations in the computational domain, which maintains regular Cartesian meshes. The mesh deformation is realized by a mesh transformation PDE, which controls the mesh redistribution by a source term. The source term consists of a monitor function, which builds in mesh contraction rules. Both interface geometry based deformed meshes and solution gradient based deformed meshes are constructed to reduce the L∞ and L2 errors in solving elliptic interface problems. The proposed adaptively deformed mesh based interface method is extensively validated by many numerical experiments. Numerical results indicate that the adaptively deformed mesh based interface method outperforms the original MIB method for dealing with elliptic interface problems. PMID:22586356
An adaptive mesh refinement algorithm for the discrete ordinates method
Jessee, J.P.; Fiveland, W.A.; Howell, L.H.; Colella, P.; Pember, R.B.
1996-03-01
The discrete ordinates form of the radiative transport equation (RTE) is spatially discretized and solved using an adaptive mesh refinement (AMR) algorithm. This technique permits the local grid refinement to minimize spatial discretization error of the RTE. An error estimator is applied to define regions for local grid refinement; overlapping refined grids are recursively placed in these regions; and the RTE is then solved over the entire domain. The procedure continues until the spatial discretization error has been reduced to a sufficient level. The following aspects of the algorithm are discussed: error estimation, grid generation, communication between refined levels, and solution sequencing. This initial formulation employs the step scheme, and is valid for absorbing and isotopically scattering media in two-dimensional enclosures. The utility of the algorithm is tested by comparing the convergence characteristics and accuracy to those of the standard single-grid algorithm for several benchmark cases. The AMR algorithm provides a reduction in memory requirements and maintains the convergence characteristics of the standard single-grid algorithm; however, the cases illustrate that efficiency gains of the AMR algorithm will not be fully realized until three-dimensional geometries are considered.
CONSTRAINED-TRANSPORT MAGNETOHYDRODYNAMICS WITH ADAPTIVE MESH REFINEMENT IN CHARM
Miniati, Francesco; Martin, Daniel F. E-mail: DFMartin@lbl.gov
2011-07-01
We present the implementation of a three-dimensional, second-order accurate Godunov-type algorithm for magnetohydrodynamics (MHD) in the adaptive-mesh-refinement (AMR) cosmological code CHARM. The algorithm is based on the full 12-solve spatially unsplit corner-transport-upwind (CTU) scheme. The fluid quantities are cell-centered and are updated using the piecewise-parabolic method (PPM), while the magnetic field variables are face-centered and are evolved through application of the Stokes theorem on cell edges via a constrained-transport (CT) method. The so-called multidimensional MHD source terms required in the predictor step for high-order accuracy are applied in a simplified form which reduces their complexity in three dimensions without loss of accuracy or robustness. The algorithm is implemented on an AMR framework which requires specific synchronization steps across refinement levels. These include face-centered restriction and prolongation operations and a reflux-curl operation, which maintains a solenoidal magnetic field across refinement boundaries. The code is tested against a large suite of test problems, including convergence tests in smooth flows, shock-tube tests, classical two- and three-dimensional MHD tests, a three-dimensional shock-cloud interaction problem, and the formation of a cluster of galaxies in a fully cosmological context. The magnetic field divergence is shown to remain negligible throughout.
Constrained-transport Magnetohydrodynamics with Adaptive Mesh Refinement in CHARM
NASA Astrophysics Data System (ADS)
Miniati, Francesco; Martin, Daniel F.
2011-07-01
We present the implementation of a three-dimensional, second-order accurate Godunov-type algorithm for magnetohydrodynamics (MHD) in the adaptive-mesh-refinement (AMR) cosmological code CHARM. The algorithm is based on the full 12-solve spatially unsplit corner-transport-upwind (CTU) scheme. The fluid quantities are cell-centered and are updated using the piecewise-parabolic method (PPM), while the magnetic field variables are face-centered and are evolved through application of the Stokes theorem on cell edges via a constrained-transport (CT) method. The so-called multidimensional MHD source terms required in the predictor step for high-order accuracy are applied in a simplified form which reduces their complexity in three dimensions without loss of accuracy or robustness. The algorithm is implemented on an AMR framework which requires specific synchronization steps across refinement levels. These include face-centered restriction and prolongation operations and a reflux-curl operation, which maintains a solenoidal magnetic field across refinement boundaries. The code is tested against a large suite of test problems, including convergence tests in smooth flows, shock-tube tests, classical two- and three-dimensional MHD tests, a three-dimensional shock-cloud interaction problem, and the formation of a cluster of galaxies in a fully cosmological context. The magnetic field divergence is shown to remain negligible throughout.
AMR++: Object-oriented design for adaptive mesh refinement
Quinlan, D.
1998-12-01
The development of object-oriented libraries for scientific computing is complicated by the wide range of applications that are targeted and the complexity and wide range of numerical methods that are used. A problem is to design a library that can be customized to handle a wide range of target applications and increasingly complex numerical methods while maintaining a sufficiently useful library for simple problems. These problems have been classically at odds with one another and have compromised the design of many object-oriented library solutions. In this paper the authors detail the mechanisms used within AMR**, and object-oriented library for Adaptive Mesh Refinement (AMR), to provide the level of extensibility that is required to make AMR++ easily customizable for the more obscure applications while remaining small and simple for less complex applications. The goal has been to have a complex applications. The goal has been to have a complexity that matches the complexity of the target application. These mechanisms are general and extend to other libraries as well.
A New Approach to Parallel Dynamic Partitioning for Adaptive Unstructured Meshes
NASA Technical Reports Server (NTRS)
Heber, Gerd; Biswas, Rupak; Gao, Guang R.
1999-01-01
Classical mesh partitioning algorithms were designed for rather static situations, and their straightforward application in a dynamical framework may lead to unsatisfactory results, e.g., excessive data migration among processors. Furthermore, special attention should be paid to their amenability to parallelization. In this paper, a novel parallel method for the dynamic partitioning of adaptive unstructured meshes is described. It is based on a linear representation of the mesh using self-avoiding walks.
Global Load Balancing with Parallel Mesh Adaption on Distributed-Memory Systems
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Oliker, Leonid; Sohn, Andrew
1996-01-01
Dynamic mesh adaptation on unstructured grids is a powerful tool for efficiently computing unsteady problems to resolve solution features of interest. Unfortunately, this causes load inbalances among processors on a parallel machine. This paper described the parallel implementation of a tetrahedral mesh adaption scheme and a new global load balancing method. A heuristic remapping algorithm is presented that assigns partitions to processors such that the redistribution coast is minimized. Results indicate that the parallel performance of the mesh adaption code depends on the nature of the adaption region and show a 35.5X speedup on 64 processors of an SP2 when 35 percent of the mesh is randomly adapted. For large scale scientific computations, our load balancing strategy gives an almost sixfold reduction in solver execution times over non-balanced loads. Furthermore, our heuristic remappier yields processor assignments that are less than 3 percent of the optimal solutions, but requires only 1 percent of the computational time.
Global Load Balancing with Parallel Mesh Adaption on Distributed-Memory Systems
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Oliker, Leonid; Sohn, Andrew
1996-01-01
Dynamic mesh adaption on unstructured grids is a powerful tool for efficiently computing unsteady problems to resolve solution features of interest. Unfortunately, this causes load imbalance among processors on a parallel machine. This paper describes the parallel implementation of a tetrahedral mesh adaption scheme and a new global load balancing method. A heuristic remapping algorithm is presented that assigns partitions to processors such that the redistribution cost is minimized. Results indicate that the parallel performance of the mesh adaption code depends on the nature of the adaption region and show a 35.5X speedup on 64 processors of an SP2 when 35% of the mesh is randomly adapted. For large-scale scientific computations, our load balancing strategy gives almost a sixfold reduction in solver execution times over non-balanced loads. Furthermore, our heuristic remapper yields processor assignments that are less than 3% off the optimal solutions but requires only 1% of the computational time.
A Robust and Scalable Software Library for Parallel Adaptive Refinement on Unstructured Meshes
NASA Technical Reports Server (NTRS)
Lou, John Z.; Norton, Charles D.; Cwik, Thomas A.
1999-01-01
The design and implementation of Pyramid, a software library for performing parallel adaptive mesh refinement (PAMR) on unstructured meshes, is described. This software library can be easily used in a variety of unstructured parallel computational applications, including parallel finite element, parallel finite volume, and parallel visualization applications using triangular or tetrahedral meshes. The library contains a suite of well-designed and efficiently implemented modules that perform operations in a typical PAMR process. Among these are mesh quality control during successive parallel adaptive refinement (typically guided by a local-error estimator), parallel load-balancing, and parallel mesh partitioning using the ParMeTiS partitioner. The Pyramid library is implemented in Fortran 90 with an interface to the Message-Passing Interface (MPI) library, supporting code efficiency, modularity, and portability. An EM waveguide filter application, adaptively refined using the Pyramid library, is illustrated.
RAM: a Relativistic Adaptive Mesh Refinement Hydrodynamics Code
Zhang, Wei-Qun; MacFadyen, Andrew I.; /Princeton, Inst. Advanced Study
2005-06-06
The authors have developed a new computer code, RAM, to solve the conservative equations of special relativistic hydrodynamics (SRHD) using adaptive mesh refinement (AMR) on parallel computers. They have implemented a characteristic-wise, finite difference, weighted essentially non-oscillatory (WENO) scheme using the full characteristic decomposition of the SRHD equations to achieve fifth-order accuracy in space. For time integration they use the method of lines with a third-order total variation diminishing (TVD) Runge-Kutta scheme. They have also implemented fourth and fifth order Runge-Kutta time integration schemes for comparison. The implementation of AMR and parallelization is based on the FLASH code. RAM is modular and includes the capability to easily swap hydrodynamics solvers, reconstruction methods and physics modules. In addition to WENO they have implemented a finite volume module with the piecewise parabolic method (PPM) for reconstruction and the modified Marquina approximate Riemann solver to work with TVD Runge-Kutta time integration. They examine the difficulty of accurately simulating shear flows in numerical relativistic hydrodynamics codes. They show that under-resolved simulations of simple test problems with transverse velocity components produce incorrect results and demonstrate the ability of RAM to correctly solve these problems. RAM has been tested in one, two and three dimensions and in Cartesian, cylindrical and spherical coordinates. they have demonstrated fifth-order accuracy for WENO in one and two dimensions and performed detailed comparison with other schemes for which they show significantly lower convergence rates. Extensive testing is presented demonstrating the ability of RAM to address challenging open questions in relativistic astrophysics.
NASA Technical Reports Server (NTRS)
Ashford, Gregory A.; Powell, Kenneth G.
1995-01-01
A method for generating high quality unstructured triangular grids for high Reynolds number Navier-Stokes calculations about complex geometries is described. Careful attention is paid in the mesh generation process to resolving efficiently the disparate length scales which arise in these flows. First the surface mesh is constructed in a way which ensures that the geometry is faithfully represented. The volume mesh generation then proceeds in two phases thus allowing the viscous and inviscid regions of the flow to be meshed optimally. A solution-adaptive remeshing procedure which allows the mesh to adapt itself to flow features is also described. The procedure for tracking wakes and refinement criteria appropriate for shock detection are described. Although at present it has only been implemented in two dimensions, the grid generation process has been designed with the extension to three dimensions in mind. An implicit, higher-order, upwind method is also presented for computing compressible turbulent flows on these meshes. Two recently developed one-equation turbulence models have been implemented to simulate the effects of the fluid turbulence. Results for flow about a RAE 2822 airfoil and a Douglas three-element airfoil are presented which clearly show the improved resolution obtainable.
NASA Astrophysics Data System (ADS)
Ashford, Gregory A.; Powell, Kenneth G.
1995-10-01
A method for generating high quality unstructured triangular grids for high Reynolds number Navier-Stokes calculations about complex geometries is described. Careful attention is paid in the mesh generation process to resolving efficiently the disparate length scales which arise in these flows. First the surface mesh is constructed in a way which ensures that the geometry is faithfully represented. The volume mesh generation then proceeds in two phases thus allowing the viscous and inviscid regions of the flow to be meshed optimally. A solution-adaptive remeshing procedure which allows the mesh to adapt itself to flow features is also described. The procedure for tracking wakes and refinement criteria appropriate for shock detection are described. Although at present it has only been implemented in two dimensions, the grid generation process has been designed with the extension to three dimensions in mind. An implicit, higher-order, upwind method is also presented for computing compressible turbulent flows on these meshes. Two recently developed one-equation turbulence models have been implemented to simulate the effects of the fluid turbulence. Results for flow about a RAE 2822 airfoil and a Douglas three-element airfoil are presented which clearly show the improved resolution obtainable.
NASA Technical Reports Server (NTRS)
Barnard, Stephen T.; Simon, Horst; Lasinski, T. A. (Technical Monitor)
1994-01-01
The design of a parallel implementation of multilevel recursive spectral bisection is described. The goal is to implement a code that is fast enough to enable dynamic repartitioning of adaptive meshes.
Kohn, S.; Weare, J.; Ong, E.; Baden, S.
1997-05-01
We have applied structured adaptive mesh refinement techniques to the solution of the LDA equations for electronic structure calculations. Local spatial refinement concentrates memory resources and numerical effort where it is most needed, near the atomic centers and in regions of rapidly varying charge density. The structured grid representation enables us to employ efficient iterative solver techniques such as conjugate gradient with FAC multigrid preconditioning. We have parallelized our solver using an object- oriented adaptive mesh refinement framework.
Improved Simulation of Electrodiffusion in the Node of Ranvier by Mesh Adaptation.
Dione, Ibrahima; Deteix, Jean; Briffard, Thomas; Chamberland, Eric; Doyon, Nicolas
2016-01-01
In neural structures with complex geometries, numerical resolution of the Poisson-Nernst-Planck (PNP) equations is necessary to accurately model electrodiffusion. This formalism allows one to describe ionic concentrations and the electric field (even away from the membrane) with arbitrary spatial and temporal resolution which is impossible to achieve with models relying on cable theory. However, solving the PNP equations on complex geometries involves handling intricate numerical difficulties related either to the spatial discretization, temporal discretization or the resolution of the linearized systems, often requiring large computational resources which have limited the use of this approach. In the present paper, we investigate the best ways to use the finite elements method (FEM) to solve the PNP equations on domains with discontinuous properties (such as occur at the membrane-cytoplasm interface). 1) Using a simple 2D geometry to allow comparison with analytical solution, we show that mesh adaptation is a very (if not the most) efficient way to obtain accurate solutions while limiting the computational efforts, 2) We use mesh adaptation in a 3D model of a node of Ranvier to reveal details of the solution which are nearly impossible to resolve with other modelling techniques. For instance, we exhibit a non linear distribution of the electric potential within the membrane due to the non uniform width of the myelin and investigate its impact on the spatial profile of the electric field in the Debye layer. PMID:27548674
Improved Simulation of Electrodiffusion in the Node of Ranvier by Mesh Adaptation
Dione, Ibrahima; Briffard, Thomas; Doyon, Nicolas
2016-01-01
In neural structures with complex geometries, numerical resolution of the Poisson-Nernst-Planck (PNP) equations is necessary to accurately model electrodiffusion. This formalism allows one to describe ionic concentrations and the electric field (even away from the membrane) with arbitrary spatial and temporal resolution which is impossible to achieve with models relying on cable theory. However, solving the PNP equations on complex geometries involves handling intricate numerical difficulties related either to the spatial discretization, temporal discretization or the resolution of the linearized systems, often requiring large computational resources which have limited the use of this approach. In the present paper, we investigate the best ways to use the finite elements method (FEM) to solve the PNP equations on domains with discontinuous properties (such as occur at the membrane-cytoplasm interface). 1) Using a simple 2D geometry to allow comparison with analytical solution, we show that mesh adaptation is a very (if not the most) efficient way to obtain accurate solutions while limiting the computational efforts, 2) We use mesh adaptation in a 3D model of a node of Ranvier to reveal details of the solution which are nearly impossible to resolve with other modelling techniques. For instance, we exhibit a non linear distribution of the electric potential within the membrane due to the non uniform width of the myelin and investigate its impact on the spatial profile of the electric field in the Debye layer. PMID:27548674
Importance of dynamic mesh adaptivity for simulation of viscous fingering in porous media
NASA Astrophysics Data System (ADS)
Mostaghimi, P.; Jackson, M.; Pain, C.; Gorman, G.
2014-12-01
Viscous fingering is a major concern in many natural and engineered processes such as water flooding of heavy-oil reservoirs. Common reservoir simulators employ low-order finite volume/difference methods on structured grids to resolve this phenomenon. However, their approach suffers from a significant numerical dispersion error along the fingering patterns due to insufficient mesh resolution and smears out some important features of the flow. We propose use of an unstructured control volume finite element method for simulation of viscous fingering in porous media. Our approach is equipped with anisotropic mesh adaptivity where the mesh resolution is optimized based on the evolving features of flow. The adaptive algorithm uses a metric tensor field based on solution error estimates to locally control the size and shape of elements in the metric. We resolve the viscous fingering patterns accurately and reduce the numerical dispersion error significantly. The mesh optimization, generates an unstructured coarse mesh in other regions of the computational domain which significantly decreases the computational cost. The effect of grid resolution on the resolved fingers is thoroughly investigated. We analyze the computational cost of mesh adaptivty on unstructured mesh and compare it with common finite volume methods. The results of this study suggests that mesh adaptivity is an efficient and accurate approach for resolving complex behaviors and instabilities of flow in porous media such as viscous fingering.
NASA Astrophysics Data System (ADS)
Todarello, Giovanni; Vonck, Floris; Bourasseau, Sébastien; Peter, Jacques; Désidéri, Jean-Antoine
2016-05-01
A new goal-oriented mesh adaptation method for finite volume/finite difference schemes is extended from the structured mesh framework to a more suitable setting for adaptation of unstructured meshes. The method is based on the total derivative of the goal with respect to volume mesh nodes that is computable after the solution of the goal discrete adjoint equation. The asymptotic behaviour of this derivative is assessed on regularly refined unstructured meshes. A local refinement criterion is derived from the requirement of limiting the first order change in the goal that an admissible node displacement may cause. Mesh adaptations are then carried out for classical test cases of 2D Euler flows. Efficiency and local density of the adapted meshes are presented. They are compared with those obtained with a more classical mesh adaptation method in the framework of finite volume/finite difference schemes [46]. Results are very close although the present method only makes usage of the current grid.
New methods and astrophysical applications of adaptive mesh fluid simulations
NASA Astrophysics Data System (ADS)
Wang, Peng
The formation of stars, galaxies and supermassive black holes are among the most interesting unsolved problems in astrophysics. Those problems are highly nonlinear and involve enormous dynamical ranges. Thus numerical simulations with spatial adaptivity are crucial in understanding those processes. In this thesis, we discuss the development and application of adaptive mesh refinement (AMR) multi-physics fluid codes to simulate those nonlinear structure formation problems. To simulate the formation of star clusters, we have developed an AMR magnetohydrodynamics (MHD) code, coupled with radiative cooling. We have also developed novel algorithms for sink particle creation, accretion, merging and outflows, all of which are coupled with the fluid algorithms using operator splitting. With this code, we have been able to perform the first AMR-MHD simulation of star cluster formation for several dynamical times, including sink particle and protostellar outflow feedbacks. The results demonstrated that protostellar outflows can drive supersonic turbulence in dense clumps and explain the observed slow and inefficient star formation. We also suggest that global collapse rate is the most important factor in controlling massive star accretion rate. In the topics of galaxy formation, we discuss the results of three projects. In the first project, using cosmological AMR hydrodynamics simulations, we found that isolated massive star still forms in cosmic string wakes even though the mega-parsec scale structure has been perturbed significantly by the cosmic strings. In the second project, we calculated the dynamical heating rate in galaxy formation. We found that by balancing our heating rate with the atomic cooling rate, it gives a critical halo mass which agrees with the result of numerical simulations. This demonstrates that the effect of dynamical heating should be put into semi-analytical works in the future. In the third project, using our AMR-MHD code coupled with radiative
Star formation with adaptive mesh refinement and magnetohydrodynamics
NASA Astrophysics Data System (ADS)
Collins, David C.
2009-01-01
In this thesis, we develop an adaptive mesh refinement (AMR) code including magnetic fields, and use it to perform high resolution simulations of magnetized molecular clouds. The purpose of these simulations is to study present day star formation in the presence of turbulence and magnetic fields. We first present MHDEnzo, the extension of the cosmology and astrophysics code Enzo to include the effects magnetic fields. We use a higher order Godunov Riemann solver for the computation of interface fluxes; constrained transport to compute the electric field from those interface fluxes, which advances the induction equation in a divergence free manner; divergence free reconstruction technique to interpolate the magnetic fields to fine grids; operator splitting to include gravity and cosmological expansion. We present a series of test problems to demonstrate the quality of solution achieved. Additionally, we present several other solvers that were developed along the way. Finally we present the results from several AMR simulations that study isothermal turbulence in the presence of magnetic fields and self gravity. Ten simulations with initial Mach number 8.9 were studied varying several parameters; virial parameter a from 0.52 to 3.1; whether they were continuously stirred or allowed to decay; and the number of refinement levels (4 or 6). Measurements of the density probability density function (PDF) were made, showing both the expected log normal distribution and an additional power law. Measurements of the line of sight magnetic field vs. column density are done, giving excellent agreement with recent observations. The line width vs. size relationship is measured and compared with good agreement to observations, reproducing both turbulent and collapse signatures The core mass distribution is measured and agrees well with observations of Serpens and Perseus core samples, but the power-law distribution in Ophiuchus is not reproduced by our simulations. Finally we
Vortex-dominated conical-flow computations using unstructured adaptively-refined meshes
NASA Technical Reports Server (NTRS)
Batina, John T.
1989-01-01
A conical Euler/Navier-Stokes algorithm is presented for the computation of vortex-dominated flows. The flow solver involves a multistage Runge-Kutta time stepping scheme which uses a finite-volume spatial discretization on an unstructured grid made up of triangles. The algorithm also employs an adaptive mesh refinement procedure which enriches the mesh locally to more accurately resolve the vortical flow features. Results are presented for several highly-swept delta wing and circular cone cases at high angles of attack and at supersonic freestream flow conditions. Accurate solutions were obtained more efficiently when adaptive mesh refinement was used in contrast with refining the grid globally. The paper presents descriptions of the conical Euler/Navier-Stokes flow solver and adaptive mesh refinement procedures along with results which demonstrate the capability.
Zonal multigrid solution of compressible flow problems on unstructured and adaptive meshes
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1989-01-01
The simultaneous use of adaptive meshing techniques with a multigrid strategy for solving the 2-D Euler equations in the context of unstructured meshes is studied. To obtain optimal efficiency, methods capable of computing locally improved solutions without recourse to global recalculations are pursued. A method for locally refining an existing unstructured mesh, without regenerating a new global mesh is employed, and the domain is automatically partitioned into refined and unrefined regions. Two multigrid strategies are developed. In the first, time-stepping is performed on a global fine mesh covering the entire domain, and convergence acceleration is achieved through the use of zonal coarse grid accelerator meshes, which lie under the adaptively refined regions of the global fine mesh. Both schemes are shown to produce similar convergence rates to each other, and also with respect to a previously developed global multigrid algorithm, which performs time-stepping throughout the entire domain, on each mesh level. However, the present schemes exhibit higher computational efficiency due to the smaller number of operations on each level.
The adaptive GRP scheme for compressible fluid flows over unstructured meshes
NASA Astrophysics Data System (ADS)
Li, Jiequan; Zhang, Yongjin
2013-06-01
Unstructured mesh methods have attracted much attention in CFD community due to the flexibility for dealing with complex geometries and the ability to easily incorporate adaptive (moving) mesh strategies. When the finite volume framework is applied, a reliable solver is crucial for the construction of numerical fluxes, for which the generalized Riemann problem (GRP) scheme undertakes such a task in the sense of second order accuracy. Combining these techniques yields a second order accurate adaptive generalized Riemann problem (AGRP) scheme for two dimensional compressible fluid flows over unstructured triangular meshes. Besides the generation of meshes, the main process of this combination consists of two ingredients: Fluid dynamical evolution and mesh redistribution. The fluid dynamical evolution ingredient serves to evolve the compressible fluid flows on a fixed nonuniform triangular mesh with the direct Eulerian GRP solver. The role of the mesh redistribution is to redistribute mesh points on which a conservative interpolation formula is adopted to calculate the cell-averages for the conservative variables, and the gradients of primitive variables are reconstructed using the least squares method. Several examples are taken from various contexts to demonstrate the performance of such a program.
Yaqi Wang; Jean C. Ragusa
2011-02-01
Standard and goal-oriented adaptive mesh refinement (AMR) techniques are presented for the linear Boltzmann transport equation. A posteriori error estimates are employed to drive the AMR process and are based on angular-moment information rather than on directional information, leading to direction-independent adapted meshes. An error estimate based on a two-mesh approach and a jump-based error indicator are compared for various test problems. In addition to the standard AMR approach, where the global error in the solution is diminished, a goal-oriented AMR procedure is devised and aims at reducing the error in user-specified quantities of interest. The quantities of interest are functionals of the solution and may include, for instance, point-wise flux values or average reaction rates in a subdomain. A high-order (up to order 4) Discontinuous Galerkin technique with standard upwinding is employed for the spatial discretization; the discrete ordinates method is used to treat the angular variable.
Three-dimensional numerical simulations of falling films using an adaptive unstructured mesh
NASA Astrophysics Data System (ADS)
Pain, Chris; Xie, Zhihua; Matar, Omar
2015-11-01
Falling liquid films have rich wave dynamics, often occurring in many industrial applications, such as condensers, evaporators and chemical reactors. A number of numerical studies featuring falling liquid films are available in the literature; the majority of them, however, have focused on two-dimensional falling films. Far fewer studies have considered three-dimensional falling films, and those that have only studied the flow in a periodic domain. The objective of this study is to investigate flow dynamics of developing three-dimensional falling films using the Navier-Stokes equations coupled with interface capturing approach over extended domains. An adaptive, unstructured mesh modelling framework is employed here to study this problem, which can modify and adapt three-dimensional meshes to better represent the underlying physics of multiphase problems and reduce computational effort without sacrificing accuracy. Numerical examples of three-dimensional falling films in a long domain are presented and discussed. EPSRC Programme Grant, MEMPHIS, EP/K0039761/1.
Higher-order schemes with CIP method and adaptive Soroban grid towards mesh-free scheme
NASA Astrophysics Data System (ADS)
Yabe, Takashi; Mizoe, Hiroki; Takizawa, Kenji; Moriki, Hiroshi; Im, Hyo-Nam; Ogata, Youichi
2004-02-01
A new class of body-fitted grid system that can keep the third-order accuracy in time and space is proposed with the help of the CIP (constrained interpolation profile/cubic interpolated propagation) method. The grid system consists of the straight lines and grid points moving along these lines like abacus - Soroban in Japanese. The length of each line and the number of grid points in each line can be different. The CIP scheme is suitable to this mesh system and the calculation of large CFL (>10) at locally refined mesh is easily performed. Mesh generation and searching of upstream departure point are very simple and almost mesh-free treatment is possible. Adaptive grid movement and local mesh refinement are demonstrated.
An Adaptive Mesh Refinement Strategy for Immersed Boundary/Interface Methods.
Li, Zhilin; Song, Peng
2012-01-01
An adaptive mesh refinement strategy is proposed in this paper for the Immersed Boundary and Immersed Interface methods for two-dimensional elliptic interface problems involving singular sources. The interface is represented by the zero level set of a Lipschitz function φ(x,y). Our adaptive mesh refinement is done within a small tube of |φ(x,y)|≤ δ with finer Cartesian meshes. The discrete linear system of equations is solved by a multigrid solver. The AMR methods could obtain solutions with accuracy that is similar to those on a uniform fine grid by distributing the mesh more economically, therefore, reduce the size of the linear system of the equations. Numerical examples presented show the efficiency of the grid refinement strategy. PMID:22670155
Multifluid adaptive-mesh simulation of the solar wind interaction with the local interstellar medium
Kryukov, I. A.; Borovikov, S. N.; Pogorelov, N. V.; Zank, G. P.
2006-09-26
DOE's SciDAC adaptive mesh refinement code Chombo has been modified for solution of compressible MHD flows with the application of high resolution, shock-capturing numerical schemes. The code developed is further extended to involve multiple fluids and applied to the problem of the solar wind interaction with the local interstellar medium. For this purpose, a set of MHD equations is solved together with a few sets of the Euler gas dynamics equations, depending on the number of neutral fluids included in the model. Our first results are presented that were obtained in the framework of an axially symmetric multifluid model which is applicable to magnetic-field-aligned flows. Details are shown of the generation and development of Rayleigh-Taylor and Kelvin-Helmholtz instabilities of the heliopause. A comparison is given of the results obtained with a two- and four-fluid models.
Fast animation of lightning using an adaptive mesh.
Kim, Theodore; Lin, Ming C
2007-01-01
We present a fast method for simulating, animating, and rendering lightning using adaptive grids. The "dielectric breakdown model" is an elegant algorithm for electrical pattern formation that we extend to enable animation of lightning. The simulation can be slow, particularly in 3D, because it involves solving a large Poisson problem. Losasso et al. recently proposed an octree data structure for simulating water and smoke, and we show that this discretization can be applied to the problem of lightning simulation as well. However, implementing the incomplete Cholesky conjugate gradient (ICCG) solver for this problem can be daunting, so we provide an extensive discussion of implementation issues. ICCG solvers can usually be accelerated using "Eisenstat's trick," but the trick cannot be directly applied to the adaptive case. Fortunately, we show that an "almost incomplete Cholesky" factorization can be computed so that Eisenstat's trick can still be used. We then present a fast rendering method based on convolution that is competitive with Monte Carlo ray tracing but orders of magnitude faster, and we also show how to further improve the visual results using jittering. PMID:17218754
A simple adaptive mesh generator for 2-D finite element calculations
Fernandez, F.A.; Yong, Y.C.; Ettinger, R.D. )
1993-03-01
A strategy for adaptive mesh generation is proposed. The method consists of the use of a suitably defined density function', which can either be defined by the user or be calculated from a previous approximate solution, to guide the generation of a new mesh. This new mesh is built starting from a minimal number of triangular elements which are then in several sweeps, repeatedly refined according to the density function. The Delaunay algorithm is used in each stage to keep the shape of the triangles as equilateral as possible.
A Numerical Study of Mesh Adaptivity in Multiphase Flows with Non-Newtonian Fluids
NASA Astrophysics Data System (ADS)
Percival, James; Pavlidis, Dimitrios; Xie, Zhihua; Alberini, Federico; Simmons, Mark; Pain, Christopher; Matar, Omar
2014-11-01
We present an investigation into the computational efficiency benefits of dynamic mesh adaptivity in the numerical simulation of transient multiphase fluid flow problems involving Non-Newtonian fluids. Such fluids appear in a range of industrial applications, from printing inks to toothpastes and introduce new challenges for mesh adaptivity due to the additional ``memory'' of viscoelastic fluids. Nevertheless, the multiscale nature of these flows implies huge potential benefits for a successful implementation. The study is performed using the open source package Fluidity, which couples an unstructured mesh control volume finite element solver for the multiphase Navier-Stokes equations to a dynamic anisotropic mesh adaptivity algorithm, based on estimated solution interpolation error criteria, and conservative mesh-to-mesh interpolation routine. The code is applied to problems involving rheologies ranging from simple Newtonian to shear-thinning to viscoelastic materials and verified against experimental data for various industrial and microfluidic flows. This work was undertaken as part of the EPSRC MEMPHIS programme grant EP/K003976/1.
ENZO+MORAY: radiation hydrodynamics adaptive mesh refinement simulations with adaptive ray tracing
NASA Astrophysics Data System (ADS)
Wise, John H.; Abel, Tom
2011-07-01
We describe a photon-conserving radiative transfer algorithm, using a spatially-adaptive ray-tracing scheme, and its parallel implementation into the adaptive mesh refinement cosmological hydrodynamics code ENZO. By coupling the solver with the energy equation and non-equilibrium chemistry network, our radiation hydrodynamics framework can be utilized to study a broad range of astrophysical problems, such as stellar and black hole feedback. Inaccuracies can arise from large time-steps and poor sampling; therefore, we devised an adaptive time-stepping scheme and a fast approximation of the optically-thin radiation field with multiple sources. We test the method with several radiative transfer and radiation hydrodynamics tests that are given in Iliev et al. We further test our method with more dynamical situations, for example, the propagation of an ionization front through a Rayleigh-Taylor instability, time-varying luminosities and collimated radiation. The test suite also includes an expanding H II region in a magnetized medium, utilizing the newly implemented magnetohydrodynamics module in ENZO. This method linearly scales with the number of point sources and number of grid cells. Our implementation is scalable to 512 processors on distributed memory machines and can include the radiation pressure and secondary ionizations from X-ray radiation. It is included in the newest public release of ENZO.
3D Adaptive Mesh Refinement Simulations of Pellet Injection in Tokamaks
R. Samtaney; S.C. Jardin; P. Colella; D.F. Martin
2003-10-20
We present results of Adaptive Mesh Refinement (AMR) simulations of the pellet injection process, a proven method of refueling tokamaks. AMR is a computationally efficient way to provide the resolution required to simulate realistic pellet sizes relative to device dimensions. The mathematical model comprises of single-fluid MHD equations with source terms in the continuity equation along with a pellet ablation rate model. The numerical method developed is an explicit unsplit upwinding treatment of the 8-wave formulation, coupled with a MAC projection method to enforce the solenoidal property of the magnetic field. The Chombo framework is used for AMR. The role of the E x B drift in mass redistribution during inside and outside pellet injections is emphasized.
White Dwarf Mergers on Adaptive Meshes. I. Methodology and Code Verification
NASA Astrophysics Data System (ADS)
Katz, Max P.; Zingale, Michael; Calder, Alan C.; Swesty, F. Douglas; Almgren, Ann S.; Zhang, Weiqun
2016-03-01
The Type Ia supernova (SN Ia) progenitor problem is one of the most perplexing and exciting problems in astrophysics, requiring detailed numerical modeling to complement observations of these explosions. One possible progenitor that has merited recent theoretical attention is the white dwarf (WD) merger scenario, which has the potential to naturally explain many of the observed characteristics of SNe Ia. To date there have been relatively few self-consistent simulations of merging WD systems using mesh-based hydrodynamics. This is the first paper in a series describing simulations of these systems using a hydrodynamics code with adaptive mesh refinement. In this paper we describe our numerical methodology and discuss our implementation in the compressible hydrodynamics code CASTRO, which solves the Euler equations, and the Poisson equation for self-gravity, and couples the gravitational and rotation forces to the hydrodynamics. Standard techniques for coupling gravitation and rotation forces to the hydrodynamics do not adequately conserve the total energy of the system for our problem, but recent advances in the literature allow progress and we discuss our implementation here. We present a set of test problems demonstrating the extent to which our software sufficiently models a system where large amounts of mass are advected on the computational domain over long timescales. Future papers in this series will describe our treatment of the initial conditions of these systems and will examine the early phases of the merger to determine its viability for triggering a thermonuclear detonation.
Development of a scalable gas-dynamics solver with adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Korkut, Burak
There are various computational physics areas in which Direct Simulation Monte Carlo (DSMC) and Particle in Cell (PIC) methods are being employed. The accuracy of results from such simulations depend on the fidelity of the physical models being used. The computationally demanding nature of these problems make them ideal candidates to make use of modern supercomputers. The software developed to run such simulations also needs special attention so that the maintainability and extendability is considered with the recent numerical methods and programming paradigms. Suited for gas-dynamics problems, a software called SUGAR (Scalable Unstructured Gas dynamics with Adaptive mesh Refinement) has recently been developed and written in C++ and MPI. Physical and numerical models were added to this framework to simulate ion thruster plumes. SUGAR is used to model the charge-exchange (CEX) reactions occurring between the neutral and ion species as well as the induced electric field effect due to ions. Multiple adaptive mesh refinement (AMR) meshes were used in order to capture different physical length scales present in the flow. A multiple-thruster configuration was run to extend the studies to cases for which there is no axial or radial symmetry present that could only be modeled with a three-dimensional simulation capability. The combined plume structure showed interactions between individual thrusters where AMR capability captured this in an automated way. The back flow for ions was found to occur when CEX and momentum-exchange (MEX) collisions are present and strongly enhanced when the induced electric field is considered. The ion energy distributions in the back flow region were obtained and it was found that the inclusion of the electric field modeling is the most important factor in determining its shape. The plume back flow structure was also examined for a triple-thruster, 3-D geometry case and it was found that the ion velocity in the back flow region appears to be
Adaptive unstructured meshing for thermal stress analysis of built-up structures
NASA Technical Reports Server (NTRS)
Dechaumphai, Pramote
1992-01-01
An adaptive unstructured meshing technique for mechanical and thermal stress analysis of built-up structures has been developed. A triangular membrane finite element and a new plate bending element are evaluated on a panel with a circular cutout and a frame stiffened panel. The adaptive unstructured meshing technique, without a priori knowledge of the solution to the problem, generates clustered elements only where needed. An improved solution accuracy is obtained at a reduced problem size and analysis computational time as compared to the results produced by the standard finite element procedure.
Using Multi-threading for the Automatic Load Balancing of 2D Adaptive Finite Element Meshes
NASA Technical Reports Server (NTRS)
Heber, Gerd; Biswas, Rupak; Thulasiraman, Parimala; Gao, Guang R.; Saini, Subhash (Technical Monitor)
1998-01-01
In this paper, we present a multi-threaded approach for the automatic load balancing of adaptive finite element (FE) meshes The platform of our choice is the EARTH multi-threaded system which offers sufficient capabilities to tackle this problem. We implement the adaption phase of FE applications oil triangular meshes and exploit the EARTH token mechanism to automatically balance the resulting irregular and highly nonuniform workload. We discuss the results of our experiments oil EARTH-SP2, on implementation of EARTH on the IBM SP2 with different load balancing strategies that are built into the runtime system.
Parallelization of Unsteady Adaptive Mesh Refinement for Unstructured Navier-Stokes Solvers
NASA Technical Reports Server (NTRS)
Schwing, Alan M.; Nompelis, Ioannis; Candler, Graham V.
2014-01-01
This paper explores the implementation of the MPI parallelization in a Navier-Stokes solver using adaptive mesh re nement. Viscous and inviscid test problems are considered for the purpose of benchmarking, as are implicit and explicit time advancement methods. The main test problem for comparison includes e ects from boundary layers and other viscous features and requires a large number of grid points for accurate computation. Ex- perimental validation against double cone experiments in hypersonic ow are shown. The adaptive mesh re nement shows promise for a staple test problem in the hypersonic com- munity. Extension to more advanced techniques for more complicated ows is described.
Thickness-based adaptive mesh refinement methods for multi-phase flow simulations with thin regions
Chen, Xiaodong; Yang, Vigor
2014-07-15
In numerical simulations of multi-scale, multi-phase flows, grid refinement is required to resolve regions with small scales. A notable example is liquid-jet atomization and subsequent droplet dynamics. It is essential to characterize the detailed flow physics with variable length scales with high fidelity, in order to elucidate the underlying mechanisms. In this paper, two thickness-based mesh refinement schemes are developed based on distance- and topology-oriented criteria for thin regions with confining wall/plane of symmetry and in any situation, respectively. Both techniques are implemented in a general framework with a volume-of-fluid formulation and an adaptive-mesh-refinement capability. The distance-oriented technique compares against a critical value, the ratio of an interfacial cell size to the distance between the mass center of the cell and a reference plane. The topology-oriented technique is developed from digital topology theories to handle more general conditions. The requirement for interfacial mesh refinement can be detected swiftly, without the need of thickness information, equation solving, variable averaging or mesh repairing. The mesh refinement level increases smoothly on demand in thin regions. The schemes have been verified and validated against several benchmark cases to demonstrate their effectiveness and robustness. These include the dynamics of colliding droplets, droplet motions in a microchannel, and atomization of liquid impinging jets. Overall, the thickness-based refinement technique provides highly adaptive meshes for problems with thin regions in an efficient and fully automatic manner.
NASA Astrophysics Data System (ADS)
Tang, Qiuyan; Wang, Jing; Lv, Pin; Sun, Quan
2015-10-01
Propagation simulation method and choosing mesh grid are both very important to get the correct propagation results in wave optics simulation. A new angular spectrum propagation method with alterable mesh grid based on the traditional angular spectrum method and the direct FFT method is introduced. With this method, the sampling space after propagation is not limited to propagation methods no more, but freely alterable. However, choosing mesh grid on target board influences the validity of simulation results directly. So an adaptive mesh choosing method based on wave characteristics is proposed with the introduced propagation method. We can calculate appropriate mesh grids on target board to get satisfying results. And for complex initial wave field or propagation through inhomogeneous media, we can also calculate and set the mesh grid rationally according to above method. Finally, though comparing with theoretical results, it's shown that the simulation result with the proposed method coinciding with theory. And by comparing with the traditional angular spectrum method and the direct FFT method, it's known that the proposed method is able to adapt to a wider range of Fresnel number conditions. That is to say, the method can simulate propagation results efficiently and correctly with propagation distance of almost zero to infinity. So it can provide better support for more wave propagation applications such as atmospheric optics, laser propagation and so on.
Detached Eddy Simulation of the UH-60 Rotor Wake Using Adaptive Mesh Refinement
NASA Technical Reports Server (NTRS)
Chaderjian, Neal M.; Ahmad, Jasim U.
2012-01-01
Time-dependent Navier-Stokes flow simulations have been carried out for a UH-60 rotor with simplified hub in forward flight and hover flight conditions. Flexible rotor blades and flight trim conditions are modeled and established by loosely coupling the OVERFLOW Computational Fluid Dynamics (CFD) code with the CAMRAD II helicopter comprehensive code. High order spatial differences, Adaptive Mesh Refinement (AMR), and Detached Eddy Simulation (DES) are used to obtain highly resolved vortex wakes, where the largest turbulent structures are captured. Special attention is directed towards ensuring the dual time accuracy is within the asymptotic range, and verifying the loose coupling convergence process using AMR. The AMR/DES simulation produced vortical worms for forward flight and hover conditions, similar to previous results obtained for the TRAM rotor in hover. AMR proved to be an efficient means to capture a rotor wake without a priori knowledge of the wake shape.
On the Computation of Integral Curves in Adaptive Mesh Refinement Vector Fields
Deines, Eduard; Weber, Gunther H.; Garth, Christoph; Van Straalen, Brian; Borovikov, Sergey; Martin, Daniel F.; Joy, Kenneth I.
2011-06-27
Integral curves, such as streamlines, streaklines, pathlines, and timelines, are an essential tool in the analysis of vector field structures, offering straightforward and intuitive interpretation of visualization results. While such curves have a long-standing tradition in vector field visualization, their application to Adaptive Mesh Refinement (AMR) simulation results poses unique problems. AMR is a highly effective discretization method for a variety of physical simulation problems and has recently been applied to the study of vector fields in flow and magnetohydrodynamic applications. The cell-centered nature of AMR data and discontinuities in the vector field representation arising from AMR level boundaries complicate the application of numerical integration methods to compute integral curves. In this paper, we propose a novel approach to alleviate these problems and show its application to streamline visualization in an AMR model of the magnetic field of the solar system as well as to a simulation of two incompressible viscous vortex rings merging.
The parallelization of an advancing-front, all-quadrilateral meshing algorithm for adaptive analysis
Lober, R.R.; Tautges, T.J.; Cairncross, R.A.
1995-11-01
The ability to perform effective adaptive analysis has become a critical issue in the area of physical simulation. Of the multiple technologies required to realize a parallel adaptive analysis capability, automatic mesh generation is an enabling technology, filling a critical need in the appropriate discretization of a problem domain. The paving algorithm`s unique ability to generate a function-following quadrilateral grid is a substantial advantage in Sandia`s pursuit of a modified h-method adaptive capability. This characteristic combined with a strong transitioning ability allow the paving algorithm to place elements where an error function indicates more mesh resolution is needed. Although the original paving algorithm is highly serial, a two stage approach has been designed to parallelize the algorithm but also retain the nice qualities of the serial algorithm. The authors approach also allows the subdomain decomposition used by the meshing code to be shared with the finite element physics code, eliminating the need for data transfer across the processors between the analysis and remeshing steps. In addition, the meshed subdomains are adjusted with a dynamic load balancer to improve the original decomposition and maintain load efficiency each time the mesh has been regenerated. This initial parallel implementation assumes an approach of restarting the physics problem from time zero at each interaction, with a refined mesh adapting to the previous iterations objective function. The remeshing tools are being developed to enable real time remeshing and geometry regeneration. Progress on the redesign of the paving algorithm for parallel operation is discussed including extensions allowing adaptive control and geometry regeneration.
Modeling study of mesh conductors and their electroluminescent devices
NASA Astrophysics Data System (ADS)
Hu, Bin; Li, Dapeng; Manandhar, Prakash; Fan, Qinguo; Kasilingam, Dayalan; Calvert, Paul
2015-02-01
Numerical models were established to correlate with the experimentally measured properties of mesh conductors previously developed through a combined process of dip coating carbon nanotubes and inkjet printing poly 3,4-ethylenedioxythiophene: poly styrene sulfonate. The electroluminescent (EL) devices assembled with such mesh conductors as front electrodes were modeled by commercially available finite element method software COMSOL Multiphysics. The modeling results are in agreement with those from the experiments and suggest that an optimized fiber arrangement is the key for further improving the performance of EL devices based on mesh conductors.
Implementation of a mesh adaptive scheme based on an element-level error indicator
NASA Technical Reports Server (NTRS)
Keating, Scott; Felippa, Carlos A.; Militello, Carmelo
1993-01-01
We investigate the formulation and application of element-level error indicators based on parametrized variational principles. The qualifier 'element-level' means that no information from adjacent elements is used for error estimation. This property is ideally suited to drive adaptive mesh refinement on parallel computers where access to neighboring elements resident on different processors may incur significant computational overhead. Furthermore, such indicators are not affected by physical jumps at junctures or interfaces. An element-level indicator has been derived from the higher-order element energy and applied to r and h mesh adaptation of meshes in plates and shell structures. We report on our initial experiments with a cylindrical shell that intersects with fist plates forming a simplified 'wing-body intersection' benchmark problem.
Adaptive mesh refinement techniques for the immersed interface method applied to flow problems.
Li, Zhilin; Song, Peng
2013-06-01
In this paper, we develop an adaptive mesh refinement strategy of the Immersed Interface Method for flow problems with a moving interface. The work is built on the AMR method developed for two-dimensional elliptic interface problems in the paper [12] (CiCP, 12(2012), 515-527). The interface is captured by the zero level set of a Lipschitz continuous function φ(x, y, t). Our adaptive mesh refinement is built within a small band of |φ(x, y, t)| ≤ δ with finer Cartesian meshes. The AMR-IIM is validated for Stokes and Navier-Stokes equations with exact solutions, moving interfaces driven by the surface tension, and classical bubble deformation problems. A new simple area preserving strategy is also proposed in this paper for the level set method. PMID:23794763
Hornung, R.D.
1996-12-31
An adaptive local mesh refinement (AMR) algorithm originally developed for unsteady gas dynamics is extended to multi-phase flow in porous media. Within the AMR framework, we combine specialized numerical methods to treat the different aspects of the partial differential equations. Multi-level iteration and domain decomposition techniques are incorporated to accommodate elliptic/parabolic behavior. High-resolution shock capturing schemes are used in the time integration of the hyperbolic mass conservation equations. When combined with AMR, these numerical schemes provide high resolution locally in a more efficient manner than if they were applied on a uniformly fine computational mesh. We will discuss the interplay of physical, mathematical, and numerical concerns in the application of adaptive mesh refinement to flow in porous media problems of practical interest.
Adaptive mesh refinement with spectral accuracy for magnetohydrodynamics in two space dimensions
NASA Astrophysics Data System (ADS)
Rosenberg, D.; Pouquet, A.; Mininni, P. D.
2007-08-01
We examine the effect of accuracy of high-order spectral element methods, with or without adaptive mesh refinement (AMR), in the context of a classical configuration of magnetic reconnection in two space dimensions, the so-called Orszag-Tang (OT) vortex made up of a magnetic X-point centred on a stagnation point of the velocity. A recently developed spectral-element adaptive refinement incompressible magnetohydrodynamic (MHD) code is applied to simulate this problem. The MHD solver is explicit, and uses the Elsässer formulation on high-order elements. It automatically takes advantage of the adaptive grid mechanics that have been described elsewhere in the fluid context (Rosenberg et al 2006 J. Comput. Phys. 215 59-80) the code allows both statically refined and dynamically refined grids. Tests of the algorithm using analytic solutions are described, and comparisons of the OT solutions with pseudo-spectral computations are performed. We demonstrate for moderate Reynolds numbers that the algorithms using both static and refined grids reproduce the pseudo-spectral solutions quite well. We show that low-order truncation—even with a comparable number of global degrees of freedom—fails to correctly model some strong (sup-norm) quantities in this problem, even though it satisfies adequately the weak (integrated) balance diagnostics.
Parallel, Gradient-Based Anisotropic Mesh Adaptation for Re-entry Vehicle Configurations
NASA Technical Reports Server (NTRS)
Bibb, Karen L.; Gnoffo, Peter A.; Park, Michael A.; Jones, William T.
2006-01-01
Two gradient-based adaptation methodologies have been implemented into the Fun3d refine GridEx infrastructure. A spring-analogy adaptation which provides for nodal movement to cluster mesh nodes in the vicinity of strong shocks has been extended for general use within Fun3d, and is demonstrated for a 70 sphere cone at Mach 2. A more general feature-based adaptation metric has been developed for use with the adaptation mechanics available in Fun3d, and is applicable to any unstructured, tetrahedral, flow solver. The basic functionality of general adaptation is explored through a case of flow over the forebody of a 70 sphere cone at Mach 6. A practical application of Mach 10 flow over an Apollo capsule, computed with the Felisa flow solver, is given to compare the adaptive mesh refinement with uniform mesh refinement. The examples of the paper demonstrate that the gradient-based adaptation capability as implemented can give an improvement in solution quality.
Failure of Anisotropic Unstructured Mesh Adaption Based on Multidimensional Residual Minimization
NASA Technical Reports Server (NTRS)
Wood, William A.; Kleb, William L.
2003-01-01
An automated anisotropic unstructured mesh adaptation strategy is proposed, implemented, and assessed for the discretization of viscous flows. The adaption criteria is based upon the minimization of the residual fluctuations of a multidimensional upwind viscous flow solver. For scalar advection, this adaption strategy has been shown to use fewer grid points than gradient based adaption, naturally aligning mesh edges with discontinuities and characteristic lines. The adaption utilizes a compact stencil and is local in scope, with four fundamental operations: point insertion, point deletion, edge swapping, and nodal displacement. Evaluation of the solution-adaptive strategy is performed for a two-dimensional blunt body laminar wind tunnel case at Mach 10. The results demonstrate that the strategy suffers from a lack of robustness, particularly with regard to alignment of the bow shock in the vicinity of the stagnation streamline. In general, constraining the adaption to such a degree as to maintain robustness results in negligible improvement to the solution. Because the present method fails to consistently or significantly improve the flow solution, it is rejected in favor of simple uniform mesh refinement.
SU-D-207-04: GPU-Based 4D Cone-Beam CT Reconstruction Using Adaptive Meshing Method
Zhong, Z; Gu, X; Iyengar, P; Mao, W; Wang, J; Guo, X
2015-06-15
Purpose: Due to the limited number of projections at each phase, the image quality of a four-dimensional cone-beam CT (4D-CBCT) is often degraded, which decreases the accuracy of subsequent motion modeling. One of the promising methods is the simultaneous motion estimation and image reconstruction (SMEIR) approach. The objective of this work is to enhance the computational speed of the SMEIR algorithm using adaptive feature-based tetrahedral meshing and GPU-based parallelization. Methods: The first step is to generate the tetrahedral mesh based on the features of a reference phase 4D-CBCT, so that the deformation can be well captured and accurately diffused from the mesh vertices to voxels of the image volume. After the mesh generation, the updated motion model and other phases of 4D-CBCT can be obtained by matching the 4D-CBCT projection images at each phase with the corresponding forward projections of the deformed reference phase of 4D-CBCT. The entire process of this 4D-CBCT reconstruction method is implemented on GPU, resulting in significantly increasing the computational efficiency due to its tremendous parallel computing ability. Results: A 4D XCAT digital phantom was used to test the proposed mesh-based image reconstruction algorithm. The image Result shows both bone structures and inside of the lung are well-preserved and the tumor position can be well captured. Compared to the previous voxel-based CPU implementation of SMEIR, the proposed method is about 157 times faster for reconstructing a 10 -phase 4D-CBCT with dimension 256×256×150. Conclusion: The GPU-based parallel 4D CBCT reconstruction method uses the feature-based mesh for estimating motion model and demonstrates equivalent image Result with previous voxel-based SMEIR approach, with significantly improved computational speed.
CRASH: A BLOCK-ADAPTIVE-MESH CODE FOR RADIATIVE SHOCK HYDRODYNAMICS-IMPLEMENTATION AND VERIFICATION
Van der Holst, B.; Toth, G.; Sokolov, I. V.; Myra, E. S.; Fryxell, B.; Drake, R. P.; Powell, K. G.; Holloway, J. P.; Stout, Q.; Adams, M. L.; Morel, J. E.; Karni, S.
2011-06-01
We describe the Center for Radiative Shock Hydrodynamics (CRASH) code, a block-adaptive-mesh code for multi-material radiation hydrodynamics. The implementation solves the radiation diffusion model with a gray or multi-group method and uses a flux-limited diffusion approximation to recover the free-streaming limit. Electrons and ions are allowed to have different temperatures and we include flux-limited electron heat conduction. The radiation hydrodynamic equations are solved in the Eulerian frame by means of a conservative finite-volume discretization in either one-, two-, or three-dimensional slab geometry or in two-dimensional cylindrical symmetry. An operator-split method is used to solve these equations in three substeps: (1) an explicit step of a shock-capturing hydrodynamic solver; (2) a linear advection of the radiation in frequency-logarithm space; and (3) an implicit solution of the stiff radiation diffusion, heat conduction, and energy exchange. We present a suite of verification test problems to demonstrate the accuracy and performance of the algorithms. The applications are for astrophysics and laboratory astrophysics. The CRASH code is an extension of the Block-Adaptive Tree Solarwind Roe Upwind Scheme (BATS-R-US) code with a new radiation transfer and heat conduction library and equation-of-state and multi-group opacity solvers. Both CRASH and BATS-R-US are part of the publicly available Space Weather Modeling Framework.
CRASH: A Block-adaptive-mesh Code for Radiative Shock Hydrodynamics—Implementation and Verification
NASA Astrophysics Data System (ADS)
van der Holst, B.; Tóth, G.; Sokolov, I. V.; Powell, K. G.; Holloway, J. P.; Myra, E. S.; Stout, Q.; Adams, M. L.; Morel, J. E.; Karni, S.; Fryxell, B.; Drake, R. P.
2011-06-01
We describe the Center for Radiative Shock Hydrodynamics (CRASH) code, a block-adaptive-mesh code for multi-material radiation hydrodynamics. The implementation solves the radiation diffusion model with a gray or multi-group method and uses a flux-limited diffusion approximation to recover the free-streaming limit. Electrons and ions are allowed to have different temperatures and we include flux-limited electron heat conduction. The radiation hydrodynamic equations are solved in the Eulerian frame by means of a conservative finite-volume discretization in either one-, two-, or three-dimensional slab geometry or in two-dimensional cylindrical symmetry. An operator-split method is used to solve these equations in three substeps: (1) an explicit step of a shock-capturing hydrodynamic solver; (2) a linear advection of the radiation in frequency-logarithm space; and (3) an implicit solution of the stiff radiation diffusion, heat conduction, and energy exchange. We present a suite of verification test problems to demonstrate the accuracy and performance of the algorithms. The applications are for astrophysics and laboratory astrophysics. The CRASH code is an extension of the Block-Adaptive Tree Solarwind Roe Upwind Scheme (BATS-R-US) code with a new radiation transfer and heat conduction library and equation-of-state and multi-group opacity solvers. Both CRASH and BATS-R-US are part of the publicly available Space Weather Modeling Framework.
CRASH: A Block-Adaptive-Mesh Code for Radiative Shock Hydrodynamics
NASA Astrophysics Data System (ADS)
van der Holst, B.; Toth, G.; Sokolov, I. V.; Powell, K. G.; Holloway, J. P.; Myra, E. S.; Stout, Q.; Adams, M. L.; Morel, J. E.; Drake, R. P.
2011-01-01
We describe the CRASH (Center for Radiative Shock Hydrodynamics) code, a block adaptive mesh code for multi-material radiation hydrodynamics. The implementation solves the radiation diffusion model with the gray or multigroup method and uses a flux limited diffusion approximation to recover the free-streaming limit. The electrons and ions are allowed to have different temperatures and we include a flux limited electron heat conduction. The radiation hydrodynamic equations are solved in the Eulerian frame by means of a conservative finite volume discretization in either one, two, or three-dimensional slab geometry or in two-dimensional cylindrical symmetry. An operator split method is used to solve these equations in three substeps: (1) solve the hydrodynamic equations with shock-capturing schemes, (2) a linear advection of the radiation in frequency-logarithm space, and (3) an implicit solve of the stiff radiation diffusion, heat conduction, and energy exchange. We present a suite of verification test problems to demonstrate the accuracy and performance of the algorithms. The CRASH code is an extension of the Block-Adaptive Tree Solarwind Roe Upwind Scheme (BATS-R-US) code with this new radiation transfer and heat conduction library and equation-of-state and multigroup opacity solvers. Both CRASH and BATS-R-US are part of the publicly available Space Weather Modeling Framework (SWMF).
Time-dependent grid adaptation for meshes of triangles and tetrahedra
NASA Technical Reports Server (NTRS)
Rausch, Russ D.
1993-01-01
This paper presents in viewgraph form a method of optimizing grid generation for unsteady CFD flow calculations that distributes the numerical error evenly throughout the mesh. Adaptive meshing is used to locally enrich in regions of relatively large errors and to locally coarsen in regions of relatively small errors. The enrichment/coarsening procedures are robust for isotropic cells; however, enrichment of high aspect ratio cells may fail near boundary surfaces with relatively large curvature. The enrichment indicator worked well for the cases shown, but in general requires user supervision for a more efficient solution.
Using Multithreading for the Automatic Load Balancing of 2D Adaptive Finite Element Meshes
NASA Technical Reports Server (NTRS)
Heber, Gerd; Biswas, Rupak; Thulasiraman, Parimala; Gao, Guang R.; Bailey, David H. (Technical Monitor)
1998-01-01
In this paper, we present a multi-threaded approach for the automatic load balancing of adaptive finite element (FE) meshes. The platform of our choice is the EARTH multi-threaded system which offers sufficient capabilities to tackle this problem. We implement the question phase of FE applications on triangular meshes, and exploit the EARTH token mechanism to automatically balance the resulting irregular and highly nonuniform workload. We discuss the results of our experiments on EARTH-SP2, an implementation of EARTH on the IBM SP2, with different load balancing strategies that are built into the runtime system.
NASA Astrophysics Data System (ADS)
Burago, N. G.; Nikitin, I. S.; Yakushev, V. L.
2016-06-01
Techniques that improve the accuracy of numerical solutions and reduce their computational costs are discussed as applied to continuum mechanics problems with complex time-varying geometry. The approach combines shock-capturing computations with the following methods: (1) overlapping meshes for specifying complex geometry; (2) elastic arbitrarily moving adaptive meshes for minimizing the approximation errors near shock waves, boundary layers, contact discontinuities, and moving boundaries; (3) matrix-free implementation of efficient iterative and explicit-implicit finite element schemes; (4) balancing viscosity (version of the stabilized Petrov-Galerkin method); (5) exponential adjustment of physical viscosity coefficients; and (6) stepwise correction of solutions for providing their monotonicity and conservativeness.
Adaptive Mesh Refinement in Curvilinear Body-Fitted Grid Systems
NASA Technical Reports Server (NTRS)
Steinthorsson, Erlendur; Modiano, David; Colella, Phillip
1995-01-01
To be truly compatible with structured grids, an AMR algorithm should employ a block structure for the refined grids to allow flow solvers to take advantage of the strengths of unstructured grid systems, such as efficient solution algorithms for implicit discretizations and multigrid schemes. One such algorithm, the AMR algorithm of Berger and Colella, has been applied to and adapted for use with body-fitted structured grid systems. Results are presented for a transonic flow over a NACA0012 airfoil (AGARD-03 test case) and a reflection of a shock over a double wedge.
The design of a parallel adaptive paving all-quadrilateral meshing algorithm
Tautges, T.J.; Lober, R.R.; Vaughan, C.
1995-08-01
Adaptive finite element analysis demands a great deal of computational resources, and as such is most appropriately solved in a massively parallel computer environment. This analysis will require other parallel algorithms before it can fully utilize MP computers, one of which is parallel adaptive meshing. A version of the paving algorithm is being designed which operates in parallel but which also retains the robustness and other desirable features present in the serial algorithm. Adaptive paving in a production mode is demonstrated using a Babuska-Rheinboldt error estimator on a classic linearly elastic plate problem. The design of the parallel paving algorithm is described, and is based on the decomposition of a surface into {open_quotes}virtual{close_quotes} surfaces. The topology of the virtual surface boundaries is defined using mesh entities (mesh nodes and edges) so as to allow movement of these boundaries with smoothing and other operations. This arrangement allows the use of the standard paving algorithm on subdomain interiors, after the negotiation of the boundary mesh.
PLUM: Parallel Load Balancing for Unstructured Adaptive Meshes. Degree awarded by Colorado Univ.
NASA Technical Reports Server (NTRS)
Oliker, Leonid
1998-01-01
Dynamic mesh adaption on unstructured grids is a powerful tool for computing large-scale problems that require grid modifications to efficiently resolve solution features. By locally refining and coarsening the mesh to capture physical phenomena of interest, such procedures make standard computational methods more cost effective. Unfortunately, an efficient parallel implementation of these adaptive methods is rather difficult to achieve, primarily due to the load imbalance created by the dynamically-changing nonuniform grid. This requires significant communication at runtime, leading to idle processors and adversely affecting the total execution time. Nonetheless, it is generally thought that unstructured adaptive- grid techniques will constitute a significant fraction of future high-performance supercomputing. Various dynamic load balancing methods have been reported to date; however, most of them either lack a global view of loads across processors or do not apply their techniques to realistic large-scale applications.
An adaptive embedded mesh procedure for leading-edge vortex flows
NASA Technical Reports Server (NTRS)
Powell, Kenneth G.; Beer, Michael A.; Law, Glenn W.
1989-01-01
A procedure for solving the conical Euler equations on an adaptively refined mesh is presented, along with a method for determining which cells to refine. The solution procedure is a central-difference cell-vertex scheme. The adaptation procedure is made up of a parameter on which the refinement decision is based, and a method for choosing a threshold value of the parameter. The refinement parameter is a measure of mesh-convergence, constructed by comparison of locally coarse- and fine-grid solutions. The threshold for the refinement parameter is based on the curvature of the curve relating the number of cells flagged for refinement to the value of the refinement threshold. Results for three test cases are presented. The test problem is that of a delta wing at angle of attack in a supersonic free-stream. The resulting vortices and shocks are captured efficiently by the adaptive code.
Spatial adaptation procedures on tetrahedral meshes for unsteady aerodynamic flow calculations
NASA Technical Reports Server (NTRS)
Rausch, Russ D.; Batina, John T.; Yang, Henry T. Y.
1993-01-01
Spatial adaptation procedures for the accurate and efficient solution of steady and unsteady inviscid flow problems are described. The adaptation procedures were developed and implemented within a three-dimensional, unstructured-grid, upwind-type Euler code. These procedures involve mesh enrichment and mesh coarsening to either add points in high gradient regions of the flow or remove points where they are not needed, respectively, to produce solutions of high spatial accuracy at minimal computational cost. A detailed description of the enrichment and coarsening procedures are presented and comparisons with experimental data for an ONERA M6 wing and an exact solution for a shock-tube problem are presented to provide an assessment of the accuracy and efficiency of the capability. Steady and unsteady results, obtained using spatial adaptation procedures, are shown to be of high spatial accuracy, primarily in that discontinuities such as shock waves are captured very sharply.
Kolobov, Vladimir; Arslanbekov, Robert; Frolova, Anna
2014-12-09
The paper describes an Adaptive Mesh in Phase Space (AMPS) technique for solving kinetic equations with deterministic mesh-based methods. The AMPS technique allows automatic generation of adaptive Cartesian mesh in both physical and velocity spaces using a Tree-of-Trees data structure. We illustrate advantages of AMPS for simulations of rarefied gas dynamics and electron kinetics on low temperature plasmas. In particular, we consider formation of the velocity distribution functions in hypersonic flows, particle kinetics near oscillating boundaries, and electron kinetics in a radio-frequency sheath. AMPS provide substantial savings in computational cost and increased efficiency of the mesh-based kinetic solvers.
NASA Astrophysics Data System (ADS)
Kolobov, Vladimir; Arslanbekov, Robert; Frolova, Anna
2014-12-01
The paper describes an Adaptive Mesh in Phase Space (AMPS) technique for solving kinetic equations with deterministic mesh-based methods. The AMPS technique allows automatic generation of adaptive Cartesian mesh in both physical and velocity spaces using a Tree-of-Trees data structure. We illustrate advantages of AMPS for simulations of rarefied gas dynamics and electron kinetics on low temperature plasmas. In particular, we consider formation of the velocity distribution functions in hypersonic flows, particle kinetics near oscillating boundaries, and electron kinetics in a radio-frequency sheath. AMPS provide substantial savings in computational cost and increased efficiency of the mesh-based kinetic solvers.
NASA Astrophysics Data System (ADS)
Hatori, Tomoharu; Ito, Atsushi M.; Nunami, Masanori; Usui, Hideyuki; Miura, Hideaki
2016-08-01
We propose a numerical method to determine the artificial viscosity in magnetohydrodynamics (MHD) simulations with adaptive mesh refinement (AMR) method, where the artificial viscosity is adaptively changed due to the resolution level of the AMR hierarchy. Although the suitable value of the artificial viscosity depends on the governing equations and the model of target problem, it can be determined by von Neumann stability analysis. By means of the new method, "level-by-level artificial viscosity method," MHD simulations of Rayleigh-Taylor instability (RTI) are carried out with the AMR method. The validity of the level-by-level artificial viscosity method is confirmed by the comparison of the linear growth rates of RTI between the AMR simulations and the simple simulations with uniform grid and uniform artificial viscosity whose resolution is the same as that in the highest level of the AMR simulation. Moreover, in the nonlinear phase of RTI, the secondary instability is clearly observed where the hierarchical data structure of AMR calculation is visualized as high resolution region floats up like terraced fields. In the applications of the method to general fluid simulations, the growth of small structures can be sufficiently reproduced, while the divergence of numerical solutions can be suppressed.
Practical improvements of multi-grid iteration for adaptive mesh refinement method
NASA Astrophysics Data System (ADS)
Miyashita, Hisashi; Yamada, Yoshiyuki
2005-03-01
Adaptive mesh refinement(AMR) is a powerful tool to efficiently solve multi-scaled problems. However, the vanilla AMR method has a well-known critical demerit, i.e., it cannot be applied to non-local problems. Although multi-grid iteration (MGI) can be regarded as a good remedy for a non-local problem such as the Poisson equation, we observed fundamental difficulties in applying the MGI technique in AMR to realistic problems under complicated mesh layouts because it does not converge or it requires too many iterations even if it does converge. To cope with the problem, when updating the next approximation in the MGI process, we calculate the precise total corrections that are relatively accurate to the current residual by introducing a new iteration for such a total correction. This procedure greatly accelerates the MGI convergence speed especially under complicated mesh layouts.
Parallel Implementation and Scaling of an Adaptive Mesh Discrete Ordinates Algorithm for Transport
Howell, L H
2004-11-29
Block-structured adaptive mesh refinement (AMR) uses a mesh structure built up out of locally-uniform rectangular grids. In the BoxLib parallel framework used by the Raptor code, each processor operates on one or more of these grids at each refinement level. The decomposition of the mesh into grids and the distribution of these grids among processors may change every few timesteps as a calculation proceeds. Finer grids use smaller timesteps than coarser grids, requiring additional work to keep the system synchronized and ensure conservation between different refinement levels. In a paper for NECDC 2002 I presented preliminary results on implementation of parallel transport sweeps on the AMR mesh, conjugate gradient acceleration, accuracy of the AMR solution, and scalar speedup of the AMR algorithm compared to a uniform fully-refined mesh. This paper continues with a more in-depth examination of the parallel scaling properties of the scheme, both in single-level and multi-level calculations. Both sweeping and setup costs are considered. The algorithm scales with acceptable performance to several hundred processors. Trends suggest, however, that this is the limit for efficient calculations with traditional transport sweeps, and that modifications to the sweep algorithm will be increasingly needed as job sizes in the thousands of processors become common.
NASA Astrophysics Data System (ADS)
Pantano, Carlos
2005-11-01
We describe a hybrid finite difference method for large-eddy simulation (LES) of compressible flows with a low-numerical dissipation scheme and structured adaptive mesh refinement (SAMR). Numerical experiments and validation calculations are presented including a turbulent jet and the strongly shock-driven mixing of a Richtmyer-Meshkov instability. The approach is a conservative flux-based SAMR formulation and as such, it utilizes refinement to computational advantage. The numerical method for the resolved scale terms encompasses the cases of scheme alternation and internal mesh interfaces resulting from SAMR. An explicit centered scheme that is consistent with a skew-symmetric finite difference formulation is used in turbulent flow regions while a weighted essentially non-oscillatory (WENO) scheme is employed to capture shocks. The subgrid stresses and transports are calculated by means of the streched-vortex model, Misra & Pullin (1997)
NASA Astrophysics Data System (ADS)
Calder, A. C.; Curtis, B. C.; Dursi, L. J.; Fryxell, B.; Henry, G.; MacNeice, P.; Olson, K.; Ricker, P.; Rosner, R.; Timmes, F. X.; Tufo, H. M.; Truran, J. W.; Zingale, M.
We present simulations and performance results of nuclear burning fronts in supernovae on the largest domain and at the finest spatial resolution studied to date. These simulations were performed on the Intel ASCI-Red machine at Sandia National Laboratories using FLASH, a code developed at the Center for Astrophysical Thermonuclear Flashes at the University of Chicago. FLASH is a modular, adaptive mesh, parallel simulation code capable of handling compressible, reactive fluid flows in astrophysical environments. FLASH is written primarily in Fortran 90, uses the Message-Passing Interface library for inter-processor communication and portability, and employs the PARAMESH package to manage a block-structured adaptive mesh that places blocks only where the resolution is required and tracks rapidly changing flow features, such as detonation fronts, with ease. We describe the key algorithms and their implementation as well as the optimizations required to achieve sustained performance of 238 GLOPS on 6420 processors of ASCI-Red in 64-bit arithmetic.
Implementation of Implicit Adaptive Mesh Refinement in an Unstructured Finite-Volume Flow Solver
NASA Technical Reports Server (NTRS)
Schwing, Alan M.; Nompelis, Ioannis; Candler, Graham V.
2013-01-01
This paper explores the implementation of adaptive mesh refinement in an unstructured, finite-volume solver. Unsteady and steady problems are considered. The effect on the recovery of high-order numerics is explored and the results are favorable. Important to this work is the ability to provide a path for efficient, implicit time advancement. A method using a simple refinement sensor based on undivided differences is discussed and applied to a practical problem: a shock-shock interaction on a hypersonic, inviscid double-wedge. Cases are compared to uniform grids without the use of adapted meshes in order to assess error and computational expense. Discussion of difficulties, advances, and future work prepare this method for additional research. The potential for this method in more complicated flows is described.
NASA Astrophysics Data System (ADS)
Adam, A.; Pavlidis, D.; Percival, J. R.; Salinas, P.; Xie, Z.; Fang, F.; Pain, C. C.; Muggeridge, A. H.; Jackson, M. D.
2016-09-01
A general, higher-order, conservative and bounded interpolation for the dynamic and adaptive meshing of control-volume fields dual to continuous and discontinuous finite element representations is presented. Existing techniques such as node-wise interpolation are not conservative and do not readily generalise to discontinuous fields, whilst conservative methods such as Grandy interpolation are often too diffusive. The new method uses control-volume Galerkin projection to interpolate between control-volume fields. Bounded solutions are ensured by using a post-interpolation diffusive correction. Example applications of the method to interface capturing during advection and also to the modelling of multiphase porous media flow are presented to demonstrate the generality and robustness of the approach.
Turner, C. David; Kotulski, Joseph Daniel; Pasik, Michael Francis
2005-12-01
This report investigates the feasibility of applying Adaptive Mesh Refinement (AMR) techniques to a vector finite element formulation for the wave equation in three dimensions. Possible error estimators are considered first. Next, approaches for refining tetrahedral elements are reviewed. AMR capabilities within the Nevada framework are then evaluated. We summarize our conclusions on the feasibility of AMR for time-domain vector finite elements and identify a path forward.
A 3-D adaptive mesh refinement algorithm for multimaterial gas dynamics
Puckett, E.G. ); Saltzman, J.S. )
1991-08-12
Adaptive Mesh Refinement (AMR) in conjunction with high order upwind finite difference methods has been used effectively on a variety of problems. In this paper we discuss an implementation of an AMR finite difference method that solves the equations of gas dynamics with two material species in three dimensions. An equation for the evolution of volume fractions augments the gas dynamics system. The material interface is preserved and tracked from the volume fractions using a piecewise linear reconstruction technique. 14 refs., 4 figs.
NASA Astrophysics Data System (ADS)
Wilson, C. R.; Kramer, S. C.; Collins, G. S.
2010-12-01
Linear wave models cannot reproduce the highly nonlinear generation mechanisms required to accurately predict the consequences of landslide-generated tsunamis. Models based on the nonlinear Navier-Stokes equations can simulate complex landslide-water interactions at realistic scales; however, the computing power required for such a simulation can be prohibitively high for large domains with realistic bathymetries. The variable resolution available with the use of unstructured adaptive meshes allows larger domains to be modeled at the same resolution for a lower computational cost than on structured meshes; they are also better at representing complex geometries and bathymetries. However, unstructured meshes introduce extra numerical challenges requiring the use of novel interface preservation techniques coupled with velocity-pressure discretisations that ensure the conservation and boundedness of all materials in the simulation. In this study we describe some of the challenges encountered extending the finite element, finite volume multiple-material fluid dynamics model Fluidity to large-scale landslide-generated tsunami simulations. In particular, we focus on the ability of the model to preserve the balance between the buoyancy and pressure gradient forces. Failure to discretely satisfy this relationship is shown to result in spurious waves that contaminate any physical tsunami signal. However, ensuring that balance is preserved in a computationally efficient manner imposes extra constraints on the dynamic mesh optimisation process. Incorporating these restrictions allows us to validate our model against multi-scale experimental simulations of landslide generated tsunami (see figure). Experimental (top, taken from Di Risio et. al. 2009, doi:10.1029/2008JC004858) and equivalent numerical simulation (bottom) of a subaerial landslide impacting into water. In the experiment the 80cm long landslide produces waves of amplitude 1-2cm around a 9m diameter island in a 50x
RCS and antenna modeling with MOM using hybrid meshes
Putnam, J.M.; Kotulski, J.D.
1997-02-01
In this presentation, the authors will investigate the use of hybrid meshes for modeling RCS and antenna problems in three dimensions. They will consider two classes of hybrid basis functions. These include combinations of quadrilateral and triangular meshes for arbitrary 3D geometries, and combinations of axisymmetric body-of-revolution (BOR) basis functions and triangular facets. In particular, they will focus on the problem of enforcing current continuity between two surfaces which are represented by different types of surface discretizations and unknown basis function representations. They will illustrate the use of an operator-based code architecture for the implementation of these formulations, and how it facilitates the incorporation of the various types of boundary conditions in the code. Both serial and parallel code implementation issues for the formulations will be discussed. Results will be presented for both scattering and antenna problems. The emphasis will be on accuracy, and robustness of the techniques. Comparisons of accuracy between triangular meshed and quadrilateral meshed geometries will be shown. The use of hybrid meshes for modeling BORs with attached appendages will also be presented.
NASA Astrophysics Data System (ADS)
Teyssier, Romain; Fromang, Sébastien; Dormy, Emmanuel
2006-10-01
We propose to extend the well-known MUSCL-Hancock scheme for Euler equations to the induction equation modeling the magnetic field evolution in kinematic dynamo problems. The scheme is based on an integral form of the underlying conservation law which, in our formulation, results in a “finite-surface” scheme for the induction equation. This naturally leads to the well-known “constrained transport” method, with additional continuity requirement on the magnetic field representation. The second ingredient in the MUSCL scheme is the predictor step that ensures second order accuracy both in space and time. We explore specific constraints that the mathematical properties of the induction equations place on this predictor step, showing that three possible variants can be considered. We show that the most aggressive formulations (referred to as C-MUSCL and U-MUSCL) reach the same level of accuracy as the other one (referred to as Runge Kutta), at a lower computational cost. More interestingly, these two schemes are compatible with the adaptive mesh refinement (AMR) framework. It has been implemented in the AMR code RAMSES. It offers a novel and efficient implementation of a second order scheme for the induction equation. We have tested it by solving two kinematic dynamo problems in the low diffusion limit. The construction of this scheme for the induction equation constitutes a step towards solving the full MHD set of equations using an extension of our current methodology.
4D cone-beam CT reconstruction using multi-organ meshes for sliding motion modeling.
Zhong, Zichun; Gu, Xuejun; Mao, Weihua; Wang, Jing
2016-02-01
A simultaneous motion estimation and image reconstruction (SMEIR) strategy was proposed for 4D cone-beam CT (4D-CBCT) reconstruction and showed excellent results in both phantom and lung cancer patient studies. In the original SMEIR algorithm, the deformation vector field (DVF) was defined on voxel grid and estimated by enforcing a global smoothness regularization term on the motion fields. The objective of this work is to improve the computation efficiency and motion estimation accuracy of SMEIR for 4D-CBCT through developing a multi-organ meshing model. Feature-based adaptive meshes were generated to reduce the number of unknowns in the DVF estimation and accurately capture the organ shapes and motion. Additionally, the discontinuity in the motion fields between different organs during respiration was explicitly considered in the multi-organ mesh model. This will help with the accurate visualization and motion estimation of the tumor on the organ boundaries in 4D-CBCT. To further improve the computational efficiency, a GPU-based parallel implementation was designed. The performance of the proposed algorithm was evaluated on a synthetic sliding motion phantom, a 4D NCAT phantom, and four lung cancer patients. The proposed multi-organ mesh based strategy outperformed the conventional Feldkamp-Davis-Kress, iterative total variation minimization, original SMEIR and single meshing method based on both qualitative and quantitative evaluations. PMID:26758496
4D cone-beam CT reconstruction using multi-organ meshes for sliding motion modeling
NASA Astrophysics Data System (ADS)
Zhong, Zichun; Gu, Xuejun; Mao, Weihua; Wang, Jing
2016-02-01
A simultaneous motion estimation and image reconstruction (SMEIR) strategy was proposed for 4D cone-beam CT (4D-CBCT) reconstruction and showed excellent results in both phantom and lung cancer patient studies. In the original SMEIR algorithm, the deformation vector field (DVF) was defined on voxel grid and estimated by enforcing a global smoothness regularization term on the motion fields. The objective of this work is to improve the computation efficiency and motion estimation accuracy of SMEIR for 4D-CBCT through developing a multi-organ meshing model. Feature-based adaptive meshes were generated to reduce the number of unknowns in the DVF estimation and accurately capture the organ shapes and motion. Additionally, the discontinuity in the motion fields between different organs during respiration was explicitly considered in the multi-organ mesh model. This will help with the accurate visualization and motion estimation of the tumor on the organ boundaries in 4D-CBCT. To further improve the computational efficiency, a GPU-based parallel implementation was designed. The performance of the proposed algorithm was evaluated on a synthetic sliding motion phantom, a 4D NCAT phantom, and four lung cancer patients. The proposed multi-organ mesh based strategy outperformed the conventional Feldkamp-Davis-Kress, iterative total variation minimization, original SMEIR and single meshing method based on both qualitative and quantitative evaluations.
Patched based methods for adaptive mesh refinement solutions of partial differential equations
Saltzman, J.
1997-09-02
This manuscript contains the lecture notes for a course taught from July 7th through July 11th at the 1997 Numerical Analysis Summer School sponsored by C.E.A., I.N.R.I.A., and E.D.F. The subject area was chosen to support the general theme of that year`s school which is ``Multiscale Methods and Wavelets in Numerical Simulation.`` The first topic covered in these notes is a description of the problem domain. This coverage is limited to classical PDEs with a heavier emphasis on hyperbolic systems and constrained hyperbolic systems. The next topic is difference schemes. These schemes are the foundation for the adaptive methods. After the background material is covered, attention is focused on a simple patched based adaptive algorithm and its associated data structures for square grids and hyperbolic conservation laws. Embellishments include curvilinear meshes, embedded boundary and overset meshes. Next, several strategies for parallel implementations are examined. The remainder of the notes contains descriptions of elliptic solutions on the mesh hierarchy, elliptically constrained flow solution methods and elliptically constrained flow solution methods with diffusion.
Error estimation and adaptive mesh refinement for parallel analysis of shell structures
NASA Technical Reports Server (NTRS)
Keating, Scott C.; Felippa, Carlos A.; Park, K. C.
1994-01-01
The formulation and application of element-level, element-independent error indicators is investigated. This research culminates in the development of an error indicator formulation which is derived based on the projection of element deformation onto the intrinsic element displacement modes. The qualifier 'element-level' means that no information from adjacent elements is used for error estimation. This property is ideally suited for obtaining error values and driving adaptive mesh refinements on parallel computers where access to neighboring elements residing on different processors may incur significant overhead. In addition such estimators are insensitive to the presence of physical interfaces and junctures. An error indicator qualifies as 'element-independent' when only visible quantities such as element stiffness and nodal displacements are used to quantify error. Error evaluation at the element level and element independence for the error indicator are highly desired properties for computing error in production-level finite element codes. Four element-level error indicators have been constructed. Two of the indicators are based on variational formulation of the element stiffness and are element-dependent. Their derivations are retained for developmental purposes. The second two indicators mimic and exceed the first two in performance but require no special formulation of the element stiffness mesh refinement which we demonstrate for two dimensional plane stress problems. The parallelizing of substructures and adaptive mesh refinement is discussed and the final error indicator using two-dimensional plane-stress and three-dimensional shell problems is demonstrated.
Delanaye, M.; Essers, J.A.
1997-04-01
This paper presents a new finite volume cell-centered scheme for solving the two-dimensional Euler equations. The technique for computing the advective derivatives is based on a high-order Gauss quadrature and an original quadratic reconstruction of the conservative variables for each control volume. A very sensitive detector identifying discontinuity regions switches the scheme to a TVD scheme, and ensures the monotonicity of the solution. The code uses unstructured meshes whose cells are polygons with any number of edges. A mesh adaptation based on cell division is performed in order to increase the resolution of shocks. The accuracy, insensitivity to grid distortions, and shock capturing properties of the scheme are demonstrated for different cascade flow computations.
NASA Astrophysics Data System (ADS)
Ghattas, O.; Burstedde, C.; Stadler, G.; Wilcox, L. C.; Tu, T.; Issac, T.; Gurnis, M.; Alisic, L.; Tan, E.; Zhong, S.
2009-12-01
Many problems in solid earth geophysics are characterized by dynamics occurring on a wide range of length and time scales, placing the solution of the governing partial differential equations (PDEs) for such problems among the grand challenges of computational geophysics. One approach to overcoming the tyranny of scales is adaptive mesh refinement (AMR), which locally and dynamically adapts the mesh to resolve spatio-temporal scales and features of interest. For example, we are interested in modeling global mantle convection with nonlinear rheology and kilometer-scale resolution at faulted plate boundaries. Another problem of interest is modeling the dynamics of polar ice sheets with fine resolution in the vicinity of stick-slip transitions. Geophysical inverse problems characterized by a wide range of medium properties can also benefit from AMR as the earth model is updated. While AMR promises to help overcome the challenges inherent in modeling multiscale problems, the benefits are difficult to achieve in practice, particularly on petascale computers that are essential for frontier problems. Due to the complex dynamic data structures and communication patterns, and frequent data exchange and redistribution, scaling dynamic AMR to tens of thousands of processors has long been considered a challenge. Another difficulty is extending parallel AMR techniques to high-order-accurate, complex-geometry-conforming finite element methods that are favored for many classes of solid earth geophysical problems. Here, we present the ALPS (Adaptive Large-scale Parallel Simulations) framework for parallel adaptive solution of PDEs. ALPS includes the octor and p4est libraries for parallel dynamic mesh adaptivity on single-octree-based and forest-of-octree-based geometries, respectively, and the mangll library for arbitrary-order hexahedral continuous and discontinuous finite/spectral element discretizations on general multi-octree geometries. ALPS has been shown to scale well
Mesh type tradeoffs in 2D hydrodynamic modeling of flooding with a Godunov-based flow solver
NASA Astrophysics Data System (ADS)
Kim, Byunghyun; Sanders, Brett F.; Schubert, Jochen E.; Famiglietti, James S.
2014-06-01
The effect of mesh type on the accuracy and computational demands of a two-dimensional Godunov-type flood inundation model is critically examined. Cartesian grids, constrained and unconstrained triangular grids, constrained quadrilateral grids, and mixed meshes are considered, with and without local time stepping (LTS), to determine the approach that maximizes computational efficiency defined as accuracy relative to computational effort. A mixed-mesh numerical scheme is introduced so all grids are processed by the same solver. Analysis focuses on a wide range of dam-break type test cases, where Godunov-type flood models have proven very successful. Results show that different mesh types excel under different circumstances. Cartesian grids are 2-3 times more efficient with relatively simple terrain features such as rectilinear channels that call for a uniform grid resolution, while unstructured grids are about twice as efficient in complex domains with irregular terrain features that call for localized refinements. The superior efficiency of locally refined, unstructured grids in complex terrain is attributable to LTS; the locally refined unstructured grid becomes less efficient using global time stepping. These results point to mesh-type tradeoffs that should be considered in flood modeling applications. A mixed mesh model formulation with LTS is recommended as a general purpose solver because the mesh type can be adapted to maximize computational efficiency.
NASA Astrophysics Data System (ADS)
Shi, Lei; Wang, Z. J.
2015-08-01
Adjoint-based mesh adaptive methods are capable of distributing computational resources to areas which are important for predicting an engineering output. In this paper, we develop an adjoint-based h-adaptation approach based on the high-order correction procedure via reconstruction formulation (CPR) to minimize the output or functional error. A dual-consistent CPR formulation of hyperbolic conservation laws is developed and its dual consistency is analyzed. Super-convergent functional and error estimate for the output with the CPR method are obtained. Factors affecting the dual consistency, such as the solution point distribution, correction functions, boundary conditions and the discretization approach for the non-linear flux divergence term, are studied. The presented method is then used to perform simulations for the 2D Euler and Navier-Stokes equations with mesh adaptation driven by the adjoint-based error estimate. Several numerical examples demonstrate the ability of the presented method to dramatically reduce the computational cost comparing with uniform grid refinement.
A new adaptive mesh refinement data structure with an application to detonation
NASA Astrophysics Data System (ADS)
Ji, Hua; Lien, Fue-Sang; Yee, Eugene
2010-11-01
A new Cell-based Structured Adaptive Mesh Refinement (CSAMR) data structure is developed. In our CSAMR data structure, Cartesian-like indices are used to identify each cell. With these stored indices, the information on the parent, children and neighbors of a given cell can be accessed simply and efficiently. Owing to the usage of these indices, the computer memory required for storage of the proposed AMR data structure is only {5}/{8} word per cell, in contrast to the conventional oct-tree [P. MacNeice, K.M. Olson, C. Mobary, R. deFainchtein, C. Packer, PARAMESH: a parallel adaptive mesh refinement community toolkit, Comput. Phys. Commun. 330 (2000) 126] and the fully threaded tree (FTT) [A.M. Khokhlov, Fully threaded tree algorithms for adaptive mesh fluid dynamics simulations, J. Comput. Phys. 143 (1998) 519] data structures which require, respectively, 19 and 2{3}/{8} words per cell for storage of the connectivity information. Because the connectivity information (e.g., parent, children and neighbors) of a cell in our proposed AMR data structure can be accessed using only the cell indices, a tree structure which was required in previous approaches for the organization of the AMR data is no longer needed for this new data structure. Instead, a much simpler hash table structure is used to maintain the AMR data, with the entry keys in the hash table obtained directly from the explicitly stored cell indices. The proposed AMR data structure simplifies the implementation and parallelization of an AMR code. Two three-dimensional test cases are used to illustrate and evaluate the computational performance of the new CSAMR data structure.
An adaptive mesh method for phase-field simulation of alloy solidification in three dimensions
NASA Astrophysics Data System (ADS)
Bollada, P. C.; Jimack, P. K.; Mullis, A. M.
2015-06-01
We present our computational method for binary alloy solidification which takes advantage of high performance computing where up to 1024 cores are employed. Much of the simulation at a sufficiently fine resolution is possible on a modern 12 core PC and the 1024 core simulation is only necessary for very mature dendrite and convergence testing where high resolution puts extreme demands on memory. In outline, the method uses implicit time stepping in conjunction with an iterative solver, adaptive meshing and a scheme for dividing the work load across processors. We include three dimensional results for a Lewis number of 100 and a snapshot for a mature dendrite for a Lewis number of 40.
Development of a Godunov method for Maxwell's equations with Adaptive Mesh Refinement
NASA Astrophysics Data System (ADS)
Barbas, Alfonso; Velarde, Pedro
2015-11-01
In this paper we present a second order 3D method for Maxwell's equations based on a Godunov scheme with Adaptive Mesh Refinement (AMR). In order to achieve it, we apply a limiter which better preserves extrema and boundary conditions based on a characteristic fields decomposition. Despite being more complex, simplifications in the boundary conditions make the resulting method competitive in computer time consumption and accuracy compared to FDTD. AMR allows us to simulate systems with a sharp step in material properties with negligible rebounds and also large domains with accuracy in small wavelengths.
Compact integration factor methods for complex domains and adaptive mesh refinement
Liu, Xinfeng; Nie, Qing
2010-01-01
Implicit integration factor (IIF) method, a class of efficient semi-implicit temporal scheme, was introduced recently for stiff reaction-diffusion equations. To reduce cost of IIF, compact implicit integration factor (cIIF) method was later developed for efficient storage and calculation of exponential matrices associated with the diffusion operators in two and three spatial dimensions for Cartesian coordinates with regular meshes. Unlike IIF, cIIF cannot be directly extended to other curvilinear coordinates, such as polar and spherical coordinate, due to the compact representation for the diffusion terms in cIIF. In this paper, we present a method to generalize cIIF for other curvilinear coordinates through examples of polar and spherical coordinates. The new cIIF method in polar and spherical coordinates has similar computational efficiency and stability properties as the cIIF in Cartesian coordinate. In addition, we present a method for integrating cIIF with adaptive mesh refinement (AMR) to take advantage of the excellent stability condition for cIIF. Because the second order cIIF is unconditionally stable, it allows large time steps for AMR, unlike a typical explicit temporal scheme whose time step is severely restricted by the smallest mesh size in the entire spatial domain. Finally, we apply those methods to simulating a cell signaling system described by a system of stiff reaction-diffusion equations in both two and three spatial dimensions using AMR, curvilinear and Cartesian coordinates. Excellent performance of the new methods is observed. PMID:20543883
Biomechanical analyses of prosthetic mesh repair in a hiatal hernia model.
Alizai, Patrick Hamid; Schmid, Sofie; Otto, Jens; Klink, Christian Daniel; Roeth, Anjali; Nolting, Jochen; Neumann, Ulf Peter; Klinge, Uwe
2014-10-01
Recurrence rate of hiatal hernia can be reduced with prosthetic mesh repair; however, type and shape of the mesh are still a matter of controversy. The purpose of this study was to investigate the biomechanical properties of four conventional meshes: pure polypropylene mesh (PP-P), polypropylene/poliglecaprone mesh (PP-U), polyvinylidenefluoride/polypropylene mesh (PVDF-I), and pure polyvinylidenefluoride mesh (PVDF-S). Meshes were tested either in warp direction (parallel to production direction) or perpendicular to the warp direction. A Zwick testing machine was used to measure elasticity and effective porosity of the textile probes. Stretching of the meshes in warp direction required forces that were up to 85-fold higher than the same elongation in perpendicular direction. Stretch stress led to loss of effective porosity in most meshes, except for PVDF-S. Biomechanical impact of the mesh was additionally evaluated in a hiatal hernia model. The different meshes were used either as rectangular patches or as circular meshes. Circular meshes led to a significant reinforcement of the hiatus, largely unaffected by the orientation of the warp fibers. In contrast, rectangular meshes provided a significant reinforcement only when warp fibers ran perpendicular to the crura. Anisotropic elasticity of prosthetic meshes should therefore be considered in hiatal closure with rectangular patches. PMID:24599834
Earthquake Rupture Dynamics using Adaptive Mesh Refinement and High-Order Accurate Numerical Methods
NASA Astrophysics Data System (ADS)
Kozdon, J. E.; Wilcox, L.
2013-12-01
Our goal is to develop scalable and adaptive (spatial and temporal) numerical methods for coupled, multiphysics problems using high-order accurate numerical methods. To do so, we are developing an opensource, parallel library known as bfam (available at http://bfam.in). The first application to be developed on top of bfam is an earthquake rupture dynamics solver using high-order discontinuous Galerkin methods and summation-by-parts finite difference methods. In earthquake rupture dynamics, wave propagation in the Earth's crust is coupled to frictional sliding on fault interfaces. This coupling is two-way, required the simultaneous simulation of both processes. The use of laboratory-measured friction parameters requires near-fault resolution that is 4-5 orders of magnitude higher than that needed to resolve the frequencies of interest in the volume. This, along with earlier simulations using a low-order, finite volume based adaptive mesh refinement framework, suggest that adaptive mesh refinement is ideally suited for this problem. The use of high-order methods is motivated by the high level of resolution required off the fault in earlier the low-order finite volume simulations; we believe this need for resolution is a result of the excessive numerical dissipation of low-order methods. In bfam spatial adaptivity is handled using the p4est library and temporal adaptivity will be accomplished through local time stepping. In this presentation we will present the guiding principles behind the library as well as verification of code against the Southern California Earthquake Center dynamic rupture code validation test problems.
NASA Astrophysics Data System (ADS)
Bajc, Iztok; Hecht, Frédéric; Žumer, Slobodan
2016-09-01
This paper presents a 3D mesh adaptivity strategy on unstructured tetrahedral meshes by a posteriori error estimates based on metrics derived from the Hessian of a solution. The study is made on the case of a nonlinear finite element minimization scheme for the Landau-de Gennes free energy functional of nematic liquid crystals. Newton's iteration for tensor fields is employed with steepest descent method possibly stepping in. Aspects relating the driving of mesh adaptivity within the nonlinear scheme are considered. The algorithmic performance is found to depend on at least two factors: when to trigger each single mesh adaptation, and the precision of the correlated remeshing. Each factor is represented by a parameter, with its values possibly varying for every new mesh adaptation. We empirically show that the time of the overall algorithm convergence can vary considerably when different sequences of parameters are used, thus posing a question about optimality. The extensive testings and debugging done within this work on the simulation of systems of nematic colloids substantially contributed to the upgrade of an open source finite element-oriented programming language to its 3D meshing possibilities, as also to an outer 3D remeshing module.
NASA Astrophysics Data System (ADS)
De Colle, Fabio; Granot, Jonathan; López-Cámara, Diego; Ramirez-Ruiz, Enrico
2012-02-01
We report on the development of Mezcal-SRHD, a new adaptive mesh refinement, special relativistic hydrodynamics (SRHD) code, developed with the aim of studying the highly relativistic flows in gamma-ray burst sources. The SRHD equations are solved using finite-volume conservative solvers, with second-order interpolation in space and time. The correct implementation of the algorithms is verified by one-dimensional (1D) and multi-dimensional tests. The code is then applied to study the propagation of 1D spherical impulsive blast waves expanding in a stratified medium with ρvpropr -k , bridging between the relativistic and Newtonian phases (which are described by the Blandford-McKee and Sedov-Taylor self-similar solutions, respectively), as well as to a two-dimensional (2D) cylindrically symmetric impulsive jet propagating in a constant density medium. It is shown that the deceleration to nonrelativistic speeds in one dimension occurs on scales significantly larger than the Sedov length. This transition is further delayed with respect to the Sedov length as the degree of stratification of the ambient medium is increased. This result, together with the scaling of position, Lorentz factor, and the shock velocity as a function of time and shock radius, is explained here using a simple analytical model based on energy conservation. The method used for calculating the afterglow radiation by post-processing the results of the simulations is described in detail. The light curves computed using the results of 1D numerical simulations during the relativistic stage correctly reproduce those calculated assuming the self-similar Blandford-McKee solution for the evolution of the flow. The jet dynamics from our 2D simulations and the resulting afterglow light curves, including the jet break, are in good agreement with those presented in previous works. Finally, we show how the details of the dynamics critically depend on properly resolving the structure of the relativistic flow.
NASA Astrophysics Data System (ADS)
Grosges, T.; Borouchaki, H.; Barchiesi, D.
2010-12-01
We present an improved adaptive mesh process based on Riemannian transformation to control the accuracy in high field gradient representation for diffraction problems. Such an adaptive meshing is applied in representing the electromagnetic intensity around a metallic submicronic spherical particle, which is known to present high gradients in limited zones of space including the interference pattern of the electromagnetic field. We show that, the precision of the field variation being controlled, this improved scheme permits drastically decreasing the computational time as well as the memory requirements by adapting the number and the position of nodes where the electromagnetic field must be computed and represented.
Ragusa, Jean C.
2015-01-01
In this paper, we propose a piece-wise linear discontinuous (PWLD) finite element discretization of the diffusion equation for arbitrary polygonal meshes. It is based on the standard diffusion form and uses the symmetric interior penalty technique, which yields a symmetric positive definite linear system matrix. A preconditioned conjugate gradient algorithm is employed to solve the linear system. Piece-wise linear approximations also allow a straightforward implementation of local mesh adaptation by allowing unrefined cells to be interpreted as polygons with an increased number of vertices. Several test cases, taken from the literature on the discretization of the radiation diffusion equation, are presented: random, sinusoidal, Shestakov, and Z meshes are used. The last numerical example demonstrates the application of the PWLD discretization to adaptive mesh refinement.
NASA Astrophysics Data System (ADS)
Foks, Nathan Leon
The interpretation of geophysical data plays an important role in the analysis of potential field data in resource exploration industries. Two categories of interpretation techniques are discussed in this thesis; boundary detection and geophysical inversion. Fault or boundary detection is a method to interpret the locations of subsurface boundaries from measured data, while inversion is a computationally intensive method that provides 3D information about subsurface structure. My research focuses on these two aspects of interpretation techniques. First, I develop a method to aid in the interpretation of faults and boundaries from magnetic data. These processes are traditionally carried out using raster grid and image processing techniques. Instead, I use unstructured meshes of triangular facets that can extract inferred boundaries using mesh edges. Next, to address the computational issues of geophysical inversion, I develop an approach to reduce the number of data in a data set. The approach selects the data points according to a user specified proxy for its signal content. The approach is performed in the data domain and requires no modification to existing inversion codes. This technique adds to the existing suite of compressive inversion algorithms. Finally, I develop an algorithm to invert gravity data for an interfacing surface using an unstructured mesh of triangular facets. A pertinent property of unstructured meshes is their flexibility at representing oblique, or arbitrarily oriented structures. This flexibility makes unstructured meshes an ideal candidate for geometry based interface inversions. The approaches I have developed provide a suite of algorithms geared towards large-scale interpretation of potential field data, by using an unstructured representation of both the data and model parameters.
Efficient Modelling and Prediction of Meshing Noise from Chain Drives
NASA Astrophysics Data System (ADS)
ZHENG, H.; WANG, Y. Y.; LIU, G. R.; LAM, K. Y.; QUEK, K. P.; ITO, T.; NOGUCHI, Y.
2001-08-01
This paper presents a practical approach for predicting the meshing noise due to the impact of chain rollers against the sprocket of chain drives. An acoustical model relating dynamic response of rollers and its induced sound pressure is developed based on the fact that the acoustic field is mainly created by oscillating rigid cylindrical rollers. Finite element techniques and numerical software codes are employed to model and simulate the acceleration response of each chain roller which is necessary for noise level prediction of a chain drive under varying operation conditions and different sprocket configurations. The predicted acoustic pressure levels of meshing noise are compared with the available experimental measurements. It is shown that the predictions are in reasonable agreement with the experiments and the approach enables designers to obtain required information on the noise level of a selected chain drive in a time- and cost-efficient manner.
NASA Astrophysics Data System (ADS)
Fromang, S.; Hennebelle, P.; Teyssier, R.
2006-10-01
Aims. In this paper, we present a new method to perform numerical simulations of astrophysical MHD flows using the Adaptive Mesh Refinement framework and Constrained Transport. Methods: . The algorithm is based on a previous work in which the MUSCL-Hancock scheme was used to evolve the induction equation. In this paper, we detail the extension of this scheme to the full MHD equations and discuss its properties. Results: . Through a series of test problems, we illustrate the performances of this new code using two different MHD Riemann solvers (Lax-Friedrich and Roe) and the need of the Adaptive Mesh Refinement capabilities in some cases. Finally, we show its versatility by applying it to two completely different astrophysical situations well studied in the past years: the growth of the magnetorotational instability in the shearing box and the collapse of magnetized cloud cores. Conclusions: . We have implemented a new Godunov scheme to solve the ideal MHD equations in the AMR code RAMSES. We have shown that it results in a powerful tool that can be applied to a great variety of astrophysical problems, ranging from galaxies formation in the early universe to high resolution studies of molecular cloud collapse in our galaxy.
Multi-dimensional upwind fluctuation splitting scheme with mesh adaption for hypersonic viscous flow
NASA Astrophysics Data System (ADS)
Wood, William Alfred, III
production is shown relative to DMFDSFV. Remarkably the fluctuation splitting scheme shows grid converged skin friction coefficients with only five points in the boundary layer for this case. A viscous Mach 17.6 (perfect gas) cylinder case demonstrates solution monotonicity and heat transfer capability with the fluctuation splitting scheme. While fluctuation splitting is recommended over DMFDSFV, the difference in performance between the schemes is not so great as to obsolete DMFDSFV. The second half of the dissertation develops a local, compact, anisotropic unstructured mesh adaption scheme in conjunction with the multi-dimensional upwind solver, exhibiting a characteristic alignment behavior for scalar problems. This alignment behavior stands in contrast to the curvature clustering nature of the local, anisotropic unstructured adaption strategy based upon a posteriori error estimation that is used for comparison. The characteristic alignment is most pronounced for linear advection, with reduced improvement seen for the more complex non-linear advection and advection-diffusion cases. The adaption strategy is extended to the two-dimensional and axisymmetric Navier-Stokes equations of motion through the concept of fluctuation minimization. The system test case for the adaption strategy is a sting mounted capsule at Mach-10 wind tunnel conditions, considered in both two-dimensional and axisymmetric configurations. For this complex flowfield the adaption results are disappointing since feature alignment does not emerge from the local operations. Aggressive adaption is shown to result in a loss of robustness for the solver, particularly in the bow shock/stagnation point interaction region. Reducing the adaption strength maintains solution robustness but fails to produce significant improvement in the surface heat transfer predictions.
Goffin, Mark A.; Baker, Christopher M.J.; Buchan, Andrew G.; Pain, Christopher C.; Eaton, Matthew D.; Smith, Paul N.
2013-06-01
This article presents a method for goal-based anisotropic adaptive methods for the finite element method applied to the Boltzmann transport equation. The neutron multiplication factor, k{sub eff}, is used as the goal of the adaptive procedure. The anisotropic adaptive algorithm requires error measures for k{sub eff} with directional dependence. General error estimators are derived for any given functional of the flux and applied to k{sub eff} to acquire the driving force for the adaptive procedure. The error estimators require the solution of an appropriately formed dual equation. Forward and dual error indicators are calculated by weighting the Hessian of each solution with the dual and forward residual respectively. The Hessian is used as an approximation of the interpolation error in the solution which gives rise to the directional dependence. The two indicators are combined to form a single error metric that is used to adapt the finite element mesh. The residual is approximated using a novel technique arising from the sub-grid scale finite element discretisation. Two adaptive routes are demonstrated: (i) a single mesh is used to solve all energy groups, and (ii) a different mesh is used to solve each energy group. The second method aims to capture the benefit from representing the flux from each energy group on a specifically optimised mesh. The k{sub eff} goal-based adaptive method was applied to three examples which illustrate the superior accuracy in criticality problems that can be obtained.
Ying, Wenjun; Henriquez, Craig S.
2015-01-01
A both space and time adaptive algorithm is presented for simulating electrical wave propagation in the Purkinje system of the heart. The equations governing the distribution of electric potential over the system are solved in time with the method of lines. At each timestep, by an operator splitting technique, the space-dependent but linear diffusion part and the nonlinear but space-independent reactions part in the partial differential equations are integrated separately with implicit schemes, which have better stability and allow larger timesteps than explicit ones. The linear diffusion equation on each edge of the system is spatially discretized with the continuous piecewise linear finite element method. The adaptive algorithm can automatically recognize when and where the electrical wave starts to leave or enter the computational domain due to external current/voltage stimulation, self-excitation, or local change of membrane properties. Numerical examples demonstrating efficiency and accuracy of the adaptive algorithm are presented. PMID:26581455
Adaptive Urban Dispersion Integrated Model
Wissink, A; Chand, K; Kosovic, B; Chan, S; Berger, M; Chow, F K
2005-11-03
Numerical simulations represent a unique predictive tool for understanding the three-dimensional flow fields and associated concentration distributions from contaminant releases in complex urban settings (Britter and Hanna 2003). Utilization of the most accurate urban models, based on fully three-dimensional computational fluid dynamics (CFD) that solve the Navier-Stokes equations with incorporated turbulence models, presents many challenges. We address two in this work; first, a fast but accurate way to incorporate the complex urban terrain, buildings, and other structures to enforce proper boundary conditions in the flow solution; second, ways to achieve a level of computational efficiency that allows the models to be run in an automated fashion such that they may be used for emergency response and event reconstruction applications. We have developed a new integrated urban dispersion modeling capability based on FEM3MP (Gresho and Chan 1998, Chan and Stevens 2000), a CFD model from Lawrence Livermore National Lab. The integrated capability incorporates fast embedded boundary mesh generation for geometrically complex problems and full three-dimensional Cartesian adaptive mesh refinement (AMR). Parallel AMR and embedded boundary gridding support are provided through the SAMRAI library (Wissink et al. 2001, Hornung and Kohn 2002). Embedded boundary mesh generation has been demonstrated to be an automatic, fast, and efficient approach for problem setup. It has been used for a variety of geometrically complex applications, including urban applications (Pullen et al. 2005). The key technology we introduce in this work is the application of AMR, which allows the application of high-resolution modeling to certain important features, such as individual buildings and high-resolution terrain (including important vegetative and land-use features). It also allows the urban scale model to be readily interfaced with coarser resolution meso or regional scale models. This talk
Longest, P Worth; Vinchurkar, Samir
2007-04-01
A number of research studies have employed a wide variety of mesh styles and levels of grid convergence to assess velocity fields and particle deposition patterns in models of branching biological systems. Generating structured meshes based on hexahedral elements requires significant time and effort; however, these meshes are often associated with high quality solutions. Unstructured meshes that employ tetrahedral elements can be constructed much faster but may increase levels of numerical diffusion, especially in tubular flow systems with a primary flow direction. The objective of this study is to better establish the effects of mesh generation techniques and grid convergence on velocity fields and particle deposition patterns in bifurcating respiratory models. In order to achieve this objective, four widely used mesh styles including structured hexahedral, unstructured tetrahedral, flow adaptive tetrahedral, and hybrid grids have been considered for two respiratory airway configurations. Initial particle conditions tested are based on the inlet velocity profile or the local inlet mass flow rate. Accuracy of the simulations has been assessed by comparisons to experimental in vitro data available in the literature for the steady-state velocity field in a single bifurcation model as well as the local particle deposition fraction in a double bifurcation model. Quantitative grid convergence was assessed based on a grid convergence index (GCI), which accounts for the degree of grid refinement. The hexahedral mesh was observed to have GCI values that were an order of magnitude below the unstructured tetrahedral mesh values for all resolutions considered. Moreover, the hexahedral mesh style provided GCI values of approximately 1% and reduced run times by a factor of 3. Based on comparisons to empirical data, it was shown that inlet particle seedings should be consistent with the local inlet mass flow rate. Furthermore, the mesh style was found to have an observable
Adaptive Mesh Refinement Cosmological Simulations of Cosmic Rays in Galaxy Clusters
NASA Astrophysics Data System (ADS)
Skillman, Samuel William
2013-12-01
Galaxy clusters are unique astrophysical laboratories that contain many thermal and non-thermal phenomena. In particular, they are hosts to cosmic shocks, which propagate through the intracluster medium as a by-product of structure formation. It is believed that at these shock fronts, magnetic field inhomogeneities in a compressing flow may lead to the acceleration of cosmic ray electrons and ions. These relativistic particles decay and radiate through a variety of mechanisms, and have observational signatures in radio, hard X-ray, and Gamma-ray wavelengths. We begin this dissertation by developing a method to find shocks in cosmological adaptive mesh refinement simulations of structure formation. After describing the evolution of shock properties through cosmic time, we make estimates for the amount of kinetic energy processed and the total number of cosmic ray protons that could be accelerated at these shocks. We then use this method of shock finding and a model for the acceleration of and radio synchrotron emission from cosmic ray electrons to estimate the radio emission properties in large scale structures. By examining the time-evolution of the radio emission with respect to the X-ray emission during a galaxy cluster merger, we find that the relative timing of the enhancements in each are important consequences of the shock dynamics. By calculating the radio emission expected from a given mass galaxy cluster, we make estimates for future large-area radio surveys. Next, we use a state-of-the-art magnetohydrodynamic simulation to follow the electron acceleration in a massive merging galaxy cluster. We use the magnetic field information to calculate not only the total radio emission, but also create radio polarization maps that are compared to recent observations. We find that we can naturally reproduce Mpc-scale radio emission that resemble many of the known double radio relic systems. Finally, motivated by our previous studies, we develop and introduce a
Unstructured mesh algorithms for aerodynamic calculations
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1992-01-01
The use of unstructured mesh techniques for solving complex aerodynamic flows is discussed. The principle advantages of unstructured mesh strategies, as they relate to complex geometries, adaptive meshing capabilities, and parallel processing are emphasized. The various aspects required for the efficient and accurate solution of aerodynamic flows are addressed. These include mesh generation, mesh adaptivity, solution algorithms, convergence acceleration, and turbulence modeling. Computations of viscous turbulent two-dimensional flows and inviscid three-dimensional flows about complex configurations are demonstrated. Remaining obstacles and directions for future research are also outlined.
Goal functional evaluations for phase-field fracture using PU-based DWR mesh adaptivity
NASA Astrophysics Data System (ADS)
Wick, Thomas
2016-03-01
In this study, a posteriori error estimation and goal-oriented mesh adaptivity are developed for phase-field fracture propagation. Goal functionals are computed with the dual-weighted residual (DWR) method, which is realized by a recently introduced novel localization technique based on a partition-of-unity (PU). This technique is straightforward to apply since the weak residual is used. The influence of neighboring cells is gathered by the PU. Consequently, neither strong residuals nor jumps over element edges are required. Therefore, this approach facilitates the application of the DWR method to coupled (nonlinear) multiphysics problems such as fracture propagation. These developments then allow for a systematic investigation of the discretization error for certain quantities of interest. Specifically, our focus on the relationship between the phase-field regularization and the spatial discretization parameter in terms of goal functional evaluations is novel.
Galaxy Mergers with Adaptive Mesh Refinement: Star Formation and Hot Gas Outflow
Kim, Ji-hoon; Wise, John H.; Abel, Tom; /KIPAC, Menlo Park /Stanford U., Phys. Dept.
2011-06-22
In hierarchical structure formation, merging of galaxies is frequent and known to dramatically affect their properties. To comprehend these interactions high-resolution simulations are indispensable because of the nonlinear coupling between pc and Mpc scales. To this end, we present the first adaptive mesh refinement (AMR) simulation of two merging, low mass, initially gas-rich galaxies (1.8 x 10{sup 10} M{sub {circle_dot}} each), including star formation and feedback. With galaxies resolved by {approx} 2 x 10{sup 7} total computational elements, we achieve unprecedented resolution of the multiphase interstellar medium, finding a widespread starburst in the merging galaxies via shock-induced star formation. The high dynamic range of AMR also allows us to follow the interplay between the galaxies and their embedding medium depicting how galactic outflows and a hot metal-rich halo form. These results demonstrate that AMR provides a powerful tool in understanding interacting galaxies.
A consistent approach to large eddy simulation using adaptive mesh refinement
Cook, A.W.
1999-09-01
The large eddy simulation of turbulent flows is discussed with particular attention paid to the issue of commutation of differentiation and filtering. Multi-level adaptive mesh refinement is proposed as a means of mostly avoiding commutation errors where increased grid resolution is required to capture key flow features. The strategy is to employ multiple uniform grids in a nested hierarchy using a constant-width filter for each grid. It is shown that commutivity of fine and coarse grid filters must be enforced in order to consistently relate variables at different refinement levels. Methods for treating fine grid boundaries and walls are also discussed. It is shown that errors associated with boundary treatments are small and localized.
NASA Astrophysics Data System (ADS)
Zanotti, O.; Dumbser, M.; Fambri, F.
2016-05-01
We describe a new method for the solution of the ideal MHD equations in special relativity which adopts the following strategy: (i) the main scheme is based on Discontinuous Galerkin (DG) methods, allowing for an arbitrary accuracy of order N+1, where N is the degree of the basis polynomials; (ii) in order to cope with oscillations at discontinuities, an ”a-posteriori” sub-cell limiter is activated, which scatters the DG polynomials of the previous time-step onto a set of 2N+1 sub-cells, over which the solution is recomputed by means of a robust finite volume scheme; (iii) a local spacetime Discontinuous-Galerkin predictor is applied both on the main grid of the DG scheme and on the sub-grid of the finite volume scheme; (iv) adaptive mesh refinement (AMR) with local time-stepping is used. We validate the new scheme and comment on its potential applications in high energy astrophysics.
AstroBEAR: Adaptive Mesh Refinement Code for Ideal Hydrodynamics & Magnetohydrodynamics
NASA Astrophysics Data System (ADS)
Cunningham, Andrew J.; Frank, Adam; Varniere, Peggy; Mitran, Sorin; Jones, Thomas W.
2011-04-01
AstroBEAR is a modular hydrodynamic & magnetohydrodynamic code environment designed for a variety of astrophysical applications. It uses the BEARCLAW package, a multidimensional, Eulerian computational code used to solve hyperbolic systems of equations. AstroBEAR allows adaptive-mesh-refinment (AMR) simulations in 2, 2.5 (i.e., cylindrical), and 3 dimensions, in either cartesian or curvilinear coordinates. Parallel applications are supported through the MPI architecture. AstroBEAR is written in Fortran 90/95 using standard libraries. AstroBEAR supports hydrodynamic (HD) and magnetohydrodynamic (MHD) applications using a variety of spatial and temporal methods. MHD simulations are kept divergence-free via the constrained transport (CT) methods of Balsara & Spicer. Three different equation of state environments are available: ideal gas, gas with differing isentropic γ, and the analytic Thomas-Fermi formulation of A.R. Bell [2]. Current work is being done to develop a more advanced real gas equation of state.
Dynamic Implicit 3D Adaptive Mesh Refinement for Non-Equilibrium Radiation Diffusion
Philip, Bobby; Wang, Zhen; Berrill, Mark A; Rodriguez Rodriguez, Manuel; Pernice, Michael
2014-01-01
The time dependent non-equilibrium radiation diffusion equations are important for solving the transport of energy through radiation in optically thick regimes and find applications in several fields including astrophysics and inertial confinement fusion. The associated initial boundary value problems that are encountered exhibit a wide range of scales in space and time and are extremely challenging to solve. To efficiently and accurately simulate these systems we describe our research on combining techniques that will also find use more broadly for long term time integration of nonlinear multiphysics systems: implicit time integration for efficient long term time integration of stiff multiphysics systems, local control theory based step size control to minimize the required global number of time steps while controlling accuracy, dynamic 3D adaptive mesh refinement (AMR) to minimize memory and computational costs, Jacobian Free Newton Krylov methods on AMR grids for efficient nonlinear solution, and optimal multilevel preconditioner components that provide level independent linear solver convergence.
Goal functional evaluations for phase-field fracture using PU-based DWR mesh adaptivity
NASA Astrophysics Data System (ADS)
Wick, Thomas
2016-06-01
In this study, a posteriori error estimation and goal-oriented mesh adaptivity are developed for phase-field fracture propagation. Goal functionals are computed with the dual-weighted residual (DWR) method, which is realized by a recently introduced novel localization technique based on a partition-of-unity (PU). This technique is straightforward to apply since the weak residual is used. The influence of neighboring cells is gathered by the PU. Consequently, neither strong residuals nor jumps over element edges are required. Therefore, this approach facilitates the application of the DWR method to coupled (nonlinear) multiphysics problems such as fracture propagation. These developments then allow for a systematic investigation of the discretization error for certain quantities of interest. Specifically, our focus on the relationship between the phase-field regularization and the spatial discretization parameter in terms of goal functional evaluations is novel.
Block structured adaptive mesh and time refinement for hybrid, hyperbolic + N-body systems
NASA Astrophysics Data System (ADS)
Miniati, Francesco; Colella, Phillip
2007-11-01
We present a new numerical algorithm for the solution of coupled collisional and collisionless systems, based on the block structured adaptive mesh and time refinement strategy (AMR). We describe the issues associated with the discretization of the system equations and the synchronization of the numerical solution on the hierarchy of grid levels. We implement a code based on a higher order, conservative and directionally unsplit Godunov’s method for hydrodynamics; a symmetric, time centered modified symplectic scheme for collisionless component; and a multilevel, multigrid relaxation algorithm for the elliptic equation coupling the two components. Numerical results that illustrate the accuracy of the code and the relative merit of various implemented schemes are also presented.
NASA Astrophysics Data System (ADS)
Kramer, Stephan C.; Piggott, Matthew D.; Cotter, Colin J.; Pain, Chris C.; Nelson, Rhodri B.
2010-05-01
There are many apparent advantages of the application of unstructured meshes in ocean modelling: a much better representation of the coastal boundaries, the ability to focus resolution in areas of interest, or areas of intensified flow, such as boundary currents, etc. In particular with adaptive mesh technology, where the mesh is adapted during the simulation as the flow evolves, one is able to resolve much smaller features in the often turbulent ocean flow, than would be possible with fixed, structured mesh models. The Imperial College Ocean Model[1], is a non-hydrostatic ocean model that employs fully unstructured adaptive meshes, that allow focussing of resolution not only in the horizontal but also in the vertical. This enables the modelling of physical processes, such as open ocean deep convection, density driven flows on a steep bottom topography, etc. that are very important for the global ocean circulation. The Imperial College Ocean Model has been applied succesfully in the modelling of many of these processes. On the other hand hydrostatic, layered ocean models have a significant advantage in large areas of the oceans where the hydrostatic assumption is valid. The fact that with fully unstructured meshes it is no longer straightforward to separate horizontal, baroptropic modes and vertical, baroclinic dynamics, has consequences for both numerical accuracy and the efficiency of the linear solvers. It has therefore been a challenge for ICOM to remain competitive in these areas with layered mesh models. These problems have been overcome by, amongst others, the development of a new mesh adaptation technique that maintains a columnar structure of the mesh in such areas. The application of multigrid techniques has improved the effiency of the non-hydrostatic pressure solve[2] in such a way that convergence is now independent of aspect ratio, which makes the pressure solve competitive with that of a hydrostatic model. In this contribution an overview will be
Single-pass GPU-raycasting for structured adaptive mesh refinement data
NASA Astrophysics Data System (ADS)
Kaehler, Ralf; Abel, Tom
2013-01-01
Structured Adaptive Mesh Refinement (SAMR) is a popular numerical technique to study processes with high spatial and temporal dynamic range. It reduces computational requirements by adapting the lattice on which the underlying differential equations are solved to most efficiently represent the solution. Particularly in astrophysics and cosmology such simulations now can capture spatial scales ten orders of magnitude apart and more. The irregular locations and extensions of the refined regions in the SAMR scheme and the fact that different resolution levels partially overlap, poses a challenge for GPU-based direct volume rendering methods. kD-trees have proven to be advantageous to subdivide the data domain into non-overlapping blocks of equally sized cells, optimal for the texture units of current graphics hardware, but previous GPU-supported raycasting approaches for SAMR data using this data structure required a separate rendering pass for each node, preventing the application of many advanced lighting schemes that require simultaneous access to more than one block of cells. In this paper we present the first single-pass GPU-raycasting algorithm for SAMR data that is based on a kD-tree. The tree is efficiently encoded by a set of 3D-textures, which allows to adaptively sample complete rays entirely on the GPU without any CPU interaction. We discuss two different data storage strategies to access the grid data on the GPU and apply them to several datasets to prove the benefits of the proposed method.
Radiation hydrodynamics including irradiation and adaptive mesh refinement with AZEuS. I. Methods
NASA Astrophysics Data System (ADS)
Ramsey, J. P.; Dullemond, C. P.
2015-02-01
Aims: The importance of radiation to the physical structure of protoplanetary disks cannot be understated. However, protoplanetary disks evolve with time, and so to understand disk evolution and by association, disk structure, one should solve the combined and time-dependent equations of radiation hydrodynamics. Methods: We implement a new implicit radiation solver in the AZEuS adaptive mesh refinement magnetohydrodynamics fluid code. Based on a hybrid approach that combines frequency-dependent ray-tracing for stellar irradiation with non-equilibrium flux limited diffusion, we solve the equations of radiation hydrodynamics while preserving the directionality of the stellar irradiation. The implementation permits simulations in Cartesian, cylindrical, and spherical coordinates, on both uniform and adaptive grids. Results: We present several hydrostatic and hydrodynamic radiation tests which validate our implementation on uniform and adaptive grids as appropriate, including benchmarks specifically designed for protoplanetary disks. Our results demonstrate that the combination of a hybrid radiation algorithm with AZEuS is an effective tool for radiation hydrodynamics studies, and produces results which are competitive with other astrophysical radiation hydrodynamics codes.
The PLUTO Code for Adaptive Mesh Computations in Astrophysical Fluid Dynamics
NASA Astrophysics Data System (ADS)
Mignone, A.; Zanni, C.; Tzeferacos, P.; van Straalen, B.; Colella, P.; Bodo, G.
2012-01-01
We present a description of the adaptive mesh refinement (AMR) implementation of the PLUTO code for solving the equations of classical and special relativistic magnetohydrodynamics (MHD and RMHD). The current release exploits, in addition to the static grid version of the code, the distributed infrastructure of the CHOMBO library for multidimensional parallel computations over block-structured, adaptively refined grids. We employ a conservative finite-volume approach where primary flow quantities are discretized at the cell center in a dimensionally unsplit fashion using the Corner Transport Upwind method. Time stepping relies on a characteristic tracing step where piecewise parabolic method, weighted essentially non-oscillatory, or slope-limited linear interpolation schemes can be handily adopted. A characteristic decomposition-free version of the scheme is also illustrated. The solenoidal condition of the magnetic field is enforced by augmenting the equations with a generalized Lagrange multiplier providing propagation and damping of divergence errors through a mixed hyperbolic/parabolic explicit cleaning step. Among the novel features, we describe an extension of the scheme to include non-ideal dissipative processes, such as viscosity, resistivity, and anisotropic thermal conduction without operator splitting. Finally, we illustrate an efficient treatment of point-local, potentially stiff source terms over hierarchical nested grids by taking advantage of the adaptivity in time. Several multidimensional benchmarks and applications to problems of astrophysical relevance assess the potentiality of the AMR version of PLUTO in resolving flow features separated by large spatial and temporal disparities.
THE PLUTO CODE FOR ADAPTIVE MESH COMPUTATIONS IN ASTROPHYSICAL FLUID DYNAMICS
Mignone, A.; Tzeferacos, P.; Zanni, C.; Bodo, G.; Van Straalen, B.; Colella, P.
2012-01-01
We present a description of the adaptive mesh refinement (AMR) implementation of the PLUTO code for solving the equations of classical and special relativistic magnetohydrodynamics (MHD and RMHD). The current release exploits, in addition to the static grid version of the code, the distributed infrastructure of the CHOMBO library for multidimensional parallel computations over block-structured, adaptively refined grids. We employ a conservative finite-volume approach where primary flow quantities are discretized at the cell center in a dimensionally unsplit fashion using the Corner Transport Upwind method. Time stepping relies on a characteristic tracing step where piecewise parabolic method, weighted essentially non-oscillatory, or slope-limited linear interpolation schemes can be handily adopted. A characteristic decomposition-free version of the scheme is also illustrated. The solenoidal condition of the magnetic field is enforced by augmenting the equations with a generalized Lagrange multiplier providing propagation and damping of divergence errors through a mixed hyperbolic/parabolic explicit cleaning step. Among the novel features, we describe an extension of the scheme to include non-ideal dissipative processes, such as viscosity, resistivity, and anisotropic thermal conduction without operator splitting. Finally, we illustrate an efficient treatment of point-local, potentially stiff source terms over hierarchical nested grids by taking advantage of the adaptivity in time. Several multidimensional benchmarks and applications to problems of astrophysical relevance assess the potentiality of the AMR version of PLUTO in resolving flow features separated by large spatial and temporal disparities.
NASA Astrophysics Data System (ADS)
Northrup, Scott A.
A new parallel implicit adaptive mesh refinement (AMR) algorithm is developed for the prediction of unsteady behaviour of laminar flames. The scheme is applied to the solution of the system of partial-differential equations governing time-dependent, two- and three-dimensional, compressible laminar flows for reactive thermally perfect gaseous mixtures. A high-resolution finite-volume spatial discretization procedure is used to solve the conservation form of these equations on body-fitted multi-block hexahedral meshes. A local preconditioning technique is used to remove numerical stiffness and maintain solution accuracy for low-Mach-number, nearly incompressible flows. A flexible block-based octree data structure has been developed and is used to facilitate automatic solution-directed mesh adaptation according to physics-based refinement criteria. The data structure also enables an efficient and scalable parallel implementation via domain decomposition. The parallel implicit formulation makes use of a dual-time-stepping like approach with an implicit second-order backward discretization of the physical time, in which a Jacobian-free inexact Newton method with a preconditioned generalized minimal residual (GMRES) algorithm is used to solve the system of nonlinear algebraic equations arising from the temporal and spatial discretization procedures. An additive Schwarz global preconditioner is used in conjunction with block incomplete LU type local preconditioners for each sub-domain. The Schwarz preconditioning and block-based data structure readily allow efficient and scalable parallel implementations of the implicit AMR approach on distributed-memory multi-processor architectures. The scheme was applied to solutions of steady and unsteady laminar diffusion and premixed methane-air combustion and was found to accurately predict key flame characteristics. For a premixed flame under terrestrial gravity, the scheme accurately predicted the frequency of the natural
NASA Astrophysics Data System (ADS)
Lu, Xiaochun; Xiao, Yijun; Chai, Zhi; Wang, Bangping
2007-11-01
An adaptive background model aiming at outdoor vehicle detection is presented in this paper. This model is an improved model of PICA (pixel intensity classification algorithm), it classifies pixels into K-distributions by color similarity, and then a hypothesis that the background pixel color appears in image sequence with a high frequency is used to evaluate all the distributions to determine which presents the current background color. As experiments show, the model presented in this paper is a robust, adaptive and flexible model, which can deal with situations like camera motions, lighting changes and so on.
Three-dimensional modeling and highly refined mesh generation of the aorta artery and its tunics
NASA Astrophysics Data System (ADS)
Cazotto, J. A.; Neves, L. A.; Machado, J. M.; Momente, J. C.; Shiyou, Y.; Godoy, M. F.; Zafalon, G. F. D.; Pinto, A. R.; Valêncio, C. R.
2013-02-01
This paper describes strategies and techniques to perform modeling and automatic mesh generation of the aorta artery and its tunics (adventitia, media and intima walls), using open source codes. The models were constructed in the Blender package and Python scripts were used to export the data necessary for the mesh generation in TetGen. The strategies proposed are able to provide meshes of complicated and irregular volumes, with a large number of mesh elements involved (12,000,000 tetrahedrons approximately). These meshes can be used to perform computational simulations by Finite Element Method (FEM).
Performance Characteristics of an Adaptive Mesh RefinementCalculation on Scalar and Vector Platforms
Welcome, Michael; Rendleman, Charles; Oliker, Leonid; Biswas, Rupak
2006-01-31
Adaptive mesh refinement (AMR) is a powerful technique thatreduces the resources necessary to solve otherwise in-tractable problemsin computational science. The AMR strategy solves the problem on arelatively coarse grid, and dynamically refines it in regions requiringhigher resolution. However, AMR codes tend to be far more complicatedthan their uniform grid counterparts due to the software infrastructurenecessary to dynamically manage the hierarchical grid framework. Despitethis complexity, it is generally believed that future multi-scaleapplications will increasingly rely on adaptive methods to study problemsat unprecedented scale and resolution. Recently, a new generation ofparallel-vector architectures have become available that promise toachieve extremely high sustained performance for a wide range ofapplications, and are the foundation of many leadership-class computingsystems worldwide. It is therefore imperative to understand the tradeoffsbetween conventional scalar and parallel-vector platforms for solvingAMR-based calculations. In this paper, we examine the HyperCLaw AMRframework to compare and contrast performance on the Cray X1E, IBM Power3and Power5, and SGI Altix. To the best of our knowledge, this is thefirst work that investigates and characterizes the performance of an AMRcalculation on modern parallel-vector systems.
Mesh refinement for uncertainty quantification through model reduction
NASA Astrophysics Data System (ADS)
Li, Jing; Stinis, Panos
2015-01-01
We present a novel way of deciding when and where to refine a mesh in probability space in order to facilitate uncertainty quantification in the presence of discontinuities in random space. A discontinuity in random space makes the application of generalized polynomial chaos expansion techniques prohibitively expensive. The reason is that for discontinuous problems, the expansion converges very slowly. An alternative to using higher terms in the expansion is to divide the random space in smaller elements where a lower degree polynomial is adequate to describe the randomness. In general, the partition of the random space is a dynamic process since some areas of the random space, particularly around the discontinuity, need more refinement than others as time evolves. In the current work we propose a way to decide when and where to refine the random space mesh based on the use of a reduced model. The idea is that a good reduced model can monitor accurately, within a random space element, the cascade of activity to higher degree terms in the chaos expansion. In turn, this facilitates the efficient allocation of computational sources to the areas of random space where they are more needed. For the Kraichnan-Orszag system, the prototypical system to study discontinuities in random space, we present theoretical results which show why the proposed method is sound and numerical results which corroborate the theory.
Mesh refinement for uncertainty quantification through model reduction
Li, Jing Stinis, Panos
2015-01-01
We present a novel way of deciding when and where to refine a mesh in probability space in order to facilitate uncertainty quantification in the presence of discontinuities in random space. A discontinuity in random space makes the application of generalized polynomial chaos expansion techniques prohibitively expensive. The reason is that for discontinuous problems, the expansion converges very slowly. An alternative to using higher terms in the expansion is to divide the random space in smaller elements where a lower degree polynomial is adequate to describe the randomness. In general, the partition of the random space is a dynamic process since some areas of the random space, particularly around the discontinuity, need more refinement than others as time evolves. In the current work we propose a way to decide when and where to refine the random space mesh based on the use of a reduced model. The idea is that a good reduced model can monitor accurately, within a random space element, the cascade of activity to higher degree terms in the chaos expansion. In turn, this facilitates the efficient allocation of computational sources to the areas of random space where they are more needed. For the Kraichnan–Orszag system, the prototypical system to study discontinuities in random space, we present theoretical results which show why the proposed method is sound and numerical results which corroborate the theory.
NASA Astrophysics Data System (ADS)
Huang, Rongzong; Wu, Huiying
2016-06-01
A total enthalpy-based lattice Boltzmann (LB) method with adaptive mesh refinement (AMR) is developed in this paper to efficiently simulate solid-liquid phase change problem where variables vary significantly near the phase interface and thus finer grid is required. For the total enthalpy-based LB method, the velocity field is solved by an incompressible LB model with multiple-relaxation-time (MRT) collision scheme, and the temperature field is solved by a total enthalpy-based MRT LB model with the phase interface effects considered and the deviation term eliminated. With a kinetic assumption that the density distribution function for solid phase is at equilibrium state, a volumetric LB scheme is proposed to accurately realize the nonslip velocity condition on the diffusive phase interface and in the solid phase. As compared with the previous schemes, this scheme can avoid nonphysical flow in the solid phase. As for the AMR approach, it is developed based on multiblock grids. An indicator function is introduced to control the adaptive generation of multiblock grids, which can guarantee the existence of overlap area between adjacent blocks for information exchange. Since MRT collision schemes are used, the information exchange is directly carried out in the moment space. Numerical tests are firstly performed to validate the strict satisfaction of the nonslip velocity condition, and then melting problems in a square cavity with different Prandtl numbers and Rayleigh numbers are simulated, which demonstrate that the present method can handle solid-liquid phase change problem with high efficiency and accuracy.
NASA Astrophysics Data System (ADS)
Mizoguchi, Tomohiro; Kanai, Satoshi
Along with the rapid growth of industrial X-ray CT scanning systems, it is now possible to non-destructively acquire the entire meshes of assemblies consisting of a set of parts. For the advanced inspections of the assemblies, such as estimation of their assembling errors or examinations of their behaviors in the motions, based on their CT scanned meshes, it is necessary to accurately decompose the mesh and to extract a set of partial meshes each of which correspond to a part. Moreover it is required to create models which can be used for the real-product based simulations. In this paper, we focus on CT scanned meshes of gear assemblies as examples and propose beneficial methods for establishing such advance inspections of the assemblies. We first propose a method that accurately decomposes the mesh into partial meshes each of which corresponds to a gear based on periodicity recognitions. The key idea is first to accurately recognize the periodicity of each gear and then to extract the partial meshes as sets of topologically connected mesh elements where periodicities are valid. Our method can robustly and accurately recognize periodicities from noisy scanned meshes. In contrast to previous methods, our method can deal with single-material CT scanned meshes and can estimate the correct boundaries of neighboring parts with no previous knowledge. Moreover it can efficiently extract the partial meshes from large scanned meshes containing about one million triangles in a few minutes. We also propose a method for creating simulation models which can be used for a gear teeth contact evaluation using extracted partial meshes and their periodicities. Such an evaluation of teeth contacts is one of the most important functions in kinematic simulations of gear assemblies for predicting the power transmission efficiency, noise and vibration. We demonstrate the effectiveness of our method on a variety of artificial and CT scanned meshes.
NASA Technical Reports Server (NTRS)
Steger, J. L.; Dougherty, F. C.; Benek, J. A.
1983-01-01
A mesh system composed of multiple overset body-conforming grids is described for adapting finite-difference procedures to complex aircraft configurations. In this so-called 'chimera mesh,' a major grid is generated about a main component of the configuration and overset minor grids are used to resolve all other features. Methods for connecting overset multiple grids and modifications of flow-simulation algorithms are discussed. Computational tests in two dimensions indicate that the use of multiple overset grids can simplify the task of grid generation without an adverse effect on flow-field algorithms and computer code complexity.
Performance of a mixed-mesh Godunov-based flood inundation model
NASA Astrophysics Data System (ADS)
Kim, B.; Sanders, B. F.; Kim, H.; Famiglietti, J. S.
2011-12-01
Godunov-based finite volume models for solving the shallow-water equations, which are seeing increasing use in hydrology for flood inundation modeling, have almost exclusively adopted either structured meshes of quadrilateral cells or unstructured meshes of triangular cells but have not focused on mixing quadrilateral and triangular cells as is common with finite element models. Triangular meshing is advantageous in complex topography arising from channel junctions, meandering channels, and artificial structures because powerful Delaunay mesh generators easily accommodate internal and external boundary constraints while retaining mesh quality attributes such as area and angle properties. On the other hand, quadrilateral meshing is advantages in the absence of boundary constraints because efficient and low-overhead mesh designs such as Cartesian grids can be used. Motivated by the need for efficient and accurate modeling of flood inundation including channel flows, overbank flows, and overtopping processes, a mixed-mesh version of the BreZo flood inundation model is presented and its ease-of-use and performance in a series of laboratory and field scale test problems is examined in comparison to versions that use either triangular or quadrilateral cells exclusively. We find that mixed meshes are not as easily prepared as triangular meshes for commonly encountered study-site geometries, but execution times and memory requirements for a similar level of accuracy are reduced. The challenges of designing high quality meshes for flood modeling highlights the need for a new class of mesh generators that can scan high resolution topographic data (e.g., lidar) for critical features (e.g., channels and levees) and automate mixed-mesh generation and parameterization.
3D Boltzmann Simulation of the Io's Plasma Environment with Adaptive Mesh and Particle Refinement
NASA Astrophysics Data System (ADS)
Lipatov, A. S.; Combi, M. R.
2002-12-01
The global dynamics of the ionized and neutral components in the environment of Io plays an important role in the interaction of Jupiter's corotating magnetospheric plasma with Io [Combi et al., 2002; 1998; Kabin et al., 2001]. The stationary simulation of this problem was done in the MHD [Combi et al., 1998; Linker et al, 1998; Kabin et al., 2001] and the electrodynamic [Saur et al., 1999] approaches. In this report, we develop a method of kinetic ion-neutral simulation, which is based on a multiscale adaptive mesh, particle and algorithm refinement. This method employs the fluid description for electrons whereas for ions the drift-kinetic and particle approaches are used. This method takes into account charge-exchange and photoionization processes. The first results of such simulation of the dynamics of ions in the Io's environment are discussed in this report. ~ M R Combi et al., J. Geophys. Res., 103, 9071, 1998. M R Combi, T I Gombosi, K Kabin, Atmospheres in the Solar System: Comparative\\ Aeronomy. Geophys. Monograph Series, 130, 151, 2002. K Kabin et al., Planetary and Space Sci., 49, 337, 2001. J A Linker et al., J. Geophys. Res., 103(E9), 19867, 1998. J Saur et al., J. Geophys. Res., 104, 25105, 1999.
AMR++: A design for parallel object-oriented adaptive mesh refinement
Quinlan, D.
1997-11-01
Adaptive mesh refinement computations are complicated by their dynamic nature. In the serial environment they require substantial infrastructures to support the regridding processes, intergrid operations, and local bookkeeping of positions of grids relative to one another. In the parallel environment the dynamic behavior is more problematic because it requires dynamic distribution support and load balancing. Parallel AMR is further complicated by the substantial task parallelism, in addition to the obvious data parallelism, this task parallelism requires additional infrastructure to support efficiently. The degree of parallelism is typically dependent upon the algorithms in use and the equations being solved. Different algorithms have significant compromises between computation and communication. Substantial research work is often required to define efficient methods and suitable infrastructure. The purpose of this paper is to introduce AMR++ as an object-oriented library which forms a part of the OVERTURE framework, a much larger object-oriented numerical framework developed and supported at Los Alamos National Laboratory and distributed on the Web for the last several years.
Advances in Rotor Performance and Turbulent Wake Simulation Using DES and Adaptive Mesh Refinement
NASA Technical Reports Server (NTRS)
Chaderjian, Neal M.
2012-01-01
Time-dependent Navier-Stokes simulations have been carried out for a rigid V22 rotor in hover, and a flexible UH-60A rotor in forward flight. Emphasis is placed on understanding and characterizing the effects of high-order spatial differencing, grid resolution, and Spalart-Allmaras (SA) detached eddy simulation (DES) in predicting the rotor figure of merit (FM) and resolving the turbulent rotor wake. The FM was accurately predicted within experimental error using SA-DES. Moreover, a new adaptive mesh refinement (AMR) procedure revealed a complex and more realistic turbulent rotor wake, including the formation of turbulent structures resembling vortical worms. Time-dependent flow visualization played a crucial role in understanding the physical mechanisms involved in these complex viscous flows. The predicted vortex core growth with wake age was in good agreement with experiment. High-resolution wakes for the UH-60A in forward flight exhibited complex turbulent interactions and turbulent worms, similar to the V22. The normal force and pitching moment coefficients were in good agreement with flight-test data.
Relativistic Flows Using Spatial And Temporal Adaptive Structured Mesh Refinement. I. Hydrodynamics
Wang, Peng; Abel, Tom; Zhang, Weiqun; /KIPAC, Menlo Park
2007-04-02
Astrophysical relativistic flow problems require high resolution three-dimensional numerical simulations. In this paper, we describe a new parallel three-dimensional code for simulations of special relativistic hydrodynamics (SRHD) using both spatially and temporally structured adaptive mesh refinement (AMR). We used method of lines to discrete SRHD equations spatially and used a total variation diminishing (TVD) Runge-Kutta scheme for time integration. For spatial reconstruction, we have implemented piecewise linear method (PLM), piecewise parabolic method (PPM), third order convex essentially non-oscillatory (CENO) and third and fifth order weighted essentially non-oscillatory (WENO) schemes. Flux is computed using either direct flux reconstruction or approximate Riemann solvers including HLL, modified Marquina flux, local Lax-Friedrichs flux formulas and HLLC. The AMR part of the code is built on top of the cosmological Eulerian AMR code enzo, which uses the Berger-Colella AMR algorithm and is parallel with dynamical load balancing using the widely available Message Passing Interface library. We discuss the coupling of the AMR framework with the relativistic solvers and show its performance on eleven test problems.
EMMA: an adaptive mesh refinement cosmological simulation code with radiative transfer
NASA Astrophysics Data System (ADS)
Aubert, Dominique; Deparis, Nicolas; Ocvirk, Pierre
2015-11-01
EMMA is a cosmological simulation code aimed at investigating the reionization epoch. It handles simultaneously collisionless and gas dynamics, as well as radiative transfer physics using a moment-based description with the M1 approximation. Field quantities are stored and computed on an adaptive three-dimensional mesh and the spatial resolution can be dynamically modified based on physically motivated criteria. Physical processes can be coupled at all spatial and temporal scales. We also introduce a new and optional approximation to handle radiation: the light is transported at the resolution of the non-refined grid and only once the dynamics has been fully updated, whereas thermo-chemical processes are still tracked on the refined elements. Such an approximation reduces the overheads induced by the treatment of radiation physics. A suite of standard tests are presented and passed by EMMA, providing a validation for its future use in studies of the reionization epoch. The code is parallel and is able to use graphics processing units (GPUs) to accelerate hydrodynamics and radiative transfer calculations. Depending on the optimizations and the compilers used to generate the CPU reference, global GPU acceleration factors between ×3.9 and ×16.9 can be obtained. Vectorization and transfer operations currently prevent better GPU performance and we expect that future optimizations and hardware evolution will lead to greater accelerations.
Dynamic implicit 3D adaptive mesh refinement for non-equilibrium radiation diffusion
B. Philip; Z. Wang; M.A. Berrill; M. Birke; M. Pernice
2014-04-01
The time dependent non-equilibrium radiation diffusion equations are important for solving the transport of energy through radiation in optically thick regimes and find applications in several fields including astrophysics and inertial confinement fusion. The associated initial boundary value problems that are encountered often exhibit a wide range of scales in space and time and are extremely challenging to solve. To efficiently and accurately simulate these systems we describe our research on combining techniques that will also find use more broadly for long term time integration of nonlinear multi-physics systems: implicit time integration for efficient long term time integration of stiff multi-physics systems, local control theory based step size control to minimize the required global number of time steps while controlling accuracy, dynamic 3D adaptive mesh refinement (AMR) to minimize memory and computational costs, Jacobian Free Newton–Krylov methods on AMR grids for efficient nonlinear solution, and optimal multilevel preconditioner components that provide level independent solver convergence.
NASA Astrophysics Data System (ADS)
Wang, Cheng; Dong, XinZhuang; Shu, Chi-Wang
2015-10-01
For numerical simulation of detonation, computational cost using uniform meshes is large due to the vast separation in both time and space scales. Adaptive mesh refinement (AMR) is advantageous for problems with vastly different scales. This paper aims to propose an AMR method with high order accuracy for numerical investigation of multi-dimensional detonation. A well-designed AMR method based on finite difference weighted essentially non-oscillatory (WENO) scheme, named as AMR&WENO is proposed. A new cell-based data structure is used to organize the adaptive meshes. The new data structure makes it possible for cells to communicate with each other quickly and easily. In order to develop an AMR method with high order accuracy, high order prolongations in both space and time are utilized in the data prolongation procedure. Based on the message passing interface (MPI) platform, we have developed a workload balancing parallel AMR&WENO code using the Hilbert space-filling curve algorithm. Our numerical experiments with detonation simulations indicate that the AMR&WENO is accurate and has a high resolution. Moreover, we evaluate and compare the performance of the uniform mesh WENO scheme and the parallel AMR&WENO method. The comparison results provide us further insight into the high performance of the parallel AMR&WENO method.
2010-10-05
MeshKit is an open-source library of mesh generation functionality. MeshKit has general mesh manipulation and generation functions such as Copoy, Move, Rotate and Extrude mesh. In addition, new quad mesh and embedded boundary Cartesian mesh algorithm (EB Mesh) are included. Interfaces to several public domain meshing algorithms (TetGen, netgen, triangle, Gmsh, camal) are also offered. This library interacts with mesh data mostly through iMesh including accessing the mesh in parallel. It also can interact withmore » iGeom interface to provide geometry functionality such as importing solid model based geometries. iGeom and IMesh are implemented in the CGM and MOAB packages, respectively. For some non-existing function in iMesh such as tree-construction and ray-tracing, MeshKit also interacts with MOAB functions directly.« less
2010-10-05
MeshKit is an open-source library of mesh generation functionality. MeshKit has general mesh manipulation and generation functions such as Copoy, Move, Rotate and Extrude mesh. In addition, new quad mesh and embedded boundary Cartesian mesh algorithm (EB Mesh) are included. Interfaces to several public domain meshing algorithms (TetGen, netgen, triangle, Gmsh, camal) are also offered. This library interacts with mesh data mostly through iMesh including accessing the mesh in parallel. It also can interact with iGeom interface to provide geometry functionality such as importing solid model based geometries. iGeom and IMesh are implemented in the CGM and MOAB packages, respectively. For some non-existing function in iMesh such as tree-construction and ray-tracing, MeshKit also interacts with MOAB functions directly.
NASA Technical Reports Server (NTRS)
Kleb, William L.; Williams, Marc H.; Batina, John T.
1990-01-01
A temporal adaptive algorithm for the time-integration of the two-dimensional Euler or Navier-Stokes equations is presented. The flow solver involves an upwind flux-split spatial discretization for the convective terms and central differencing for the shear-stress and heat flux terms on an unstructured mesh of triangles. The temporal adaptive algorithm is a time-accurate integration procedure which allows flows with high spatial and temporal gradients to be computed efficiently by advancing each grid cell near its maximum allowable time step. Results indicate that an appreciable computational savings can be achieved for both inviscid and viscous unsteady airfoil problems using unstructured meshes without degrading spatial or temporal accuracy.
Adaptive Finite Element Methods for Continuum Damage Modeling
NASA Technical Reports Server (NTRS)
Min, J. B.; Tworzydlo, W. W.; Xiques, K. E.
1995-01-01
The paper presents an application of adaptive finite element methods to the modeling of low-cycle continuum damage and life prediction of high-temperature components. The major objective is to provide automated and accurate modeling of damaged zones through adaptive mesh refinement and adaptive time-stepping methods. The damage modeling methodology is implemented in an usual way by embedding damage evolution in the transient nonlinear solution of elasto-viscoplastic deformation problems. This nonlinear boundary-value problem is discretized by adaptive finite element methods. The automated h-adaptive mesh refinements are driven by error indicators, based on selected principal variables in the problem (stresses, non-elastic strains, damage, etc.). In the time domain, adaptive time-stepping is used, combined with a predictor-corrector time marching algorithm. The time selection is controlled by required time accuracy. In order to take into account strong temperature dependency of material parameters, the nonlinear structural solution a coupled with thermal analyses (one-way coupling). Several test examples illustrate the importance and benefits of adaptive mesh refinements in accurate prediction of damage levels and failure time.
Parallel adaptive simplical re-meshing for deforming domain CFD computations
NASA Astrophysics Data System (ADS)
Menon, Sandeep; Mooney, Kyle G.; Stapf, K. G.; Schmidt, David P.
2015-10-01
Deforming domains occur in many fields of computational fluid dynamics (CFD), such as interface tracking, simulation of pumps and engines, and fluid/structure interaction. The deformation of the domain presents a challenge to the integrity of the computational mesh; substantial motion of the domain boundaries requires vertex motion and changes in mesh connectivity. For cases of simple boundary motion or structured meshes, predetermined changes to the mesh structure can be sufficient. However, without a priori knowledge of how the domain will change, a more robust solution is required. The present work offers a parallelized solution for simplical meshes that is well-suited to extremely complex geometry. The mesh continuously evolves without user intervention or the use of target meshes. Varying length scales imposed by evolving boundary curvatures and narrow gaps are resolved with a fast length-scale algorithm. The set of algorithms are incorporated into an object-oriented code structure that permits broad application to a range of CFD problems. The robustness and versatility of the algorithm is demonstrated in several examples, representing motion of internal and external boundaries, where the boundary motion may or may not be known a priori.
NASA Astrophysics Data System (ADS)
Ravindran, Prashaanth
The unstable nature of detonation waves is a result of the critical relationship between the hydrodynamic shock and the chemical reactions sustaining the shock. A perturbative analysis of the critical point is quite challenging due to the multiple spatio-temporal scales involved along with the non-linear nature of the shock-reaction mechanism. The author's research attempts to provide detailed resolution of the instabilities at the shock front. Another key aspect of the present research is to develop an understanding of the causality between the non-linear dynamics of the front and the eventual breakdown of the sub-structures. An accurate numerical simulation of detonation waves requires a very efficient solution of the Euler equations in conservation form with detailed, non-equilibrium chemistry. The difference in the flow and reaction length scales results in very stiff source terms, requiring the problem to be solved with adaptive mesh refinement. For this purpose, Berger-Colella's block-structured adaptive mesh refinement (AMR) strategy has been developed and applied to time-explicit finite volume methods. The block-structured technique uses a hierarchy of parent-child sub-grids, integrated recursively over time. One novel approach to partition the problem within a large supercomputer was the use of modified Peano-Hilbert space filling curves. The AMR framework was merged with CLAWPACK, a package providing finite volume numerical methods tailored for wave-propagation problems. The stiffness problem is bypassed by using a 1st order Godunov or a 2nd order Strang splitting technique, where the flow variables and source terms are integrated independently. A linearly explicit fourth-order Runge-Kutta integrator is used for the flow, and an ODE solver was used to overcome the numerical stiffness. Second-order spatial resolution is obtained by using a second-order Roe-HLL scheme with the inclusion of numerical viscosity to stabilize the solution near the discontinuity
Henshaw, W; Schwendeman, D
2007-11-15
This paper describes an approach for the numerical solution of time-dependent partial differential equations in complex three-dimensional domains. The domains are represented by overlapping structured grids, and block-structured adaptive mesh refinement (AMR) is employed to locally increase the grid resolution. In addition, the numerical method is implemented on parallel distributed-memory computers using a domain-decomposition approach. The implementation is flexible so that each base grid within the overlapping grid structure and its associated refinement grids can be independently partitioned over a chosen set of processors. A modified bin-packing algorithm is used to specify the partition for each grid so that the computational work is evenly distributed amongst the processors. All components of the AMR algorithm such as error estimation, regridding, and interpolation are performed in parallel. The parallel time-stepping algorithm is illustrated for initial-boundary-value problems involving a linear advection-diffusion equation and the (nonlinear) reactive Euler equations. Numerical results are presented for both equations to demonstrate the accuracy and correctness of the parallel approach. Exact solutions of the advection-diffusion equation are constructed, and these are used to check the corresponding numerical solutions for a variety of tests involving different overlapping grids, different numbers of refinement levels and refinement ratios, and different numbers of processors. The problem of planar shock diffraction by a sphere is considered as an illustration of the numerical approach for the Euler equations, and a problem involving the initiation of a detonation from a hot spot in a T-shaped pipe is considered to demonstrate the numerical approach for the reactive case. For both problems, the solutions are shown to be well resolved on the finest grid. The parallel performance of the approach is examined in detail for the shock diffraction problem.
NASA Astrophysics Data System (ADS)
Schaal, Kevin; Bauer, Andreas; Chandrashekar, Praveen; Pakmor, Rüdiger; Klingenberg, Christian; Springel, Volker
2015-11-01
Solving the Euler equations of ideal hydrodynamics as accurately and efficiently as possible is a key requirement in many astrophysical simulations. It is therefore important to continuously advance the numerical methods implemented in current astrophysical codes, especially also in light of evolving computer technology, which favours certain computational approaches over others. Here we introduce the new adaptive mesh refinement (AMR) code TENET, which employs a high-order discontinuous Galerkin (DG) scheme for hydrodynamics. The Euler equations in this method are solved in a weak formulation with a polynomial basis by means of explicit Runge-Kutta time integration and Gauss-Legendre quadrature. This approach offers significant advantages over commonly employed second-order finite-volume (FV) solvers. In particular, the higher order capability renders it computationally more efficient, in the sense that the same precision can be obtained at significantly less computational cost. Also, the DG scheme inherently conserves angular momentum in regions where no limiting takes place, and it typically produces much smaller numerical diffusion and advection errors than an FV approach. A further advantage lies in a more natural handling of AMR refinement boundaries, where a fall-back to first order can be avoided. Finally, DG requires no wide stencils at high order, and offers an improved data locality and a focus on local computations, which is favourable for current and upcoming highly parallel supercomputers. We describe the formulation and implementation details of our new code, and demonstrate its performance and accuracy with a set of two- and three-dimensional test problems. The results confirm that DG schemes have a high potential for astrophysical applications.
NASA Astrophysics Data System (ADS)
Power, C.; Read, J. I.; Hobbs, A.
2014-06-01
We simulate cosmological galaxy cluster formation using three different approaches to solving the equations of non-radiative hydrodynamics - classic smoothed particle hydrodynamics (SPH), novel SPH with a higher order dissipation switch (SPHS), and an adaptive mesh refinement (AMR) method. Comparing spherically averaged entropy profiles, we find that SPHS and AMR approaches result in a well-defined entropy core that converges rapidly with increasing mass and force resolution. In contrast, the central entropy profile in the SPH approach is sensitive to the cluster's assembly history and shows poor numerical convergence. We trace this disagreement to the known artificial surface tension in SPH that appears at phase boundaries. Varying systematically numerical dissipation in SPHS, we study the contributions of numerical and physical dissipation to the entropy core and argue that numerical dissipation is required to ensure single-valued fluid quantities in converging flows. However, provided it occurs only at the resolution limit and does not propagate errors to larger scales, its effect is benign - there is no requirement to build `sub-grid' models of unresolved turbulence for galaxy cluster simulations. We conclude that entropy cores in non-radiative galaxy cluster simulations are physical, resulting from entropy generation in shocked gas during cluster assembly.
NASA Astrophysics Data System (ADS)
Sui, Yi; Spelt, Peter D. M.; Ding, Hang
2010-11-01
Diffuse Interface (DI) methods are employed widely for the numerical simulation of two-phase flows, even with moving contact lines. In a DI method, the interface thickness should be as thin as possible to simulate spreading phenomena under realistic flow conditions, so a fine grid is required, beyond the reach of current methods that employ a uniform grid. Here we have integrated a DI method based on a uniform mesh, to a block-based adaptive mesh refinement method, so that only the regions near the interface are resolved by a fine mesh. The performance of the present method is tested by simulations including drop deformation in shear flow, Rayleigh-Taylor instability and drop spreading on a flat surface, et al. The results show that the present method can give accurate results with much smaller computational cost, compared to the original DI method based on a uniform mesh. Based on the present method, simulation of drop spreading is carried out with Cahn number of 0.001 and the contact line region is well resolved. The flow field near the contact line, the contact line speed as well as the apparent contact angle are investigated in detail and compared with previous analytical work.
NASA Astrophysics Data System (ADS)
Zheng, H. W.; Shu, C.; Chew, Y. T.
2008-07-01
In this paper, an object-oriented and quadrilateral-mesh based solution adaptive algorithm for the simulation of compressible multi-fluid flows is presented. The HLLC scheme (Harten, Lax and van Leer approximate Riemann solver with the Contact wave restored) is extended to adaptively solve the compressible multi-fluid flows under complex geometry on unstructured mesh. It is also extended to the second-order of accuracy by using MUSCL extrapolation. The node, edge and cell are arranged in such an object-oriented manner that each of them inherits from a basic object. A home-made double link list is designed to manage these objects so that the inserting of new objects and removing of the existing objects (nodes, edges and cells) are independent of the number of objects and only of the complexity of O( 1). In addition, the cells with different levels are further stored in different lists. This avoids the recursive calculation of solution of mother (non-leaf) cells. Thus, high efficiency is obtained due to these features. Besides, as compared to other cell-edge adaptive methods, the separation of nodes would reduce the memory requirement of redundant nodes, especially in the cases where the level number is large or the space dimension is three. Five two-dimensional examples are used to examine its performance. These examples include vortex evolution problem, interface only problem under structured mesh and unstructured mesh, bubble explosion under the water, bubble-shock interaction, and shock-interface interaction inside the cylindrical vessel. Numerical results indicate that there is no oscillation of pressure and velocity across the interface and it is feasible to apply it to solve compressible multi-fluid flows with large density ratio (1000) and strong shock wave (the pressure ratio is 10,000) interaction with the interface.
A Nonlinear Dynamic Model and Free Vibration Analysis of Deployable Mesh Reflectors
NASA Technical Reports Server (NTRS)
Shi, H.; Yang, B.; Thomson, M.; Fang, H.
2011-01-01
This paper presents a dynamic model of deployable mesh reflectors, in which geometric and material nonlinearities of such a space structure are fully described. Then, by linearization around an equilibrium configuration of the reflector structure, a linearized model is obtained. With this linearized model, the natural frequencies and mode shapes of a reflector can be computed. The nonlinear dynamic model of deployable mesh reflectors is verified by using commercial finite element software in numerical simulation. As shall be seen, the proposed nonlinear model is useful for shape (surface) control of deployable mesh reflectors under thermal loads.
Single fiber model of particle retention in an acoustically driven porous mesh.
Grossner, Michael T; Penrod, Alan E; Belovich, Joanne M; Feke, Donald L
2003-03-01
A method for the capture of small particles (tens of microns in diameter) from a continuously flowing suspension has recently been reported. This technique relies on a standing acoustic wave resonating in a rectangular chamber filled with a high-porosity mesh. Particles are retained in this chamber via a complex interaction between the acoustic field and the porous mesh. Although the mesh has a pore size two orders of magnitude larger than the particle diameter, collection efficiencies of 90% have been measured. A mathematical model has been developed to understand the experimentally observed phenomena and to be able to predict filtration performance. By examining a small region (a single fiber) of the porous mesh, the model has duplicated several experimental events such as the focusing of particles near an element of the mesh and the levitation of particles within pores. The single-fiber analysis forms the basis of modeling the overall performance of the particle filtration system. PMID:12565069
NASA Astrophysics Data System (ADS)
Moura, R. C.; Silva, A. F. C.; Bigarella, E. D. V.; Fazenda, A. L.; Ortega, M. A.
2016-08-01
This paper proposes two important improvements to shock-capturing strategies using a discontinuous Galerkin scheme, namely, accurate shock identification via finite-time Lyapunov exponent (FTLE) operators and efficient shock treatment through a point-implicit discretization of a PDE-based artificial viscosity technique. The advocated approach is based on the FTLE operator, originally developed in the context of dynamical systems theory to identify certain types of coherent structures in a flow. We propose the application of FTLEs in the detection of shock waves and demonstrate the operator's ability to identify strong and weak shocks equally well. The detection algorithm is coupled with a mesh refinement procedure and applied to transonic and supersonic flows. While the proposed strategy can be used potentially with any numerical method, a high-order discontinuous Galerkin solver is used in this study. In this context, two artificial viscosity approaches are employed to regularize the solution near shocks: an element-wise constant viscosity technique and a PDE-based smooth viscosity model. As the latter approach is more sophisticated and preferable for complex problems, a point-implicit discretization in time is proposed to reduce the extra stiffness introduced by the PDE-based technique, making it more competitive in terms of computational cost.
NASA Astrophysics Data System (ADS)
Pantano, C.; Deiterding, R.; Hill, D. J.; Pullin, D. I.
2007-01-01
We present a methodology for the large-eddy simulation of compressible flows with a low-numerical dissipation scheme and structured adaptive mesh refinement (SAMR). A description of a conservative, flux-based hybrid numerical method that uses both centered finite-difference and a weighted essentially non-oscillatory (WENO) scheme is given, encompassing the cases of scheme alternation and internal mesh interfaces resulting from SAMR. In this method, the centered scheme is used in turbulent flow regions while WENO is employed to capture shocks. One-, two- and three-dimensional numerical experiments and example simulations are presented including homogeneous shock-free turbulence, a turbulent jet and the strongly shock-driven mixing of a Richtmyer-Meshkov instability.
NASA Technical Reports Server (NTRS)
Steinthorsson, E.; Modiano, David; Colella, Phillip
1994-01-01
A methodology for accurate and efficient simulation of unsteady, compressible flows is presented. The cornerstones of the methodology are a special discretization of the Navier-Stokes equations on structured body-fitted grid systems and an efficient solution-adaptive mesh refinement technique for structured grids. The discretization employs an explicit multidimensional upwind scheme for the inviscid fluxes and an implicit treatment of the viscous terms. The mesh refinement technique is based on the AMR algorithm of Berger and Colella. In this approach, cells on each level of refinement are organized into a small number of topologically rectangular blocks, each containing several thousand cells. The small number of blocks leads to small overhead in managing data, while their size and regular topology means that a high degree of optimization can be achieved on computers with vector processors.
FEATURE-BASED MULTIBLOCK FINITE ELEMENT MESH GENERATION
Shivanna, Kiran H.; Tadepalli, Srinivas C.; Grosland, Nicole M.
2010-01-01
Hexahedral finite element mesh development for anatomic structures and biomedical implants can be cumbersome. Moreover, using traditional meshing techniques, detailed features may be inadequately captured. In this paper, we describe methodologies to handle multi-feature datasets (i.e., feature edges and surfaces). Coupling multi-feature information with multiblock meshing techniques has enabled anatomic structures, as well as orthopaedic implants, to be readily meshed. Moreover, the projection process, node and element set creation are automated, thus reducing the user interaction during model development. To improve the mesh quality, Laplacian- and optimization-based mesh improvement algorithms have been adapted to the multi-feature datasets. PMID:21076650
FEMHD: An adaptive finite element method for MHD and edge modelling
Strauss, H.R.
1995-07-01
This paper describes the code FEMHD, an adaptive finite element MHD code, which is applied in a number of different manners to model MHD behavior and edge plasma phenomena on a diverted tokamak. The code uses an unstructured triangular mesh in 2D and wedge shaped mesh elements in 3D. The code has been adapted to look at neutral and charged particle dynamics in the plasma scrape off region, and into a full MHD-particle code.
Development of improved analytical models for mesh reflector surfaces
NASA Technical Reports Server (NTRS)
Brand, J. C.; Kauffman, J. F.
1983-01-01
Several methods for computing the reflection coefficients from mesh surfaces are discussed. Some methods mentioned have severe limitations, and the spectral approach appears to be the most attractive alternative. In spite of some inherent problems, the solutions obtained with this method will offer not only the reflection coefficients but also the currents carried on the mesh. This would allow separation of the power lost to resistive terms from that due to transmission loss. Overall, good results and rapid convergence should be obtained from this method when proper care is applied.
Approaches to the automatic generation and control of finite element meshes
NASA Technical Reports Server (NTRS)
Shephard, Mark S.
1987-01-01
The algorithmic approaches being taken to the development of finite element mesh generators capable of automatically discretizing general domains without the need for user intervention are discussed. It is demonstrated that because of the modeling demands placed on a automatic mesh generator, all the approaches taken to date produce unstructured meshes. Consideration is also given to both a priori and a posteriori mesh control devices for automatic mesh generators as well as their integration with geometric modeling and adaptive analysis procedures.
Full-hexahedral structured meshing for image-based computational vascular modeling.
De Santis, Gianluca; De Beule, Matthieu; Van Canneyt, Koen; Segers, Patrick; Verdonck, Pascal; Verhegghe, Benedict
2011-12-01
Image-based computational modeling offers a virtual access to spatially and temporally high resolution flow and structural mechanical data in vivo. Due to inter-subject morphological variability, mesh generation represents a critical step in modeling the patient-specific geometry and is usually performed using unstructured tetrahedral meshing algorithms. Although hexahedral structured meshes are known to provide higher accuracy and reduce the computational costs both for Finite Element Analysis and Computational Fluid Dynamics, their application in computational cardiovascular studies is challenging due to the complex 3D and branching topology of vascular territories. In this study, we propose a robust procedure for structured mesh generation, tailoring the mesh structure to the subject-specific vessel topology. The proposed methodology is based on centerline-based synthetic descriptors (i.e. centerlines, radii and centerlines' normals) which are used to solve the meshing problem following a bottom-up approach. First, topologically equivalent block-structures are placed inside and outside the lumen domain. Then, a projection operation is performed, returning a parametric volume mesh which fits the original triangulated model with sub-micrometric accuracy. Additionally, a three-layered arterial wall (resembling the intima, media and adventitia) is artificially generated, with the possibility of setting variable thickness (e.g. proximal-to-distal tapering) and material anisotropy (e.g. position-dependent collagen-fibers' orientation). This new meshing procedure, implemented using open-source software packages only, is demonstrated on two challenging human cases, being an aortic arch and an abdominal aortic aneurysm. High-quality meshes are generated in both cases, according to shape-quality metrics. By increasing the computation accuracy, the developed meshing tool has the potential to further add "confidence" to the use of computational methods in vascular
Lipnikov, Konstantin; Agouzal, Abdellatif; Vassilevski, Yuri
2009-01-01
We present a new technology for generating meshes minimizing the interpolation and discretization errors or their gradients. The key element of this methodology is construction of a space metric from edge-based error estimates. For a mesh with N{sub h} triangles, the error is proportional to N{sub h}{sup -1} and the gradient of error is proportional to N{sub h}{sup -1/2} which are optimal asymptotics. The methodology is verified with numerical experiments.
Spherical geodesic mesh generation
Fung, Jimmy; Kenamond, Mark Andrew; Burton, Donald E.; Shashkov, Mikhail Jurievich
2015-02-27
In ALE simulations with moving meshes, mesh topology has a direct influence on feature representation and code robustness. In three-dimensional simulations, modeling spherical volumes and features is particularly challenging for a hydrodynamics code. Calculations on traditional spherical meshes (such as spin meshes) often lead to errors and symmetry breaking. Although the underlying differencing scheme may be modified to rectify this, the differencing scheme may not be accessible. This work documents the use of spherical geodesic meshes to mitigate solution-mesh coupling. These meshes are generated notionally by connecting geodesic surface meshes to produce triangular-prismatic volume meshes. This mesh topology is fundamentally different from traditional mesh topologies and displays superior qualities such as topological symmetry. This work describes the geodesic mesh topology as well as motivating demonstrations with the FLAG hydrocode.
Karell, Mara A; Langstaff, Helen K; Halazonetis, Demetrios J; Minghetti, Caterina; Frelat, Mélanie; Kranioti, Elena F
2016-09-01
The commingling of human remains often hinders forensic/physical anthropologists during the identification process, as there are limited methods to accurately sort these remains. This study investigates a new method for pair-matching, a common individualization technique, which uses digital three-dimensional models of bone: mesh-to-mesh value comparison (MVC). The MVC method digitally compares the entire three-dimensional geometry of two bones at once to produce a single value to indicate their similarity. Two different versions of this method, one manual and the other automated, were created and then tested for how well they accurately pair-matched humeri. Each version was assessed using sensitivity and specificity. The manual mesh-to-mesh value comparison method was 100 % sensitive and 100 % specific. The automated mesh-to-mesh value comparison method was 95 % sensitive and 60 % specific. Our results indicate that the mesh-to-mesh value comparison method overall is a powerful new tool for accurately pair-matching commingled skeletal elements, although the automated version still needs improvement. PMID:26966098
NASA Technical Reports Server (NTRS)
Turon, Albert; Davila, Carlos G.; Camanho, Pedro P.; Costa, Josep
2005-01-01
This paper presents a methodology to determine the parameters used in the simulation of delamination in composite materials using decohesion finite elements. A closed-form expression is developed to define the stiffness of the cohesive layer. A novel procedure that allows the use of coarser meshes of decohesion elements in large-scale computations is proposed. The procedure ensures that the energy dissipated by the fracture process is correctly computed. It is shown that coarse-meshed models defined using the approach proposed here yield the same results as the models with finer meshes normally used in the simulation of fracture processes.
NASA Astrophysics Data System (ADS)
Weller, Hilary; Browne, Philip; Budd, Chris; Cullen, Mike
2016-03-01
An equation of Monge-Ampère type has, for the first time, been solved numerically on the surface of the sphere in order to generate optimally transported (OT) meshes, equidistributed with respect to a monitor function. Optimal transport generates meshes that keep the same connectivity as the original mesh, making them suitable for r-adaptive simulations, in which the equations of motion can be solved in a moving frame of reference in order to avoid mapping the solution between old and new meshes and to avoid load balancing problems on parallel computers. The semi-implicit solution of the Monge-Ampère type equation involves a new linearisation of the Hessian term, and exponential maps are used to map from old to new meshes on the sphere. The determinant of the Hessian is evaluated as the change in volume between old and new mesh cells, rather than using numerical approximations to the gradients. OT meshes are generated to compare with centroidal Voronoi tessellations on the sphere and are found to have advantages and disadvantages; OT equidistribution is more accurate, the number of iterations to convergence is independent of the mesh size, face skewness is reduced and the connectivity does not change. However anisotropy is higher and the OT meshes are non-orthogonal. It is shown that optimal transport on the sphere leads to meshes that do not tangle. However, tangling can be introduced by numerical errors in calculating the gradient of the mesh potential. Methods for alleviating this problem are explored. Finally, OT meshes are generated using observed precipitation as a monitor function, in order to demonstrate the potential power of the technique.
NASA Astrophysics Data System (ADS)
Farooq, Umar; Myler, Peter
2014-09-01
This work is concerned with physical testing and numerical simulations of flat and round nose drop-weight impact of carbon fibre-reinforced laminate composite panels to predict ply level failure. Majority of the existing studies on impact of composites by spherical nose impactors are experimental, computational models are simplified, and based on classical laminated plate theories where contributions of through-thickness stresses are neglected. Present work considers flat nose impact and contributions from through-thickness stresses and is mainly simulation based. A computational model was developed in ABAQUS™ software using adaptive meshing techniques. Simulation produced (2D model) stresses were numerically integrated using MATALB™ code to predict through-thickness (3D) stresses. Through-the-thickness stresses were then utilised in advanced failure criteria coded in MATLAB™ software to predict ply level failures. Simulation produced results demonstrate that the computational model can efficiently and effectively predict ply-by-ply failure status of relatively thick laminates.
Klein, R.I. |; Bell, J.; Pember, R.; Kelleher, T.
1993-04-01
The authors present results for high resolution hydrodynamic calculations of the growth and development of instabilities in shock driven imploding spherical geometries in both 2D and 3D. They solve the Eulerian equations of hydrodynamics with a high order Godunov approach using local adaptive mesh refinement to study the temporal and spatial development of the turbulent mixing layer resulting from both Richtmyer Meshkov and Rayleigh Taylor instabilities. The use of a high resolution Eulerian discretization with adaptive mesh refinement permits them to study the detailed three-dimensional growth of multi-mode perturbations far into the non-linear regime for converging geometries. They discuss convergence properties of the simulations by calculating global properties of the flow. They discuss the time evolution of the turbulent mixing layer and compare its development to a simple theory for a turbulent mix model in spherical geometry based on Plesset`s equation. Their 3D calculations show that the constant found in the planar incompressible experiments of Read and Young`s may not be universal for converging compressible flow. They show the 3D time trace of transitional onset to a mixing state using the temporal evolution of volume rendered imaging. Their preliminary results suggest that the turbulent mixing layer loses memory of its initial perturbations for classical Richtmyer Meshkov and Rayleigh Taylor instabilities in spherically imploding shells. They discuss the time evolution of mixed volume fraction and the role of vorticity in converging 3D flows in enhancing the growth of a turbulent mixing layer.
Application of adaptive mesh refinement to particle-in-cell simulations of plasmas and beams
Vay, J.-L.; Colella, P.; Kwan, J.W.; McCorquodale, P.; Serafini, D.B.; Friedman, A.; Grote, D.P.; Westenskow, G.; Adam, J.-C.; Heron, A.; Haber, I.
2003-11-04
Plasma simulations are often rendered challenging by the disparity of scales in time and in space which must be resolved. When these disparities are in distinctive zones of the simulation domain, a method which has proven to be effective in other areas (e.g. fluid dynamics simulations) is the mesh refinement technique. We briefly discuss the challenges posed by coupling this technique with plasma Particle-In-Cell simulations, and present examples of application in Heavy Ion Fusion and related fields which illustrate the effectiveness of the approach. We also report on the status of a collaboration under way at Lawrence Berkeley National Laboratory between the Applied Numerical Algorithms Group (ANAG) and the Heavy Ion Fusion group to upgrade ANAG's mesh refinement library Chombo to include the tools needed by Particle-In-Cell simulation codes.
Cunningham, Andrew J.; Frank, Adam; Varniere, Peggy; Mitran, Sorin; Jones, Thomas W.
2009-06-15
A description is given of the algorithms implemented in the AstroBEAR adaptive mesh-refinement code for ideal magnetohydrodynamics. The code provides several high-resolution shock-capturing schemes which are constructed to maintain conserved quantities of the flow in a finite-volume sense. Divergence-free magnetic field topologies are maintained to machine precision by collating the components of the magnetic field on a cell-interface staggered grid and utilizing the constrained transport approach for integrating the induction equations. The maintenance of magnetic field topologies on adaptive grids is achieved using prolongation and restriction operators which preserve the divergence and curl of the magnetic field across collocated grids of different resolutions. The robustness and correctness of the code is demonstrated by comparing the numerical solution of various tests with analytical solutions or previously published numerical solutions obtained by other codes.
NASA Astrophysics Data System (ADS)
Fang, F.; Zhang, T.; Pavlidis, D.; Pain, C. C.; Buchan, A. G.; Navon, I. M.
2014-10-01
A novel reduced order model (ROM) based on proper orthogonal decomposition (POD) has been developed for a finite-element (FE) adaptive mesh air pollution model. A quadratic expansion of the non-linear terms is employed to ensure the method remained efficient. This is the first time such an approach has been applied to air pollution LES turbulent simulation through three dimensional landscapes. The novelty of this work also includes POD's application within a FE-LES turbulence model that uses adaptive resolution. The accuracy of the reduced order model is assessed and validated for a range of 2D and 3D urban street canyon flow problems. By comparing the POD solutions against the fine detail solutions obtained from the full FE model it is shown that the accuracy is maintained, where fine details of the air flows are captured, whilst the computational requirements are reduced. In the examples presented below the size of the reduced order models is reduced by factors up to 2400 in comparison to the full FE model while the CPU time is reduced by up to 98% of that required by the full model.
Adaptive Models for Gene Networks
Shin, Yong-Jun; Sayed, Ali H.; Shen, Xiling
2012-01-01
Biological systems are often treated as time-invariant by computational models that use fixed parameter values. In this study, we demonstrate that the behavior of the p53-MDM2 gene network in individual cells can be tracked using adaptive filtering algorithms and the resulting time-variant models can approximate experimental measurements more accurately than time-invariant models. Adaptive models with time-variant parameters can help reduce modeling complexity and can more realistically represent biological systems. PMID:22359614
Evaluation of discretization procedures for transition elements in adaptive mesh refinement
NASA Technical Reports Server (NTRS)
Park, K. C.; Levit, Itzak; Stanley, Gary M.
1991-01-01
Three transition interpolation schemes for use in h-or r-refinement have been analyzed in terms of accuracy, implementation ease and extendability. They include blending-function interpolation, displacement averaging, and strain matching at discrete points along the transition edge lines. The results suggest that the choice of matching depends strongly on the element formulations, (viz. displacement or assumed strain, etc.) and mesh refinement criteria employed, and to a lesser extent the choice of computer architecture (serial vs. parallel) and the equation solution procedures. A recommended pairing of some of the elements with the choice factors is suggested.
An adaptive-mesh finite-difference solution method for the Navier-Stokes equations
NASA Astrophysics Data System (ADS)
Luchini, Paolo
1987-02-01
An adjustable variable-spacing grid is presented which permits the addition or deletion of single points during iterative solutions of the Navier-Stokes equations by finite difference methods. The grid is designed for application to two-dimensional steady-flow problems which can be described by partial differential equations whose second derivatives are constrained to the Laplacian operator. An explicit Navier-Stokes equations solution technique defined for use with the grid incorporates a hybrid form of the convective terms. Three methods are developed for automatic modifications of the mesh during calculations.
NASA Technical Reports Server (NTRS)
Lutz, R. J.; Spar, J.
1978-01-01
The Hansen atmospheric model was used to compute five monthly forecasts (October 1976 through February 1977). The comparison is based on an energetics analysis, meridional and vertical profiles, error statistics, and prognostic and observed mean maps. The monthly mean model simulations suffer from several defects. There is, in general, no skill in the simulation of the monthly mean sea-level pressure field, and only marginal skill is indicated for the 850 mb temperatures and 500 mb heights. The coarse-mesh model appears to generate a less satisfactory monthly mean simulation than the finer mesh GISS model.
Effects of Chitosan Coatings on Polypropylene Mesh for Implantation in a Rat Abdominal Wall Model
Udpa, Natasha; Iyer, Shama R.; Rajoria, Rohit; Breyer, Kate E.; Valentine, Helen; Singh, Bhupinder; McDonough, Sean P.; Brown, Bryan N.; Bonassar, Lawrence J.
2013-01-01
Hernia repair and pelvic floor reconstruction are usually accompanied with the implantation of a surgical mesh, which frequently results in a foreign body response with associated complications. An ideal surgical mesh that allows force generation of muscle tissues without significant granulation tissue and/or fibrosis is of significant clinical interest. The objective of the present study was to evaluate the in vitro and in vivo responses of a chitosan coating on polypropylene mesh (Ch-PPM) in comparison with commercially available meshes. We found that application of a 0.5% (w/v) Ch-PPM elicited preferential attachment of myoblasts over fibroblast attachment in vitro. Therefore, we test the hypothesis that 0.5% Ch-PPM will encourage skeletal muscle tissue ingrowth and decrease fibrosis formation in vivo. We implanted 0.5% Ch-PPM, collagen-coated polypropylene mesh (Pelvitex™; C.R. Bard), and polypropylene (Avaulta Solo®; C.R. Bard) alone using a rat abdominal defect model. Force generation capacity and inflammatory response of each mesh were evaluated 2, 4, and 12 weeks postimplantation. We found that chitosan coating is associated with the restoration of functional skeletal muscle with histomorphologic characteristics that resemble native muscle and an early macrophage phenotypic response that has previously been shown to lead to more functional outcomes. PMID:23859182
NASA Astrophysics Data System (ADS)
Angelidis, Dionysios; Sotiropoulos, Fotis
2015-11-01
The geometrical details of wind turbines determine the structure of the turbulence in the near and far wake and should be taken in account when performing high fidelity calculations. Multi-resolution simulations coupled with an immersed boundary method constitutes a powerful framework for high-fidelity calculations past wind farms located over complex terrains. We develop a 3D Immersed-Boundary Adaptive Mesh Refinement flow solver (IB-AMR) which enables turbine-resolving LES of wind turbines. The idea of using a hybrid staggered/non-staggered grid layout adopted in the Curvilinear Immersed Boundary Method (CURVIB) has been successfully incorporated on unstructured meshes and the fractional step method has been employed. The overall performance and robustness of the second order accurate, parallel, unstructured solver is evaluated by comparing the numerical simulations against conforming grid calculations and experimental measurements of laminar and turbulent flows over complex geometries. We also present turbine-resolving multi-scale LES considering all the details affecting the induced flow field; including the geometry of the tower, the nacelle and especially the rotor blades of a wind tunnel scale turbine. This material is based upon work supported by the Department of Energy under Award Number DE-EE0005482 and the Sandia National Laboratories.
Cortical imaging on a head template: a simulation study using a resistor mesh model (RMM).
Chauveau, Nicolas; Franceries, Xavier; Aubry, Florent; Celsis, Pierre; Rigaud, Bernard
2008-09-01
The T1 head template model used in Statistical Parametric Mapping Version 2000 (SPM2), was segmented into five layers (scalp, skull, CSF, grey and white matter) and implemented in 2 mm voxels. We designed a resistor mesh model (RMM), based on the finite volume method (FVM) to simulate the electrical properties of this head model along the three axes for each voxel. Then, we introduced four dipoles of high eccentricity (about 0.8) in this RMM, separately and simultaneously, to compute the potentials for two sets of conductivities. We used the direct cortical imaging technique (CIT) to recover the simulated dipoles, using 60 or 107 electrodes and with or without addition of Gaussian white noise (GWN). The use of realistic conductivities gave better CIT results than standard conductivities, lowering the blurring effect on scalp potentials and displaying more accurate position areas when CIT was applied to single dipoles. Simultaneous dipoles were less accurately localized, but good qualitative and stable quantitative results were obtained up to 5% noise level for 107 electrodes and up to 10% noise level for 60 electrodes, showing that a compromise must be found to optimize both the number of electrodes and the noise level. With the RMM defined in 2 mm voxels, the standard 128-electrode cap and 5% noise appears to be the upper limit providing reliable source positions when direct CIT is used. The admittance matrix defining the RMM is easy to modify so as to adapt to different conductivities. The next step will be the adaptation of individual real head T2 images to the RMM template and the introduction of anisotropy using diffusion imaging (DI). PMID:18629625
Lopez-Camara, D.; Lazzati, Davide; Morsony, Brian J.; Begelman, Mitchell C.
2013-04-10
We present the results of special relativistic, adaptive mesh refinement, 3D simulations of gamma-ray burst jets expanding inside a realistic stellar progenitor. Our simulations confirm that relativistic jets can propagate and break out of the progenitor star while remaining relativistic. This result is independent of the resolution, even though the amount of turbulence and variability observed in the simulations is greater at higher resolutions. We find that the propagation of the jet head inside the progenitor star is slightly faster in 3D simulations compared to 2D ones at the same resolution. This behavior seems to be due to the fact that the jet head in 3D simulations can wobble around the jet axis, finding the spot of least resistance to proceed. Most of the average jet properties, such as density, pressure, and Lorentz factor, are only marginally affected by the dimensionality of the simulations and therefore results from 2D simulations can be considered reliable.
Armstrong, Jerawan C.; Favorite, Jeffrey A.
2012-06-20
The Levenberg-Marquardt (or simply Marquardt) and differential evolution (DE) optimization methods were recently applied to solve inverse transport problems. The Marquardt method is fast but convergence of the method is dependent on the initial guess. While it has been shown to work extremely well at finding an optimum independent of the initial guess, the DE method does not provide a global optimal solution in some problems. In this paper, we apply the Mesh Adaptive Direct Search (MADS) algorithm to solve the inverse problem of material interface location identification in one-dimensional spherical radiation source/shield systems, and we compare the results obtained by MADS to those obtained by Levenberg-Marquardt and DE.
de Tayrac, Renaud; Alves, Antoine; Thérin, Michel
2007-05-01
The aims of this study were dual. First, to evaluate the feasibility of a sheep model as an animal model for vaginal surgery with meshes. Second, to compare host response to two low-weight polypropylene (PP) meshes, a noncoated (Soft Prolene, Gynecare, Ethicon) and a coated mesh with an absorbable hydrophilic film (Ugytex, Sofradim). Thirty-six 20 x 20 mm polypropylene meshes (18 coated and 18 noncoated) were surgically implanted by the vaginal route in 12 adult ewes. Meshes were implanted in the anterior (n=12) and the posterior vaginal compartments (n=24). Animals were killed 1 (n=6) and 12 (n=6) weeks after surgery. Postimplantation evaluation included macroscopical examination, histological and immunohistochemical analysis and histomorphometrical measures of the distance between the meshes and the vaginal epithelium. The experimental procedure was feasible in all cases. Vaginal erosions were observed twice as frequently with the noncoated-PP meshes (6/18, 33.3%) as with the coated-PP meshes (3/18, 16.7%), even if that difference was not significant (p=0.4). However, no differences were observed between the two meshes in terms of shrinkage, tissue ingrowth, inflammatory response, and position of the mesh in the vaginal wall. The mechanism involved in the reduction of vaginal erosion could be due to the lesser adhesion of the coated mesh on the vaginal wound during the early postoperative period. PMID:16941070
Masterlark, Timothy; Lu, Zhiming; Rykhus, Russ
2006-01-01
Interferometric synthetic aperture radar (InSAR) imagery documents the consistent subsidence, during the interval 1992-1999, of a pyroclastic flow deposit (PFD) emplaced during the 1986 eruption of Augustine Volcano, Alaska. We construct finite element models (FEMs) that simulate thermoelastic contraction of the PFD to account for the observed subsidence. Three-dimensional problem domains of the FEMs include a thermoelastic PFD embedded in an elastic substrate. The thickness of the PFD is initially determined from the difference between post- and pre-eruption digital elevation models (DEMs). The initial excess temperature of the PFD at the time of deposition, 640 ??C, is estimated from FEM predictions and an InSAR image via standard least-squares inverse methods. Although the FEM predicts the major features of the observed transient deformation, systematic prediction errors (RMSE=2.2 cm) are most likely associated with errors in the a priori PFD thickness distribution estimated from the DEM differences. We combine an InSAR image, FEMs, and an adaptive mesh algorithm to iteratively optimize the geometry of the PFD with respect to a minimized misfit between the predicted thermoelastic deformation and observed deformation. Prediction errors from an FEM, which includes an optimized PFD geometry and the initial excess PFD temperature estimated from the least-squares analysis, are sub-millimeter (RMSE=0.3 mm). The average thickness (9.3 m), maximum thickness (126 m), and volume (2.1 ?? 107 m3) of the PFD, estimated using the adaptive mesh algorithm, are about twice as large as the respective estimations for the a priori PFD geometry. Sensitivity analyses suggest unrealistic PFD thickness distributions are required for initial excess PFD temperatures outside of the range 500-800 ??C. ?? 2005 Elsevier B.V. All rights reserved.
Three-dimensional gravity modeling and focusing inversion using rectangular meshes.
Commer, M.
2011-03-01
Rectangular grid cells are commonly used for the geophysical modeling of gravity anomalies, owing to their flexibility in constructing complex models. The straightforward handling of cubic cells in gravity inversion algorithms allows for a flexible imposition of model regularization constraints, which are generally essential in the inversion of static potential field data. The first part of this paper provides a review of commonly used expressions for calculating the gravity of a right polygonal prism, both for gravity and gradiometry, where the formulas of Plouff and Forsberg are adapted. The formulas can be cast into general forms practical for implementation. In the second part, a weighting scheme for resolution enhancement at depth is presented. Modelling the earth using highly digitized meshes, depth weighting schemes are typically applied to the model objective functional, subject to minimizing the data misfit. The scheme proposed here involves a non-linear conjugate gradient inversion scheme with a weighting function applied to the non-linear conjugate gradient scheme's gradient vector of the objective functional. The low depth resolution due to the quick decay of the gravity kernel functions is counteracted by suppressing the search directions in the parameter space that would lead to near-surface concentrations of gravity anomalies. Further, a density parameter transformation function enabling the imposition of lower and upper bounding constraints is employed. Using synthetic data from models of varying complexity and a field data set, it is demonstrated that, given an adequate depth weighting function, the gravity inversion in the transform space can recover geologically meaningful models requiring a minimum of prior information and user interaction.
Computational framework to model and design surgical meshes for hernia repair.
Hernández-Gascón, B; Espés, N; Peña, E; Pascual, G; Bellón, J M; Calvo, B
2014-08-01
Surgical procedures for hernia surgery are usually performed using prosthetic meshes. In spite of all the improvements in these biomaterials, the perfect match between the prosthesis and the implant site has not been achieved. Thus, new designs of surgical meshes are still being developed. Previous to implantation in humans, the validity of the meshes has to be addressed, and to date experimental studies have been the gold standard in testing and validating new implants. Nevertheless, these procedures involve long periods of time and are expensive. Thus, a computational framework for the simulation of prosthesis and surgical procedures may overcome some disadvantages of the experimental methods. The computational framework includes two computational models for designing and validating the behaviour of new meshes, respectively. Firstly, the beam model, which reproduces the exact geometry of the mesh, is set to design the weave and determine the stiffness of the surgical prosthesis. However, this implies a high computational cost whereas the membrane model, defined within the framework of the large deformation hyperelasticity, is a relatively inexpensive computational tool, which also enables a prosthesis to be included in more complex geometries such as human or animal bodies. PMID:23167618
Adaptive Numerical Algorithms in Space Weather Modeling
NASA Technical Reports Server (NTRS)
Toth, Gabor; vanderHolst, Bart; Sokolov, Igor V.; DeZeeuw, Darren; Gombosi, Tamas I.; Fang, Fang; Manchester, Ward B.; Meng, Xing; Nakib, Dalal; Powell, Kenneth G.; Stout, Quentin F.; Glocer, Alex; Ma, Ying-Juan; Opher, Merav
2010-01-01
Space weather describes the various processes in the Sun-Earth system that present danger to human health and technology. The goal of space weather forecasting is to provide an opportunity to mitigate these negative effects. Physics-based space weather modeling is characterized by disparate temporal and spatial scales as well as by different physics in different domains. A multi-physics system can be modeled by a software framework comprising of several components. Each component corresponds to a physics domain, and each component is represented by one or more numerical models. The publicly available Space Weather Modeling Framework (SWMF) can execute and couple together several components distributed over a parallel machine in a flexible and efficient manner. The framework also allows resolving disparate spatial and temporal scales with independent spatial and temporal discretizations in the various models. Several of the computationally most expensive domains of the framework are modeled by the Block-Adaptive Tree Solar wind Roe Upwind Scheme (BATS-R-US) code that can solve various forms of the magnetohydrodynamics (MHD) equations, including Hall, semi-relativistic, multi-species and multi-fluid MHD, anisotropic pressure, radiative transport and heat conduction. Modeling disparate scales within BATS-R-US is achieved by a block-adaptive mesh both in Cartesian and generalized coordinates. Most recently we have created a new core for BATS-R-US: the Block-Adaptive Tree Library (BATL) that provides a general toolkit for creating, load balancing and message passing in a 1, 2 or 3 dimensional block-adaptive grid. We describe the algorithms of BATL and demonstrate its efficiency and scaling properties for various problems. BATS-R-US uses several time-integration schemes to address multiple time-scales: explicit time stepping with fixed or local time steps, partially steady-state evolution, point-implicit, semi-implicit, explicit/implicit, and fully implicit numerical
Adaptive numerical algorithms in space weather modeling
NASA Astrophysics Data System (ADS)
Tóth, Gábor; van der Holst, Bart; Sokolov, Igor V.; De Zeeuw, Darren L.; Gombosi, Tamas I.; Fang, Fang; Manchester, Ward B.; Meng, Xing; Najib, Dalal; Powell, Kenneth G.; Stout, Quentin F.; Glocer, Alex; Ma, Ying-Juan; Opher, Merav
2012-02-01
Space weather describes the various processes in the Sun-Earth system that present danger to human health and technology. The goal of space weather forecasting is to provide an opportunity to mitigate these negative effects. Physics-based space weather modeling is characterized by disparate temporal and spatial scales as well as by different relevant physics in different domains. A multi-physics system can be modeled by a software framework comprising several components. Each component corresponds to a physics domain, and each component is represented by one or more numerical models. The publicly available Space Weather Modeling Framework (SWMF) can execute and couple together several components distributed over a parallel machine in a flexible and efficient manner. The framework also allows resolving disparate spatial and temporal scales with independent spatial and temporal discretizations in the various models. Several of the computationally most expensive domains of the framework are modeled by the Block-Adaptive Tree Solarwind Roe-type Upwind Scheme (BATS-R-US) code that can solve various forms of the magnetohydrodynamic (MHD) equations, including Hall, semi-relativistic, multi-species and multi-fluid MHD, anisotropic pressure, radiative transport and heat conduction. Modeling disparate scales within BATS-R-US is achieved by a block-adaptive mesh both in Cartesian and generalized coordinates. Most recently we have created a new core for BATS-R-US: the Block-Adaptive Tree Library (BATL) that provides a general toolkit for creating, load balancing and message passing in a 1, 2 or 3 dimensional block-adaptive grid. We describe the algorithms of BATL and demonstrate its efficiency and scaling properties for various problems. BATS-R-US uses several time-integration schemes to address multiple time-scales: explicit time stepping with fixed or local time steps, partially steady-state evolution, point-implicit, semi-implicit, explicit/implicit, and fully implicit
Modelling hypervelocity impact fracture of ceramic panels using a mesh-free method
NASA Astrophysics Data System (ADS)
Das, R.; Mikhail, J.; Cleary, P. W.
2010-06-01
This paper studies the application of Smoothed Particle Hydrodynamics (SPH) for modelling hyper-velocity impact fracture and fragmentation in ceramic panels. Numerical modelling of complex fracture processes is important to understand the fundamental failure mechanisms in a variety of systems. Finite Element Method (FEM) is the mesh-based method conventionally applied to numerical simulation of fracture and fragmentation. However, the mesh generation and manipulation do not often provide the desired accuracy of the solutions, especially in problems with extreme deformations and discontinuities. To overcome this, here we use a mesh-free method called Smoothed Particle Hydrodynamic (SPH) to investigate the three-dimensional fracture of ceramic panels. The effect of impact speed on the fracture pattern and energy transfer is analysed. The SPH simulations are found to be robust in understanding the fracture mechanisms and in providing crucial design parameters.
The effect of enoxaparin on seroma and mesh-tissue adhesion in a hernia model.
Ozkececi, Ziya T; Gonul, Yucel; Karavelioglu, Afra; Bozkurt, Mehmet F; Kacar, Emre; Bal, Ahmet; Ozsoy, Mustafa; Turamanlar, Ozan; Celep, Bahadir
2016-07-01
The aim of this study was to investigate whether enoxaparin (ENX) administration would increase seroma risk and worsen mesh tissue recovery in an experimental rat hernia repair model. Fifty-six adult male Wistar-Albino rats were included in the study. Rats were equally and randomly separated into seven groups: Group 1, Control, only subcutaneous dissection was performed; group 2, Sham, Hernia defect was primary sutured; Group 3, Prolene mesh; Group 4, Dual mesh; Group 5, ENX + Sham; Group 6, ENX + Prolene mesh; Group 7, ENX + Dual mesh. ENX was subcutaneously injected at a dose of 180 U/kg per day for 7 days. Rats were killed after the amount of subcutaneous seroma was determined by ultrasound on day 7 following the surgical procedure. Mesh-tissue healing was evaluated using histopathological and immunohistochemical (CD31) staining methods. The mean seroma amount significantly increased in Groups 5-7 compared to Groups 2-4. CD31 immunostaining showed a reduction in neovascularization in Groups 6 and 7, compared to Groups 3 and 4. Neovascularization decreased and hemorrhage, necrosis and oedema findings remarkably increased in Groups 6 and 7, when compared to Groups 3 and 4. Fibroblastic activity and inflammation were more prominent in Groups 3 and 4. It should be kept in mind that ENX interferes with inflammation, which is desired in the early period of healing and leads to an increase in overall seroma amount with anti-coagulant effects, which in turn may disrupt wound healing and mesh-tissue adhesions, as was indicated in our study. PMID:27113250
Feasibility of electrical impedance tomography in haemorrhagic stroke treatment using adaptive mesh
NASA Astrophysics Data System (ADS)
Nasehi Tehrani, J.; Anderson, C.; Jin, C.; van Schaik, A.; Holder, D.; McEwan, A.
2010-04-01
EIT has been proposed for acute stroke differentiation, specifically to determine the type of stroke, either ischaemia (clot) or haemorrhage (bleed) to allow the rapid use of clot-busting drugs in the former (Romsauerova et al 2006) . This addresses an important medical need, although there is little treatment offered in the case of haemorrhage. Also the demands on EIT are high with usually no availability to take a 'before' measurement, ruling out time difference imaging. Recently a new treatment option for haemorrhage has been proposed and is being studied in international randomised controlled trial: the early reduction of elevated blood pressure to attenuate the haematoma. This has been shown via CT to reduce bleeds by up to 1mL by Anderson et al 2008. The use of EIT as a continuous measure is desirable here to monitor the effect of blood pressure reduction. A 1mL increase of haemorrhagic lesion located near scalp on the right side of head caused a boundary voltage change of less than 0.05% at 50 kHz. This could be visually observed in a time difference 3D reconstruction with no change in electrode positions, mesh, background conductivity or drift when baseline noise was less than 0.005% but not when noise was increased to 0.01%. This useful result informs us that the EIT system must have noise of less than 0.005% at 50 kHz including instrumentation, physiological and other biases.
The response of cranial biomechanical finite element models to variations in mesh density.
Bright, Jen A; Rayfield, Emily J
2011-04-01
Finite element (FE) models provide discrete solutions to continuous problems. Therefore, to arrive at the correct solution, it is vital to ensure that FE models contain a sufficient number of elements to fully resolve all the detail encountered in a continuum structure. Mesh convergence testing is the process of comparing successively finer meshes to identify the point of diminishing returns; where increasing resolution has marginal effects on results and further detail would become costly and unnecessary. Historically, convergence has not been considered in most CT-based biomechanical reconstructions involving complex geometries like the skull, as generating such models has been prohibitively time-consuming. To assess how mesh convergence influences results, 18 increasingly refined CT-based models of a domestic pig skull were compared to identify the point of convergence for strain and displacement, using both linear and quadratic tetrahedral elements. Not all regions of the skull converged at the same rate, and unexpectedly, areas of high strain converged faster than low-strain regions. Linear models were slightly stiffer than their quadratic counterparts, but did not converge less rapidly. As expected, insufficiently dense models underestimated strain and displacement, and failed to resolve strain "hot-spots" notable in contour plots. In addition to quantitative differences, visual assessments of such plots often inform conclusions drawn in many comparative studies, highlighting that mesh convergence should be performed on all finite element models before further analysis takes place. PMID:21370496
Mesh Optimization for Monte Carlo-Based Optical Tomography
Edmans, Andrew; Intes, Xavier
2015-01-01
Mesh-based Monte Carlo techniques for optical imaging allow for accurate modeling of light propagation in complex biological tissues. Recently, they have been developed within an efficient computational framework to be used as a forward model in optical tomography. However, commonly employed adaptive mesh discretization techniques have not yet been implemented for Monte Carlo based tomography. Herein, we propose a methodology to optimize the mesh discretization and analytically rescale the associated Jacobian based on the characteristics of the forward model. We demonstrate that this method maintains the accuracy of the forward model even in the case of temporal data sets while allowing for significant coarsening or refinement of the mesh. PMID:26566523
ADAPT model: Model use, calibration and validation
Technology Transfer Automated Retrieval System (TEKTRAN)
This paper presents an overview of the Agricultural Drainage and Pesticide Transport (ADAPT) model and a case study to illustrate the calibration and validation steps for predicting subsurface tile drainage and nitrate-N losses from an agricultural system. The ADAPT model is a daily time step field ...
Manual for automatic generation of finite element models of spiral bevel gears in mesh
NASA Technical Reports Server (NTRS)
Bibel, G. D.; Reddy, S.; Kumar, A.
1994-01-01
The goal of this research is to develop computer programs that generate finite element models suitable for doing 3D contact analysis of faced milled spiral bevel gears in mesh. A pinion tooth and a gear tooth are created and put in mesh. There are two programs: Points.f and Pat.f to perform the analysis. Points.f is based on the equation of meshing for spiral bevel gears. It uses machine tool settings to solve for an N x M mesh of points on the four surfaces, pinion concave and convex, and gear concave and convex. Points.f creates the file POINTS.OUT, an ASCI file containing N x M points for each surface. (N is the number of node points along the length of the tooth, and M is nodes along the height.) Pat.f reads POINTS.OUT and creates the file tl.out. Tl.out is a series of PATRAN input commands. In addition to the mesh density on the tooth face, additional user specified variables are the number of finite elements through the thickness, and the number of finite elements along the tooth full fillet. A full fillet is assumed to exist for both the pinion and gear.
Medical case-based retrieval: integrating query MeSH terms for query-adaptive multi-modal fusion
NASA Astrophysics Data System (ADS)
Seco de Herrera, Alba G.; Foncubierta-Rodríguez, Antonio; Müller, Henning
2015-03-01
Advances in medical knowledge give clinicians more objective information for a diagnosis. Therefore, there is an increasing need for bibliographic search engines that can provide services helping to facilitate faster information search. The ImageCLEFmed benchmark proposes a medical case-based retrieval task. This task aims at retrieving articles from the biomedical literature that are relevant for differential diagnosis of query cases including a textual description and several images. In the context of this campaign many approaches have been investigated showing that the fusion of visual and text information can improve the precision of the retrieval. However, fusion does not always lead to better results. In this paper, a new query-adaptive fusion criterion to decide when to use multi-modal (text and visual) or only text approaches is presented. The proposed method integrates text information contained in MeSH (Medical Subject Headings) terms extracted and visual features of the images to find synonym relations between them. Given a text query, the query-adaptive fusion criterion decides when it is suitable to also use visual information for the retrieval. Results show that this approach can decide if a text or multi{modal approach should be used with 77.15% of accuracy.
woptic: Optical conductivity with Wannier functions and adaptive k-mesh refinement
NASA Astrophysics Data System (ADS)
Assmann, E.; Wissgott, P.; Kuneš, J.; Toschi, A.; Blaha, P.; Held, K.
2016-05-01
We present an algorithm for the adaptive tetrahedral integration over the Brillouin zone of crystalline materials, and apply it to compute the optical conductivity, dc conductivity, and thermopower. For these quantities, whose contributions are often localized in small portions of the Brillouin zone, adaptive integration is especially relevant. Our implementation, the woptic package, is tied into the WIEN2WANNIER framework and allows including a local many-body self energy, e.g. from dynamical mean-field theory (DMFT). Wannier functions and dipole matrix elements are computed with the DFT package WIEN2k and Wannier90. For illustration, we show DFT results for fcc-Al and DMFT results for the correlated metal SrVO3.
A Three-Dimensional Multi-Mesh Lattice Boltzmann Model for Multiphysics Simulations
NASA Astrophysics Data System (ADS)
Hashemi, Amirreza; Eshraghi, Mohsen; Felicelli, Sergio
2015-11-01
The lattice Boltzmann method (LBM) is known as an attractive computational method for modeling fluid flow and, more recently, transport phenomena. As any numerical method, the computational cost of LBM simulations depends on the density of the computational grids. The cost of simulations can become enormous when multiple equations are solved in three dimensions. In this work, the development of a multi-block multi-grid LBM model is discussed for three-dimensional (3D) multiphysics simulations. In a system of multiple coupled equations with different length scales, a multi-block mesh with different grids for each model would enhance the computational efficiency and stability of the model. Embedded-type grids facilitate the transfer of information between lattices while allowing larger time steps. In addition, a non-uniform mesh is considered within each mode that allows mesh refinement within each physical model when required. The multi-mesh method was developed to solve for transport phenomena including fluid flow, mass and heat transfer. The huge memory demands of LBM simulations in 3D was significantly reduced using this scheme. Moreover, by reducing the number of lattice points, cost communication in parallel processing was largely decreased.
A parallel second-order adaptive mesh algorithm for incompressible flow in porous media.
Pau, George S H; Almgren, Ann S; Bell, John B; Lijewski, Michael J
2009-11-28
In this paper, we present a second-order accurate adaptive algorithm for solving multi-phase, incompressible flow in porous media. We assume a multi-phase form of Darcy's law with relative permeabilities given as a function of the phase saturation. The remaining equations express conservation of mass for the fluid constituents. In this setting, the total velocity, defined to be the sum of the phase velocities, is divergence free. The basic integration method is based on a total-velocity splitting approach in which we solve a second-order elliptic pressure equation to obtain a total velocity. This total velocity is then used to recast component conservation equations as nonlinear hyperbolic equations. Our approach to adaptive refinement uses a nested hierarchy of logically rectangular grids with simultaneous refinement of the grids in both space and time. The integration algorithm on the grid hierarchy is a recursive procedure in which coarse grids are advanced in time, fine grids are advanced multiple steps to reach the same time as the coarse grids and the data at different levels are then synchronized. The single-grid algorithm is described briefly, but the emphasis here is on the time-stepping procedure for the adaptive hierarchy. Numerical examples are presented to demonstrate the algorithm's accuracy and convergence properties and to illustrate the behaviour of the method. PMID:19840985
A Parallel Second-Order Adaptive Mesh Algorithm for Incompressible Flow in Porous Media
Pau, George Shu Heng; Almgren, Ann S.; Bell, John B.; Lijewski, Michael J.
2008-04-01
In this paper we present a second-order accurate adaptive algorithm for solving multiphase, incompressible flows in porous media. We assume a multiphase form of Darcy's law with relative permeabilities given as a function of the phase saturation. The remaining equations express conservation of mass for the fluid constituents. In this setting the total velocity, defined to be the sum of the phase velocities, is divergence-free. The basic integration method is based on a total-velocity splitting approach in which we solve a second-order elliptic pressure equation to obtain a total velocity. This total velocity is then used to recast component conservation equations as nonlinear hyperbolic equations. Our approach to adaptive refinement uses a nested hierarchy of logically rectangular grids with simultaneous refinement of the grids in both space and time. The integration algorithm on the grid hierarchy is a recursive procedure in which coarse grids are advanced in time, fine grids areadvanced multiple steps to reach the same time as the coarse grids and the data atdifferent levels are then synchronized. The single grid algorithm is described briefly,but the emphasis here is on the time-stepping procedure for the adaptive hierarchy. Numerical examples are presented to demonstrate the algorithm's accuracy and convergence properties and to illustrate the behavior of the method.
On the coupling between fluid flow and mesh motion in the modelling of fluid structure interaction
NASA Astrophysics Data System (ADS)
Dettmer, Wulf G.; Perić, Djordje
2008-12-01
Partitioned Newton type solution strategies for the strongly coupled system of equations arising in the computational modelling of fluid solid interaction require the evaluation of various coupling terms. An essential part of all ALE type solution strategies is the fluid mesh motion. In this paper, we investigate the effect of the terms which couple the fluid flow with the fluid mesh motion on the convergence behaviour of the overall solution procedure. We show that the computational efficiency of the simulation of many fluid solid interaction processes, including fluid flow through flexible pipes, can be increased significantly if some of these coupling terms are calculated exactly.
Hybrid Adaptive Flight Control with Model Inversion Adaptation
NASA Technical Reports Server (NTRS)
Nguyen, Nhan
2011-01-01
This study investigates a hybrid adaptive flight control method as a design possibility for a flight control system that can enable an effective adaptation strategy to deal with off-nominal flight conditions. The hybrid adaptive control blends both direct and indirect adaptive control in a model inversion flight control architecture. The blending of both direct and indirect adaptive control provides a much more flexible and effective adaptive flight control architecture than that with either direct or indirect adaptive control alone. The indirect adaptive control is used to update the model inversion controller by an on-line parameter estimation of uncertain plant dynamics based on two methods. The first parameter estimation method is an indirect adaptive law based on the Lyapunov theory, and the second method is a recursive least-squares indirect adaptive law. The model inversion controller is therefore made to adapt to changes in the plant dynamics due to uncertainty. As a result, the modeling error is reduced that directly leads to a decrease in the tracking error. In conjunction with the indirect adaptive control that updates the model inversion controller, a direct adaptive control is implemented as an augmented command to further reduce any residual tracking error that is not entirely eliminated by the indirect adaptive control.
Left ventricle motion modeling and analysis by adaptive-size physically based models
NASA Astrophysics Data System (ADS)
Huang, Wen-Chen; Goldgof, Dmitry B.
1992-06-01
This paper presents a new physically based modeling method which employs adaptive-size meshes to model left ventricle (LV) shape and track its motion during cardiac cycle. The mesh size increases or decreases dynamically during surface reconstruction process to locate nodes near surface areas of interest and to minimize the fitting error. Further, presented with multiple 3-D data frames, the mesh size varies as the LV undergoes nonrigid motion. Simulation results illustrate the performance and accuracy of the proposed algorithm. Then, the algorithm is applied to the volumetric temporal cardiac data. The LV data was acquired by the 3-D computed tomography scanner. It was provided by Dr. Eric Hoffman at University of Pennsylvania Medical school and consists of 16 volumetric (128 by 128 by 118) images taken through the heart cycle.
Model of aircraft noise adaptation
NASA Technical Reports Server (NTRS)
Dempsey, T. K.; Coates, G. D.; Cawthorn, J. M.
1977-01-01
Development of an aircraft noise adaptation model, which would account for much of the variability in the responses of subjects participating in human response to noise experiments, was studied. A description of the model development is presented. The principal concept of the model, was the determination of an aircraft adaptation level which represents an annoyance calibration for each individual. Results showed a direct correlation between noise level of the stimuli and annoyance reactions. Attitude-personality variables were found to account for varying annoyance judgements.
Adaptive h -refinement for reduced-order models: ADAPTIVE h -refinement for reduced-order models
Carlberg, Kevin T.
2014-11-05
Our work presents a method to adaptively refine reduced-order models a posteriori without requiring additional full-order-model solves. The technique is analogous to mesh-adaptive h-refinement: it enriches the reduced-basis space online by ‘splitting’ a given basis vector into several vectors with disjoint support. The splitting scheme is defined by a tree structure constructed offline via recursive k-means clustering of the state variables using snapshot data. This method identifies the vectors to split online using a dual-weighted-residual approach that aims to reduce error in an output quantity of interest. The resulting method generates a hierarchy of subspaces online without requiring large-scale operationsmore » or full-order-model solves. Furthermore, it enables the reduced-order model to satisfy any prescribed error tolerance regardless of its original fidelity, as a completely refined reduced-order model is mathematically equivalent to the original full-order model. Experiments on a parameterized inviscid Burgers equation highlight the ability of the method to capture phenomena (e.g., moving shocks) not contained in the span of the original reduced basis.« less
3D active shape models of human brain structures: application to patient-specific mesh generation
NASA Astrophysics Data System (ADS)
Ravikumar, Nishant; Castro-Mateos, Isaac; Pozo, Jose M.; Frangi, Alejandro F.; Taylor, Zeike A.
2015-03-01
The use of biomechanics-based numerical simulations has attracted growing interest in recent years for computer-aided diagnosis and treatment planning. With this in mind, a method for automatic mesh generation of brain structures of interest, using statistical models of shape (SSM) and appearance (SAM), for personalised computational modelling is presented. SSMs are constructed as point distribution models (PDMs) while SAMs are trained using intensity profiles sampled from a training set of T1-weighted magnetic resonance images. The brain structures of interest are, the cortical surface (cerebrum, cerebellum & brainstem), lateral ventricles and falx-cerebri membrane. Two methods for establishing correspondences across the training set of shapes are investigated and compared (based on SSM quality): the Coherent Point Drift (CPD) point-set registration method and B-spline mesh-to-mesh registration method. The MNI-305 (Montreal Neurological Institute) average brain atlas is used to generate the template mesh, which is deformed and registered to each training case, to establish correspondence over the training set of shapes. 18 healthy patients' T1-weightedMRimages form the training set used to generate the SSM and SAM. Both model-training and model-fitting are performed over multiple brain structures simultaneously. Compactness and generalisation errors of the BSpline-SSM and CPD-SSM are evaluated and used to quantitatively compare the SSMs. Leave-one-out cross validation is used to evaluate SSM quality in terms of these measures. The mesh-based SSM is found to generalise better and is more compact, relative to the CPD-based SSM. Quality of the best-fit model instance from the trained SSMs, to test cases are evaluated using the Hausdorff distance (HD) and mean absolute surface distance (MASD) metrics.
Modeling, Modal Properties, and Mesh Stiffness Variation Instabilities of Planetary Gears
NASA Technical Reports Server (NTRS)
Parker, Robert G.; Lin, Jian; Krantz, Timothy L. (Technical Monitor)
2001-01-01
Planetary gear noise and vibration are primary concerns in their applications in helicopters, automobiles, aircraft engines, heavy machinery and marine vehicles. Dynamic analysis is essential to the noise and vibration reduction. This work analytically investigates some critical issues and advances the understanding of planetary gear dynamics. A lumped-parameter model is built for the dynamic analysis of general planetary gears. The unique properties of the natural frequency spectra and vibration modes are rigorously characterized. These special structures apply for general planetary gears with cyclic symmetry and, in practically important case, systems with diametrically opposed planets. The special vibration properties are useful for subsequent research. Taking advantage of the derived modal properties, the natural frequency and vibration mode sensitivities to design parameters are investigated. The key parameters include mesh stiffnesses, support/bearing stiffnesses, component masses, moments of inertia, and operating speed. The eigen-sensitivities are expressed in simple, closed-form formulae associated with modal strain and kinetic energies. As disorders (e.g., mesh stiffness variation. manufacturing and assembling errors) disturb the cyclic symmetry of planetary gears, their effects on the free vibration properties are quantitatively examined. Well-defined veering rules are derived to identify dramatic changes of natural frequencies and vibration modes under parameter variations. The knowledge of free vibration properties, eigen-sensitivities, and veering rules provide important information to effectively tune the natural frequencies and optimize structural design to minimize noise and vibration. Parametric instabilities excited by mesh stiffness variations are analytically studied for multi-mesh gear systems. The discrepancies of previous studies on parametric instability of two-stage gear chains are clarified using perturbation and numerical methods. The
A Dynamic Mesh-Based Approach to Model Melting and Shape of an ESR Electrode
NASA Astrophysics Data System (ADS)
Karimi-Sibaki, E.; Kharicha, A.; Bohacek, J.; Wu, M.; Ludwig, A.
2015-10-01
This paper presents a numerical method to investigate the shape of tip and melt rate of an electrode during electroslag remelting process. The interactions between flow, temperature, and electromagnetic fields are taken into account. A dynamic mesh-based approach is employed to model the dynamic formation of the shape of electrode tip. The effect of slag properties such as thermal and electrical conductivities on the melt rate and electrode immersion depth is discussed. The thermal conductivity of slag has a dominant influence on the heat transfer in the system, hence on melt rate of electrode. The melt rate decreases with increasing thermal conductivity of slag. The electrical conductivity of slag governs the electric current path that in turn influences flow and temperature fields. The melting of electrode is a quite unstable process due to the complex interaction between the melt rate, immersion depth, and shape of electrode tip. Therefore, a numerical adaptation of electrode position in the slag has been implemented in order to achieve steady state melting. In fact, the melt rate, immersion depth, and shape of electrode tip are interdependent parameters of process. The generated power in the system is found to be dependent on both immersion depth and shape of electrode tip. In other words, the same amount of power was generated for the systems where the shapes of tip and immersion depth were different. Furthermore, it was observed that the shape of electrode tip is very similar for the systems running with the same ratio of power generation to melt rate. Comparison between simulations and experimental results was made to verify the numerical model.
Feature-Sensitive Tetrahedral Mesh Generation with Guaranteed Quality
Wang, Jun; Yu, Zeyun
2012-01-01
Tetrahedral meshes are being extensively used in finite element methods (FEM). This paper proposes an algorithm to generate feature-sensitive and high-quality tetrahedral meshes from an arbitrary surface mesh model. A top-down octree subdivision is conducted on the surface mesh and a set of tetrahedra are constructed using adaptive body-centered cubic (BCC) lattices. Special treatments are given to the tetrahedra near the surface such that the quality of the resulting tetrahedral mesh is provably guaranteed: the smallest dihedral angle is always greater than 5.71°. The meshes generated by our method are not only adaptive from the interior to the boundary, but also feature-sensitive on the surface with denser elements in high-curvature regions where geometric feature most likely reside. A variety of experimental results are presented to demonstrate the effectiveness and robustness of this algorithm. PMID:22328787
Valiant load-balanced robust routing under hose model for WDM mesh networks
NASA Astrophysics Data System (ADS)
Zhang, Xiaoning; Li, Lemin; Wang, Sheng
2006-09-01
In this paper, we propose Valiant Load-Balanced robust routing scheme for WDM mesh networks under the model of polyhedral uncertainty (i.e., hose model), and the proposed routing scheme is implemented with traffic grooming approach. Our Objective is to maximize the hose model throughput. A mathematic formulation of Valiant Load-Balanced robust routing is presented and three fast heuristic algorithms are also proposed. When implementing Valiant Load-Balanced robust routing scheme to WDM mesh networks, a novel traffic-grooming algorithm called MHF (minimizing hop first) is proposed. We compare the three heuristic algorithms with the VPN tree under the hose model. Finally we demonstrate in the simulation results that MHF with Valiant Load-Balanced robust routing scheme outperforms the traditional traffic-grooming algorithm in terms of the throughput for the uniform/non-uniform traffic matrix under the hose model.
Rate sensitive continuum damage models and mesh dependence in finite element analyses.
Ljustina, Goran; Fagerström, Martin; Larsson, Ragnar
2014-01-01
The experiences from orthogonal machining simulations show that the Johnson-Cook (JC) dynamic failure model exhibits significant element size dependence. Such mesh dependence is a direct consequence of the utilization of local damage models. The current contribution is an investigation of the extent of the possible pathological mesh dependence. A comparison of the resulting JC model behavior combined with two types of damage evolution is considered. The first damage model is the JC dynamic failure model, where the development of the "damage" does not affect the response until the critical state is reached. The second one is a continuum damage model, where the damage variable is affecting the material response continuously during the deformation. Both the plasticity and the damage models are rate dependent, and the damage evolutions for both models are defined as a postprocessing of the effective stress response. The investigation is conducted for a series of 2D shear tests utilizing different FE representations of the plane strain plate with pearlite material properties. The results show for both damage models, using realistic pearlite material parameters, that similar extent of the mesh dependence is obtained and that the possible viscous regularization effects are absent in the current investigation. PMID:25530994
Rate Sensitive Continuum Damage Models and Mesh Dependence in Finite Element Analyses
Fagerström, Martin
2014-01-01
The experiences from orthogonal machining simulations show that the Johnson-Cook (JC) dynamic failure model exhibits significant element size dependence. Such mesh dependence is a direct consequence of the utilization of local damage models. The current contribution is an investigation of the extent of the possible pathological mesh dependence. A comparison of the resulting JC model behavior combined with two types of damage evolution is considered. The first damage model is the JC dynamic failure model, where the development of the “damage” does not affect the response until the critical state is reached. The second one is a continuum damage model, where the damage variable is affecting the material response continuously during the deformation. Both the plasticity and the damage models are rate dependent, and the damage evolutions for both models are defined as a postprocessing of the effective stress response. The investigation is conducted for a series of 2D shear tests utilizing different FE representations of the plane strain plate with pearlite material properties. The results show for both damage models, using realistic pearlite material parameters, that similar extent of the mesh dependence is obtained and that the possible viscous regularization effects are absent in the current investigation. PMID:25530994
Surface mesh to voxel data registration for patient-specific anatomical modeling
NASA Astrophysics Data System (ADS)
de Oliveira, Júlia E. E.; Giessler, Paul; Keszei, András.; Herrler, Andreas; Deserno, Thomas M.
2016-03-01
Virtual Physiological Human (VPH) models are frequently used for training, planning, and performing medical procedures. The Regional Anaesthesia Simulator and Assistant (RASimAs) project has the goal of increasing the application and effectiveness of regional anesthesia (RA) by combining a simulator of ultrasound-guided and electrical nerve-stimulated RA procedures and a subject-specific assistance system through an integration of image processing, physiological models, subject-specific data, and virtual reality. Individualized models enrich the virtual training tools for learning and improving regional anaesthesia (RA) skills. Therefore, we suggest patient-specific VPH models that are composed by registering the general mesh-based models with patient voxel data-based recordings. Specifically, the pelvis region has been focused for the support of the femoral nerve block. The processing pipeline is composed of different freely available toolboxes such as MatLab, the open Simulation framework (SOFA), and MeshLab. The approach of Gilles is applied for mesh-to-voxel registration. Personalized VPH models include anatomical as well as mechanical properties of the tissues. Two commercial VPH models (Zygote and Anatomium) were used together with 34 MRI data sets. Results are presented for the skin surface and pelvic bones. Future work will extend the registration procedure to cope with all model tissue (i.e., skin, muscle, bone, vessel, nerve, fascia) in a one-step procedure and extrapolating the personalized models to body regions actually being out of the captured field of view.
Finite element meshing of ANSYS (trademark) solid models
NASA Technical Reports Server (NTRS)
Kelley, F. S.
1987-01-01
A large scale, general purpose finite element computer program, ANSYS, developed and marketed by Swanson Analysis Systems, Inc. is discussed. ANSYS was perhaps the first commercially available program to offer truly interactive finite element model generation. ANSYS's purpose is for solid modeling. This application is briefly discussed and illustrated.
SCEC CVM-Toolkit (CVM-T) -- High Performance Meshing Tools for SCEC Community Velocity Models
NASA Astrophysics Data System (ADS)
Small, P.; Maechling, P. J.; Ely, G. P.; Olsen, K. B.; Withers, K.; Graves, R. W.; Jordan, T. H.; Plesch, A.; Shaw, J. H.
2010-12-01
The SCEC Community Velocity Model Toolkit (CVM-T) enables earthquake modelers to quickly build, visualize, and validate large-scale 3D velocity meshes using SCEC CVM-H or CVM-4. CVM-T is comprised of three main components: (1) a current SCEC community velocity model for Southern California, (2) tools for extracting meshes from this model and visualizing them, and (3) an automated test framework for evaluating new releases of CVMs using SCEC’s AWP-ODC forward wave propagation software and one, or more, ground motion goodness of fit (GoF) algorithms. CVM-T is designed to help SCEC modelers build large-scale velocity meshes by extracting material properties from the most current version of Community Velocity Model H (CVM-H) and to provide a consistent interface as new CVM-H versions are developed. The CVM-T software provides a highly-scalable interface to CVM-H 6.2 (and later) voxets. Along with an improved interface to CVM-H material properties, the CVM-T software adds a geotechnical layer (GTL) to CVM-H 6.2+ based on Ely’s Vs30-derived GTL. The initial release of CVM-T also extends the coverage region for CVM-H 6.2 with a Hadley-Kanamori 1D background. Smoothing is performed within the transition boundary between the core model and the 1D background. The user interface now includes a C API that allows applications to query the model either by elevation or depth. The Extraction and Visualization Tools (EVT) include a parallelized 3D mesh generator which can quickly generate meshes (consisting of Vp, Vs, and density) from either CVM-H or CVM-4 with over 100 billion points. Python plotting scripts can be employed to plot horizontal or profile slices from existing meshes or directly from either CVM. The Automated Test Framework (ATF) is a system for quantitatively evaluating new versions of CVM-H and ensuring that the model improves against prior versions. The ATF employs the CruiseControl build and test framework to run an AWP-ODC simulation for the 2008 Chino
The model of rough wetting for hydrophobic steel meshes that mimic Asparagus setaceus leaf.
Jiang, Zai X; Geng, Lin; Huang, Yu D; Guan, Shi A; Dong, W; Ma, Zi Y
2011-02-15
A comprehensive analytical model is proposed to provide a relationship between the macroscopic roughness and contact angle, which is used to develop macroscopic rough surface and to create biomimetic superhydrophobic surfaces. Using chemical surface modification of steel wires, an artificial hydrophobic surface was prepared. A steel mesh mimicking the Asparagus setaceus leaf was created by lowing the surface energy and enhancing macroscopic surface roughness. Water contact angles as high as 129.0° were achieved on the steel mesh with 200μm×200μm pore size. Bad agreement between the predictions based on the original Cassie-Baxter model and experiments was obtained. The version of the Cassie-Baxter model in current use could not be applied to this problem since the roughness magnitude changes from nano/microscopic to macroscopic. A new model, called macroscopic Cassie-Baxter (MCB) model, is constructed by the introduction of contact area density (δ) to original Cassie-Baxter model. It is shown that the measured data is in good agreement with the predicted data based on the MCB model. This model not only for solving macroscopic hydrophobic problems of meshes, but also can be used to solve that of other materials with macroscopic roughness. PMID:21115180
NASA Astrophysics Data System (ADS)
Rasia, Elena; Lau, Erwin T.; Borgani, Stefano; Nagai, Daisuke; Dolag, Klaus; Avestruz, Camille; Granato, Gian Luigi; Mazzotta, Pasquale; Murante, Giuseppe; Nelson, Kaylea; Ragone-Figueroa, Cinthia
2014-08-01
Analyses of cosmological hydrodynamic simulations of galaxy clusters suggest that X-ray masses can be underestimated by 10%-30%. The largest bias originates from both violation of hydrostatic equilibrium (HE) and an additional temperature bias caused by inhomogeneities in the X-ray-emitting intracluster medium (ICM). To elucidate this large dispersion among theoretical predictions, we evaluate the degree of temperature structures in cluster sets simulated either with smoothed-particle hydrodynamics (SPH) or adaptive-mesh refinement (AMR) codes. We find that the SPH simulations produce larger temperature variations connected to the persistence of both substructures and their stripped cold gas. This difference is more evident in nonradiative simulations, whereas it is reduced in the presence of radiative cooling. We also find that the temperature variation in radiative cluster simulations is generally in agreement with that observed in the central regions of clusters. Around R 500 the temperature inhomogeneities of the SPH simulations can generate twice the typical HE mass bias of the AMR sample. We emphasize that a detailed understanding of the physical processes responsible for the complex thermal structure in ICM requires improved resolution and high-sensitivity observations in order to extend the analysis to higher temperature systems and larger cluster-centric radii.
Rasia, Elena; Lau, Erwin T.; Nagai, Daisuke; Avestruz, Camille; Borgani, Stefano; Dolag, Klaus; Granato, Gian Luigi; Murante, Giuseppe; Ragone-Figueroa, Cinthia; Mazzotta, Pasquale; Nelson, Kaylea
2014-08-20
Analyses of cosmological hydrodynamic simulations of galaxy clusters suggest that X-ray masses can be underestimated by 10%-30%. The largest bias originates from both violation of hydrostatic equilibrium (HE) and an additional temperature bias caused by inhomogeneities in the X-ray-emitting intracluster medium (ICM). To elucidate this large dispersion among theoretical predictions, we evaluate the degree of temperature structures in cluster sets simulated either with smoothed-particle hydrodynamics (SPH) or adaptive-mesh refinement (AMR) codes. We find that the SPH simulations produce larger temperature variations connected to the persistence of both substructures and their stripped cold gas. This difference is more evident in nonradiative simulations, whereas it is reduced in the presence of radiative cooling. We also find that the temperature variation in radiative cluster simulations is generally in agreement with that observed in the central regions of clusters. Around R {sub 500} the temperature inhomogeneities of the SPH simulations can generate twice the typical HE mass bias of the AMR sample. We emphasize that a detailed understanding of the physical processes responsible for the complex thermal structure in ICM requires improved resolution and high-sensitivity observations in order to extend the analysis to higher temperature systems and larger cluster-centric radii.
Ly{alpha} RADIATIVE TRANSFER IN COSMOLOGICAL SIMULATIONS USING ADAPTIVE MESH REFINEMENT
Laursen, Peter; Razoumov, Alexei O.; Sommer-Larsen, Jesper E-mail: razoumov@ap.smu.ca
2009-05-01
A numerical code for solving various Ly{alpha} radiative transfer (RT) problems is presented. The code is suitable for an arbitrary, three-dimensional distribution of Ly{alpha} emissivity, gas temperature, density, and velocity field. Capable of handling Ly{alpha} RT in an adaptively refined grid-based structure, it enables detailed investigation of the effects of clumpiness of the interstellar (or intergalactic) medium. The code is tested against various geometrically and physically idealized configurations for which analytical solutions exist, and subsequently applied to three different simulated high-resolution 'Lyman-break galaxies', extracted from high-resolution cosmological simulations at redshift z = 3.6. Proper treatment of the Ly{alpha} scattering reveals a diversity of surface brightness (SB) and line profiles. Specifically, for a given galaxy the maximum observed SB can vary by an order of magnitude, and the total flux by a factor of 3-6, depending on the viewing angle. This may provide an explanation for differences in observed properties of high-redshift galaxies, and in particular a possible physical link between Lyman-break galaxies and regular Ly{alpha} emitters.
NASA Astrophysics Data System (ADS)
Pelties, C.; Käser, M.
2010-12-01
We will present recent developments concerning the extensions of the ADER-DG method to solve three dimensional dynamic rupture problems on unstructured tetrahedral meshes. The simulation of earthquake rupture dynamics and seismic wave propagation using a discontinuous Galerkin (DG) method in 2D was recently presented by J. de la Puente et al. (2009). A considerable feature of this study regarding spontaneous rupture problems was the combination of the DG scheme and a time integration method using Arbitrarily high-order DERivatives (ADER) to provide high accuracy in space and time with the discretization on unstructured meshes. In the resulting discrete velocity-stress formulation of the elastic wave equations variables are naturally discontinuous at the interfaces between elements. The so-called Riemann problem can then be solved to obtain well defined values of the variables at the discontinuity itself. This is in particular valid for the fault at which a certain friction law has to be evaluated. Hence, the fault’s geometry is honored by the computational mesh. This way, complex fault planes can be modeled adequately with small elements while fast mesh coarsening is possible with increasing distance from the fault. Due to the strict locality of the scheme using only direct neighbor communication, excellent parallel behavior can be observed. A further advantage of the scheme is that it avoids spurious high-frequency contributions in the slip rate spectra and therefore does not require artificial Kelvin-Voigt damping or filtering of synthetic seismograms. In order to test the accuracy of the ADER-DG method the Southern California Earthquake Center (SCEC) benchmark for spontaneous rupture simulations was employed. Reference: J. de la Puente, J.-P. Ampuero, and M. Käser (2009), Dynamic rupture modeling on unstructured meshes using a discontinuous Galerkin method, JOURNAL OF GEOPHYSICAL RESEARCH, VOL. 114, B10302, doi:10.1029/2008JB006271
NASA Technical Reports Server (NTRS)
Mikhaylov, Rebecca; Dawson, Douglas; Kwack, Eug
2014-01-01
NASA's Earth observing Soil Moisture Active & Passive (SMAP) Mission is scheduled to launch in November 2014 into a 685 km near-polar, sun synchronous orbit. SMAP will provide comprehensive global mapping measurements of soil moisture and freeze/thaw state in order to enhance understanding of the processes that link the water, energy, and carbon cycles. The primary objectives of SMAP are to improve worldwide weather and flood forecasting, enhance climate prediction, and refine drought and agriculture monitoring during its 3 year mission. The SMAP instrument architecture incorporates an L-band radar and an L-band radiometer which share a common feed horn and parabolic mesh reflector. The instrument rotates about the nadir axis at approximately 15 rpm, thereby providing a conically scanning wide swath antenna beam that is capable of achieving global coverage within 3 days. In order to make the necessary precise surface emission measurements from space, a temperature knowledge of 60 deg C for the mesh reflector is required. In order to show compliance, a thermal vacuum test was conducted using a portable solar simulator to illuminate a non flight, but flight-like test article through the quartz window of the vacuum chamber. The molybdenum wire of the antenna mesh is too fine to accommodate thermal sensors for direct temperature measurements. Instead, the mesh temperature was inferred from resistance measurements made during the test. The test article was rotated to five separate angles between 10 deg and 90 deg via chamber breaks to simulate the maximum expected on-orbit solar loading during the mission. The resistance measurements were converted to temperature via a resistance versus temperature calibration plot that was constructed from data collected in a separate calibration test. A simple thermal model of two different representations of the mesh (plate and torus) was created to correlate the mesh temperature predictions to within 60 deg C. The on-orbit mesh
Method For Model-Reference Adaptive Control
NASA Technical Reports Server (NTRS)
Seraji, Homayoun
1990-01-01
Relatively simple method of model-reference adaptive control (MRAC) developed from two prior classes of MRAC techniques: signal-synthesis method and parameter-adaption method. Incorporated into unified theory, which yields more general adaptation scheme.
Parameterized reduced order models from a single mesh using hyper-dual numbers
NASA Astrophysics Data System (ADS)
Brake, M. R. W.; Fike, J. A.; Topping, S. D.
2016-06-01
In order to assess the predicted performance of a manufactured system, analysts must consider random variations (both geometric and material) in the development of a model, instead of a single deterministic model of an idealized geometry with idealized material properties. The incorporation of random geometric variations, however, potentially could necessitate the development of thousands of nearly identical solid geometries that must be meshed and separately analyzed, which would require an impractical number of man-hours to complete. This research advances a recent approach to uncertainty quantification by developing parameterized reduced order models. These parameterizations are based upon Taylor series expansions of the system's matrices about the ideal geometry, and a component mode synthesis representation for each linear substructure is used to form an efficient basis with which to study the system. The numerical derivatives required for the Taylor series expansions are obtained via hyper-dual numbers, and are compared to parameterized models constructed with finite difference formulations. The advantage of using hyper-dual numbers is two-fold: accuracy of the derivatives to machine precision, and the need to only generate a single mesh of the system of interest. The theory is applied to a stepped beam system in order to demonstrate proof of concept. The results demonstrate that the hyper-dual number multivariate parameterization of geometric variations, which largely are neglected in the literature, are accurate for both sensitivity and optimization studies. As model and mesh generation can constitute the greatest expense of time in analyzing a system, the foundation to create a parameterized reduced order model based off of a single mesh is expected to reduce dramatically the necessary time to analyze multiple realizations of a component's possible geometry.
OMEGA: The operational multiscale environment model with grid adaptivity
Bacon, D.P.
1995-07-01
This review talk describes the OMEGA code, used for weather simulation and the modeling of aerosol transport through the atmosphere. Omega employs a 3D mesh of wedge shaped elements (triangles when viewed from above) that adapt with time. Because wedges are laid out in layers of triangular elements, the scheme can utilize structured storage and differencing techniques along the elevation coordinate, and is thus a hybrid of structured and unstructured methods. The utility of adaptive gridding in this moded, near geographic features such as coastlines, where material properties change discontinuously, is illustrated. Temporal adaptivity was used additionally to track moving internal fronts, such as clouds of aerosol contaminants. The author also discusses limitations specific to this problem, including manipulation of huge data bases and fixed turn-around times. In practice, the latter requires a carefully tuned optimization between accuracy and computation speed.
Computational results for parallel unstructured mesh computations
Jones, M.T.; Plassmann, P.E.
1994-12-31
The majority of finite element models in structural engineering are composed of unstructured meshes. These unstructured meshes are often very large and require significant computational resources; hence they are excellent candidates for massively parallel computation. Parallel solution of the sparse matrices that arise from such meshes has been studied heavily, and many good algorithms have been developed. Unfortunately, many of the other aspects of parallel unstructured mesh computation have gone largely ignored. The authors present a set of algorithms that allow the entire unstructured mesh computation process to execute in parallel -- including adaptive mesh refinement, equation reordering, mesh partitioning, and sparse linear system solution. They briefly describe these algorithms and state results regarding their running-time and performance. They then give results from the 512-processor Intel DELTA for a large-scale structural analysis problem. These results demonstrate that the new algorithms are scalable and efficient. The algorithms are able to achieve up to 2.2 gigaflops for this unstructured mesh problem.
MAST-2D diffusive model for flood prediction on domains with triangular Delaunay unstructured meshes
NASA Astrophysics Data System (ADS)
Aricò, C.; Sinagra, M.; Begnudelli, L.; Tucciarelli, T.
2011-11-01
A new methodology for the solution of the 2D diffusive shallow water equations over Delaunay unstructured triangular meshes is presented. Before developing the new algorithm, the following question is addressed: it is worth developing and using a simplified shallow water model, when well established algorithms for the solution of the complete one do exist? The governing Partial Differential Equations are discretized using a procedure similar to the linear conforming Finite Element Galerkin scheme, with a different flux formulation and a special flux treatment that requires Delaunay triangulation but entire solution monotonicity. A simple mesh adjustment is suggested, that attains the Delaunay condition for all the triangle sides without changing the original nodes location and also maintains the internal boundaries. The original governing system is solved applying a fractional time step procedure, that solves consecutively a convective prediction system and a diffusive correction system. The non linear components of the problem are concentrated in the prediction step, while the correction step leads to the solution of a linear system of the order of the number of computational cells. A semi-analytical procedure is applied for the solution of the prediction step. The discretized formulation of the governing equations allows to handle also wetting and drying processes without any additional specific treatment. Local energy dissipations, mainly the effect of vertical walls and hydraulic jumps, can be easily included in the model. Several numerical experiments have been carried out in order to test (1) the stability of the proposed model with regard to the size of the Courant number and to the mesh irregularity, (2) its computational performance, (3) the convergence order by means of mesh refinement. The model results are also compared with the results obtained by a fully dynamic model. Finally, the application to a real field case with a Venturi channel is presented.
NASA Astrophysics Data System (ADS)
Pantano, C.; Deiterding, R.; Hill, D. J.; Pullin, D. I.
2006-09-01
This paper describes a hybrid finite-difference method for the large-eddy simulation of compressible flows with low-numerical dissipation and structured adaptive mesh refinement (SAMR). A conservative flux-based approach is described with an explicit centered scheme used in turbulent flow regions while a weighted essentially non-oscillatory (WENO) scheme is employed to capture shocks. Three-dimensional numerical simulations of a Richtmyer-Meshkov instability are presented.
A mechanical model for deformable and mesh pattern wheel of lunar roving vehicle
NASA Astrophysics Data System (ADS)
Liang, Zhongchao; Wang, Yongfu; Chen, Gang (Sheng); Gao, Haibo
2015-12-01
As an indispensable tool for astronauts on lunar surface, the lunar roving vehicle (LRV) is of great significance for manned lunar exploration. An LRV moves on loose and soft lunar soil, so the mechanical property of its wheels directly affects the mobility performance. The wheels used for LRV have deformable and mesh pattern, therefore, the existing mechanical theory of vehicle wheel cannot be used directly for analyzing the property of LRV wheels. In this paper, a new mechanical model for LRV wheel is proposed. At first, a mechanical model for a rigid normal wheel is presented, which involves in multiple conventional parameters such as vertical load, tangential traction force, lateral force, and slip ratio. Secondly, six equivalent coefficients are introduced to amend the rigid normal wheel model to fit for the wheels with deformable and mesh-pattern in LRV application. Thirdly, the values of the six equivalent coefficients are identified by using experimental data obtained in an LRV's single wheel testing. Finally, the identified mechanical model for LRV's wheel with deformable and mesh pattern are further verified and validated by using additional experimental results.
Spatiotemporal processing of gated cardiac SPECT images using deformable mesh modeling
Brankov, Jovan G.; Yang Yongyi; Wernick, Miles N.
2005-09-15
In this paper we present a spatiotemporal processing approach, based on deformable mesh modeling, for noise reduction in gated cardiac single-photon emission computed tomography images. Because of the partial volume effect (PVE), clinical cardiac-gated perfusion images exhibit a phenomenon known as brightening--the myocardium appears to become brighter as the heart wall thickens. Although brightening is an artifact, it serves as an important diagnostic feature for assessment of wall thickening in clinical practice. Our proposed processing algorithm aims to preserve this important diagnostic feature while reducing the noise level in the images. The proposed algorithm is based on the use of a deformable mesh for modeling the cardiac motion in a gated cardiac sequence, based on which the images are processed by smoothing along space-time trajectories of object points while taking into account the PVE. Our experiments demonstrate that the proposed algorithm can yield significantly more-accurate results than several existing methods.
On Modeling and Analysis of MIMO Wireless Mesh Networks with Triangular Overlay Topology
Cao, Zhanmao; Wu, Chase Q.; Zhang, Yuanping; Shiva, Sajjan G.; Gu, Yi
2015-01-01
Multiple input multiple output (MIMO) wireless mesh networks (WMNs) aim to provide the last-mile broadband wireless access to the Internet. Along with the algorithmic development for WMNs, some fundamental mathematical problems also emerge in various aspects such as routing, scheduling, and channel assignment, all of which require an effective mathematical model and rigorous analysis of network properties. In this paper, we propose to employ Cartesian product of graphs (CPG) as a multichannel modeling approach and explore a set of unique properties of triangular WMNs. In each layer of CPG with a single channel, we design a node coordinate scheme thatmore » retains the symmetric property of triangular meshes and develop a function for the assignment of node identity numbers based on their coordinates. We also derive a necessary-sufficient condition for interference-free links and combinatorial formulas to determine the number of the shortest paths for channel realization in triangular WMNs.« less
Using adaptive-mesh refinement in SCFT simulations of surfactant adsorption
NASA Astrophysics Data System (ADS)
Sides, Scott; Kumar, Rajeev; Jamroz, Ben; Crockett, Robert; Pletzer, Alex
2013-03-01
Adsorption of surfactants at interfaces is relevant to many applications such as detergents, adhesives, emulsions and ferrofluids. Atomistic simulations of interface adsorption are challenging due to the difficulty of modeling the wide range of length scales in these problems: the thin interface region in equilibrium with a large bulk region that serves as a reservoir for the adsorbed species. Self-consistent field theory (SCFT) has been extremely useful for studying the morphologies of dense block copolymer melts. Field-theoretic simulations such as these are able to access large length and time scales that are difficult or impossible for particle-based simulations such as molecular dynamics. However, even SCFT methods can be difficult to apply to systems in which small spatial regions might require finer resolution than most of the simulation grid (eg. interface adsorption and confinement). We will present results on interface adsorption simulations using PolySwift++, an object-oriented, polymer SCFT simulation code aided by the Tech-X Chompst library that enables via block-structured AMR calculations with PETSc.
Merging for Particle-Mesh Complex Particle Kinetic Modeling of the Multiple Plasma Beams
NASA Technical Reports Server (NTRS)
Lipatov, Alexander S.
2011-01-01
We suggest a merging procedure for the Particle-Mesh Complex Particle Kinetic (PMCPK) method in case of inter-penetrating flow (multiple plasma beams). We examine the standard particle-in-cell (PIC) and the PMCPK methods in the case of particle acceleration by shock surfing for a wide range of the control numerical parameters. The plasma dynamics is described by a hybrid (particle-ion-fluid-electron) model. Note that one may need a mesh if modeling with the computation of an electromagnetic field. Our calculations use specified, time-independent electromagnetic fields for the shock, rather than self-consistently generated fields. While a particle-mesh method is a well-verified approach, the CPK method seems to be a good approach for multiscale modeling that includes multiple regions with various particle/fluid plasma behavior. However, the CPK method is still in need of a verification for studying the basic plasma phenomena: particle heating and acceleration by collisionless shocks, magnetic field reconnection, beam dynamics, etc.
NASA Astrophysics Data System (ADS)
Skutnik, Steven E.; Davis, David R.
2016-05-01
The use of passive gamma and neutron signatures from fission indicators is a common means of estimating used fuel burnup, enrichment, and cooling time. However, while characteristic fission product signatures such as 134Cs, 137Cs, 154Eu, and others are generally reliable estimators for used fuel burnup within the context where the assembly initial enrichment and the discharge time are known, in the absence of initial enrichment and/or cooling time information (such as when applying NDA measurements in a safeguards/verification context), these fission product indicators no longer yield a unique solution for assembly enrichment, burnup, and cooling time after discharge. Through the use of a new Mesh-Adaptive Direct Search (MADS) algorithm, it is possible to directly probe the shape of this "degeneracy space" characteristic of individual nuclides (and combinations thereof), both as a function of constrained parameters (such as the assembly irradiation history) and unconstrained parameters (e.g., the cooling time before measurement and the measurement precision for particular indicator nuclides). In doing so, this affords the identification of potential means of narrowing the uncertainty space of potential assembly enrichment, burnup, and cooling time combinations, thereby bounding estimates of assembly plutonium content. In particular, combinations of gamma-emitting nuclides with distinct half-lives (e.g., 134Cs with 137Cs and 154Eu) in conjunction with gross neutron counting (via 244Cm) are able to reasonably constrain the degeneracy space of possible solutions to a space small enough to perform useful discrimination and verification of fuel assemblies based on their irradiation history.
Moving Overlapping Grids with Adaptive Mesh Refinement for High-Speed Reactive and Non-reactive Flow
Henshaw, W D; Schwendeman, D W
2005-08-30
We consider the solution of the reactive and non-reactive Euler equations on two-dimensional domains that evolve in time. The domains are discretized using moving overlapping grids. In a typical grid construction, boundary-fitted grids are used to represent moving boundaries, and these grids overlap with stationary background Cartesian grids. Block-structured adaptive mesh refinement (AMR) is used to resolve fine-scale features in the flow such as shocks and detonations. Refinement grids are added to base-level grids according to an estimate of the error, and these refinement grids move with their corresponding base-level grids. The numerical approximation of the governing equations takes place in the parameter space of each component grid which is defined by a mapping from (fixed) parameter space to (moving) physical space. The mapped equations are solved numerically using a second-order extension of Godunov's method. The stiff source term in the reactive case is handled using a Runge-Kutta error-control scheme. We consider cases when the boundaries move according to a prescribed function of time and when the boundaries of embedded bodies move according to the surface stress exerted by the fluid. In the latter case, the Newton-Euler equations describe the motion of the center of mass of the each body and the rotation about it, and these equations are integrated numerically using a second-order predictor-corrector scheme. Numerical boundary conditions at slip walls are described, and numerical results are presented for both reactive and non-reactive flows in order to demonstrate the use and accuracy of the numerical approach.
NASA Astrophysics Data System (ADS)
Miyawaki, Shinjiro; Tawhai, Merryn H.; Hoffman, Eric A.; Lin, Ching-Long
2014-11-01
The authors have developed a method to automatically generate non-uniform CFD mesh for image-based human airway models. The sizes of generated tetrahedral elements vary in both radial and longitudinal directions to account for boundary layer and multiscale nature of pulmonary airflow. The proposed method takes advantage of our previously developed centerline-based geometry reconstruction method. In order to generate the mesh branch by branch in parallel, we used the open-source programs Gmsh and TetGen for surface and volume meshes, respectively. Both programs can specify element sizes by means of background mesh. The size of an arbitrary element in the domain is a function of wall distance, element size on the wall, and element size at the center of airway lumen. The element sizes on the wall are computed based on local flow rate and airway diameter. The total number of elements in the non-uniform mesh (10 M) was about half of that in the uniform mesh, although the computational time for the non-uniform mesh was about twice longer (170 min). The proposed method generates CFD meshes with fine elements near the wall and smooth variation of element size in longitudinal direction, which are required, e.g., for simulations with high flow rate. NIH Grants R01-HL094315, U01-HL114494, and S10-RR022421. Computer time provided by XSEDE.
Panaro, Fabrizio; Matos-Azevedo, Ana Maria; Fatas, José Antonio; Marin, Juan; Navarro, Francis; Zaragoza-Fernandez, Cristobal
2015-01-01
Purpose Conventional prostheses used for inguinal hernia repair are static and passive. This feasibility-study shows the features of a new 3D tension-free prosthesis in an experimental model. Methods This study was divided into two-phases: 1) aimed to test the physics intrinsic features and the anatomical adaptability of a new 3D designed mesh, and 2) aimed to evaluate the inflammatory reaction associated with different materials used. On phase-1 implantations were performed in pigs. During the first trial phase, the prostheses were also implanted on human cadavers. On phase-2, implantation was carried out on large swine. Follow-up was of 60-days, after which the animals were anaesthetized for laparoscopic assessment, and for sample collection of mesh implantation site for histological analysis. Results All animals showed good 3D mesh tolerance, and the follow-up period was uneventful. The laparoscopy showed no inflammatory lesions on the internal surface of the peritoneum. Macroscopic observation of implantation site revealed some local fibrosis and reorganization of tissue, no signs of infection, and no changes on original implant positioning. Histological analysis on phase-1 showed in most sample segments the deferent duct maintaining its central position and surrounded by vascular and nervous structures. On phase-2 differences in inflammatory lesion score could be found between subjects. Conclusions This new 3D mesh can be placed appropriately and from this preliminary animal study no untoward complications were noted over a 60 day period. PMID:27158482
Conversion of Component-Based Point Definition to VSP Model and Higher Order Meshing
NASA Technical Reports Server (NTRS)
Ordaz, Irian
2011-01-01
Vehicle Sketch Pad (VSP) has become a powerful conceptual and parametric geometry tool with numerous export capabilities for third-party analysis codes as well as robust surface meshing capabilities for computational fluid dynamics (CFD) analysis. However, a capability gap currently exists for reconstructing a fully parametric VSP model of a geometry generated by third-party software. A computer code called GEO2VSP has been developed to close this gap and to allow the integration of VSP into a closed-loop geometry design process with other third-party design tools. Furthermore, the automated CFD surface meshing capability of VSP are demonstrated for component-based point definition geometries in a conceptual analysis and design framework.
Landmark detection from 3D mesh facial models for image-based analysis of dysmorphology.
Chendeb, Marwa; Tortorici, Claudio; Al Muhairi, Hassan; Al Safar, Habiba; Linguraru, Marius; Werghi, Naoufel
2015-01-01
Facial landmark detection is a task of interest for facial dysmorphology, an important factor in the diagnosis of genetic conditions. In this paper, we propose a framework for feature points detection from 3D face images. The method is based on 3D Constrained Local Model (CLM) which learns both global variations in the 3D facial scan and local changes around every vertex landmark. Compared to state of the art methods our framework is distinguished by the following novel aspects: 1) It operates on facial surfaces, 2) It allows fusion of shape and color information on the mesh surface, 3) It introduces the use of LBP descriptors on the mesh. We showcase our landmarks detection framework on a set of scans including down syndrome and control cases. We also validate our method through a series of quantitative experiments conducted with the publicly available Bosphorus database. PMID:26736227
Parallel Adaptive Computation of Blood Flow in a 3D ``Whole'' Body Model
NASA Astrophysics Data System (ADS)
Zhou, M.; Figueroa, C. A.; Taylor, C. A.; Sahni, O.; Jansen, K. E.
2008-11-01
Accurate numerical simulations of vascular trauma require the consideration of a larger portion of the vasculature than previously considered, due to the systemic nature of the human body's response. A patient-specific 3D model composed of 78 connected arterial branches extending from the neck to the lower legs is constructed to effectively represent the entire body. Recently developed outflow boundary conditions that appropriately represent the downstream vasculature bed which is not included in the 3D computational domain are applied at 78 outlets. In this work, the pulsatile blood flow simulations are started on a fairly uniform, unstructured mesh that is subsequently adapted using a solution-based approach to efficiently resolve the flow features. The adapted mesh contains non-uniform, anisotropic elements resulting in resolution that conforms with the physical length scales present in the problem. The effects of the mesh resolution on the flow field are studied, specifically on relevant quantities of pressure, velocity and wall shear stress.
NASA Technical Reports Server (NTRS)
Wood, William A., III
2002-01-01
A multi-dimensional upwind fluctuation splitting scheme is developed and implemented for two-dimensional and axisymmetric formulations of the Navier-Stokes equations on unstructured meshes. Key features of the scheme are the compact stencil, full upwinding, and non-linear discretization which allow for second-order accuracy with enforced positivity. Throughout, the fluctuation splitting scheme is compared to a current state-of-the-art finite volume approach, a second-order, dual mesh upwind flux difference splitting scheme (DMFDSFV), and is shown to produce more accurate results using fewer computer resources for a wide range of test cases. A Blasius flat plate viscous validation case reveals a more accurate upsilon-velocity profile for fluctuation splitting, and the reduced artificial dissipation production is shown relative to DMFDSFV. Remarkably, the fluctuation splitting scheme shows grid converged skin friction coefficients with only five points in the boundary layer for this case. The second half of the report develops a local, compact, anisotropic unstructured mesh adaptation scheme in conjunction with the multi-dimensional upwind solver, exhibiting a characteristic alignment behavior for scalar problems. The adaptation strategy is extended to the two-dimensional and axisymmetric Navier-Stokes equations of motion through the concept of fluctuation minimization.
NASA Astrophysics Data System (ADS)
Schubert, Jochen E.; Sanders, Brett F.; Smith, Martin J.; Wright, Nigel G.
2008-12-01
Urban flood inundation modeling with a hydrodynamic flow solver is addressed in this paper, focusing on strategies to effectively integrate geospatial data for unstructured mesh generation, building representation and flow resistance parameterization. Data considered include Light Detection and Ranging (LiDAR) terrain height surveys, aerial imagery and vector datasets such as building footprint polygons. First, a unstructured mesh-generation technique we term the building-hole method (BH) is developed whereby building footprint data define interior domain boundaries or mesh holes. A wall boundary condition depicts the impact of buildings on flood hydrodynamics. BH provides an alternative to the more commonly used method of raising terrain heights where buildings coincide with the mesh. We term this the building-block method (BB). Application of BH and BB to a flooding site in Glasgow, Scotland identifies a number of tradeoffs to consider at resolutions ranging from 1 to 5 m. At fine resolution, BH is shown to be similarly accurate but execute faster than BB. And at coarse resolution, BH is shown to preserve the geometry of buildings and maintain better accuracy than BB, but requires a longer run time. Meshes that ignore buildings completely ( no-building method or NB) also support surprisingly good flood inundation predictions at coarse resolution compared to BH and BB. NB also supports faster execution times than BH at coarse resolution because the latter uses localized refinements that mandate a greater number of computational cells. However, with mesh refinement, NB converges to a different (and presumably less-accurate) solution compared to BH and BB. Using the same test conditions, Hunter et al. [Hunter NM, Bates PD, Neelz S, Pender G, Villanueva I, Wright NG, Liang D, et al. Benchmarking 2D hydraulic models for urban flood simulations. ICE J Water Manage 2008;161(1):13-30] compared the performance of dynamic-wave and diffusive-wave models and reported that
NASA Astrophysics Data System (ADS)
Skamarock, W. C.
2015-12-01
One of the major problems in atmospheric model applications is the representation of deep convection within the models; explicit simulation of deep convection on fine meshes performs much better than sub-grid parameterized deep convection on coarse meshes. Unfortunately, the high cost of explicit convective simulation has meant it has only been used to down-scale global simulations in weather prediction and regional climate applications, typically using traditional one-way interactive nesting technology. We have been performing real-time weather forecast tests using a global non-hydrostatic atmospheric model (the Model for Prediction Across Scales, MPAS) that employs a variable-resolution unstructured Voronoi horizontal mesh (nominally hexagons) to span hydrostatic to nonhydrostatic scales. The smoothly varying Voronoi mesh eliminates many downscaling problems encountered using traditional one- or two-way grid nesting. Our test weather forecasts cover two periods - the 2015 Spring Forecast Experiment conducted at the NOAA Storm Prediction Center during the month of May in which we used a 50-3 km mesh, and the PECAN field program examining nocturnal convection over the US during the months of June and July in which we used a 15-3 km mesh. An important aspect of this modeling system is that the model physics be scale-aware, particularly the deep convection parameterization. These MPAS simulations employ the Grell-Freitas scale-aware convection scheme. Our test forecasts show that the scheme produces a gradual transition in the deep convection, from the deep unstable convection being handled entirely by the convection scheme on the coarse mesh regions (dx > 15 km), to the deep convection being almost entirely explicit on the 3 km NA region of the meshes. We will present results illustrating the performance of critical aspects of the MPAS model in these tests.
Fabrication of compliant hybrid grafts supported with elastomeric meshes.
Kobashi, T; Matsuda, T
1999-01-01
We devised tubular hybrid medial tissues with mechanical properties similar to those of native arteries, which were composed of bovine smooth muscle cells (SMCs) and type I collagen with minimal reinforcement with knitted fabric meshes made of synthetic elastomers. Three hybrid medial tissue models that incorporated segmented polyester (mesh A) or polyurethane-nylon (mesh B) meshes were designed: the inner, sandwich, and wrapping models. Hybrid medial tissues were prepared by pouring a cold mixed solution of SMCs and collagen into a tubular glass mold consisting of an inner mandrel and an outer sheath and subsequent thermal gelation, followed by further culture for 7 days. For the inner model, the mandrel was wrapped with a mesh. For the sandwich model, a cylindrically shaped mesh was incorporated into a space between the mandrel and the sheath. The wrapping model was prepared by wrapping a 7-day-incubated nonmesh gel with a mesh. The inner diameter was 3 mm, irrespective of the model, and the length was 2.5-4.0 cm, depending on the model. The intraluminal pressure-external diameter relationship showed that nonmesh and inner models had a very low burst strength below 50 mmHg, while the sandwich model ruptured at around 110-120 mmHg; no rupturing below 240 mmHg was observed for the wrapping model, regardless of the type of mesh used. Compliance values of wrapping and sandwich models were close to those of native arteries. Pressure-dependent distensibility characteristics similar to native arteries were observed for a mesh A wrapping model, whereas a mesh B wrapping model expanded almost linearly as intraluminal pressure increased, which appeared to be due to elasticity of the incorporated mesh. Thus, design criteria for hybrid vascular grafts with appropriate biomechanical matching with host arteries were established. Such hybrid grafts may be mechanically adapted in an arterial system. PMID:10580342
Application of Open Loop H-Adaptation to an Unstructured Grid Tidal Flat Model
NASA Astrophysics Data System (ADS)
Cowles, G. W.
2008-12-01
The complex topology of tidal flats presents a challenge to coastal ocean models. Recently, several models have been developed employing unstructured grids, which can provide the flexibility in mesh resolution required to resolve the complex bathymetry and coastline. However, the distribution of element size in the initial mesh can be somewhat arbitrary, and is in general the product of the operator tailoring the resolution to the underlying bathymetry and regions of interest. In this work, the flow solution from an idealized tidal flat application is used to drive an open loop h-adaptation of the mesh. The model used for this work is the Finite Volume Coastal Ocean Model (FVCOM), an open source, terrain following model. A background length scale distribution derived from model output is used to generate a new initial mesh for the model run, thus defining an iteration of the procedure. Several metrics for computing the background length scale will be examined. These include direct estimation of spatial discretization error using Richardson's extrapolation from a sequence of meshes as well as heuristics derived from gradients in the primitive variables. Examination of grid independence, computational efficiency, and performance of the scheme for idealized tidal flats with inclusion of morphodynamics will be discussed.
Koch, M.
1995-12-31
A new mesh-adaptive 1D collocation technique has been developed to efficiently solve transient advection-dominated transport problems in porous media that are governed by a hyperbolic/parabolic (singularly perturbed) PDE. After spatial discretization a singularly perturbed ODE is obtained which is solved by a modification of the COLNEW ODE-collocation code. The latter also contains an adaptive mesh procedure that has been enhanced here to resolve linear and nonlinear transport flow problems with steep fronts where regular FD and FE methods often fail. An implicit first-order backward Euler and a third-order Taylor-Donea technique are employed for the time integration. Numerical simulations on a variety of high Peclet-number transport phenomena as they occur in realistic porous media flow situations are presented. Examples include classical linear advection-diffusion, nonlinear adsorption, two-phase Buckley-Leverett flow without and with capillary forces (Rapoport-Leas equation) and Burgers` equation for inviscid fluid flow. In most of these examples sharp fronts and/or shocks develop which are resolved in an oscillation-free manner by the present adaptive collocation method. The backward Euler method has some amount of numerical dissipation is observed when the time-steps are too large. The third-order Taylor-Donea technique is less dissipative but is more prone to numerical oscillations. The simulations show that for the efficient solution of nonlinear singularly perturbed PDE`s governing flow transport a careful balance must be struck between the optimal mesh adaptation, the nonlinear iteration method and the time-stepping procedure. More theoretical research is needed with this regard.
Raut, Samarth S; Liu, Peng; Finol, Ender A
2015-07-16
In this work, we present a computationally efficient image-derived volume mesh generation approach for vasculatures that implements spatially varying patient-specific wall thickness with a novel inward extrusion of the wall surface mesh. Multi-domain vascular meshes with arbitrary numbers, locations, and patterns of both iliac bifurcations and thrombi can be obtained without the need to specify features or landmark points as input. In addition, the mesh output is coordinate-frame independent and independent of the image grid resolution with high dimensional accuracy and mesh quality, devoid of errors typically found in off-the-shelf image-based model generation workflows. The absence of deformable template models or Cartesian grid-based methods enables the present approach to be sufficiently robust to handle aneurysmatic geometries with highly irregular shapes, arterial branches nearly parallel to the image plane, and variable wall thickness. The assessment of the methodology was based on i) estimation of the surface reconstruction accuracy, ii) validation of the output mesh using an aneurysm phantom, and iii) benchmarking the volume mesh quality against other frameworks. For the phantom image dataset (pixel size 0.105 mm; slice spacing 0.7 mm; and mean wall thickness 1.401±0.120 mm), the average wall thickness in the mesh was 1.459±0.123 mm. The absolute error in average wall thickness was 0.060±0.036 mm, or about 8.6% of the largest image grid spacing (0.7 mm) and 4.36% of the actual mean wall thickness. Mesh quality metrics and the ability to reproduce regional variations of wall thickness were found superior to similar alternative frameworks. PMID:25976018
Raut, Samarth S.; Liu, Peng; Finol, Ender A.
2015-01-01
In this work, we present a computationally efficient image-derived volume mesh generation approach for vasculatures that implements spatially varying patient-specific wall thickness with a novel inward extrusion of the wall surface mesh. Multi-domain vascular meshes with arbitrary numbers, locations, and patterns of both iliac bifurcations and thrombi can be obtained without the need to specify features or landmark points as input. In addition, the mesh output is coordinate-frame independent and independent of the image grid resolution with high dimensional accuracy and mesh quality, devoid of errors typically found in off-the-shelf image-based model generation workflows. The absence of deformable template models or Cartesian grid-based methods enables the present approach to be robust by handling aneurysmatic geometries with highly irregular shapes, arterial branches nearly parallel to the image plane, and variable wall thickness. The assessment of the methodology was based on i) estimation of the surface reconstruction accuracy, ii) validation of the output mesh using an aneurysm phantom, and iii) benchmarking the volume mesh quality against other frameworks. For the phantom image dataset (pixel size 0.105 mm; slice spacing 0.7 mm; mean wall thickness 1.401 ± 0.120 mm), the average wall thickness in the mesh was 1.459 ± 0.123 mm. The absolute error in average wall thickness was 0.060 ± 0.036 mm, or about 8.6% of the largest image grid spacing (0.7 mm) and 4.36% of the actual mean wall thickness. Mesh quality metrics and the ability to reproduce regional variations of wall thickness were found superior to similar alternative frameworks. PMID:25976018
Model for adaptive multimedia services
NASA Astrophysics Data System (ADS)
Forstadius, Jari; Ala-Kurikka, Jussi; Koivisto, Antti T.; Sauvola, Jaakko J.
2001-11-01
Development towards high-bandwidth wireless devices that are capable of processing complex, streaming multimedia is enabling a new breed of network-based media services. Coping with the diversity of network and device capabilities requires services to be flexible and able to adapt to the needs and limitations of the environment at hand. Before efficient deployment, multi-platform services require additional issues to be considered, e.g. content handling, digital rights management, adaptability of content, user profiling, provisioning, and the available access methods. The key issue is how the content and the service is being modelled and stored for inauguration. We propose a new service content model based on persistent media objects able to store and manage XHTML-based multimedia services. In our approach, media, content summaries, and other meta-information are stored within media objects that can be queried from the object database. The content of the media objects can also specify queries to the database and links to other media objects. The final presentation is created dynamically according to the service request and user profiles. Our approach allows for dynamic updating of the service database together with user group management, and provides a method for notifying the registered users by different smart messaging methods, e.g. via e-mail or a SMS message. The model is demonstrated with an 'ice-hockey service' running in our platform called Princess. The service also utilizes SMIL and key frame techniques for the video representation.
NASA Astrophysics Data System (ADS)
Gill, Stuart P. D.; Knebe, Alexander; Gibson, Brad K.; Flynn, Chris; Ibata, Rodrigo A.; Lewis, Geraint F.
2003-04-01
An adaptive multi grid approach to simulating the formation of structure from collisionless dark matter is described. MLAPM (Multi-Level Adaptive Particle Mesh) is one of the most efficient serial codes available on the cosmological "market" today. As part of Swinburne University's role in the development of the Square Kilometer Array, we are implementing hydrodynamics, feedback, and radiative transfer within the MLAPM adaptive mesh, in order to simulate baryonic processes relevant to the interstellar and intergalactic media at high redshift. We will outline our progress to date in applying the existing MLAPM to a study of the decay of satellite galaxies within massive host potentials.
Tang, Hua; Xu, Zhifei; Qin, Xiong; Wu, Bin; Wu, Lihui; Zhao, XueWei; Li, Yulin
2009-07-01
Extensive chest wall defect reconstruction remains a challenging problem for surgeons. In the past several years, little progress has been made in this area. In this study, a biodegradable polydioxanone (PDO) mesh and demineralized bone matrix (DBM) seeded with osteogenically induced bone marrow stromal cells (BMSCs) were used to reconstruct a 6 cm x 5.5 cm chest wall defect. Four experimental groups were evaluated (n=6 per group): polydioxanone (PDO) mesh/DBMs/BMSCs group, polydioxanone (PDO) mesh/DBMs group, polydioxanone (PDO) mesh group, and a blank group (no materials) in a canine model. All the animals survived except those in the blank group. In all groups receiving biomaterial implants, the polydioxanone (PDO) mesh completely degraded at 24 weeks and was replaced by fibrous tissue with thickness close to that of the normal intercostal tissue (P>0.05). In the polydioxanone (PDO) mesh/DBMs/BMSCs group, new bone formation and bone-union were observed by radiographic and histological examination. More importantly, the reconstructed rib could maintain its original radian and achieve satisfactory biomechanics close to normal ribs in terms of bending stress (P>0.05). However, in the other two groups, fibrous tissue was observed in the defect and junctions, and the reconstructed ribs were easily distorted under an outer force. Based on these results, a surgical approach utilizing biodegradable polydioxanone (PDO) mesh in combination with DBMs and BMSCs could repair the chest wall defect not only in function but also in structure. PMID:19233465
Pavarino, E.; Neves, L. A.; Machado, J. M.; de Godoy, M. F.; Shiyou, Y.; Momente, J. C.; Zafalon, G. F. D.; Pinto, A. R.; Valêncio, C. R.
2013-01-01
The Finite Element Method is a well-known technique, being extensively applied in different areas. Studies using the Finite Element Method (FEM) are targeted to improve cardiac ablation procedures. For such simulations, the finite element meshes should consider the size and histological features of the target structures. However, it is possible to verify that some methods or tools used to generate meshes of human body structures are still limited, due to nondetailed models, nontrivial preprocessing, or mainly limitation in the use condition. In this paper, alternatives are demonstrated to solid modeling and automatic generation of highly refined tetrahedral meshes, with quality compatible with other studies focused on mesh generation. The innovations presented here are strategies to integrate Open Source Software (OSS). The chosen techniques and strategies are presented and discussed, considering cardiac structures as a first application context. PMID:23762031
NASA Astrophysics Data System (ADS)
Candy, A. S.; Avdis, A.; Hill, J.; Gorman, G. J.; Piggott, M. D.
2014-09-01
Computational simulations of physical phenomena rely on an accurate discretisation of the model domain. Numerical models have increased in sophistication to a level where it is possible to support terrain-following boundaries that conform accurately to real physical interfaces, and resolve a multiscale of spatial resolutions. Whilst simulation codes are maturing in this area, pre-processing tools have not developed significantly enough to competently initialise these problems in a rigorous, efficient and recomputable manner. In the relatively disjoint field of Geographic Information Systems (GIS) however, techniques and tools for mapping and analysis of geographical data have matured significantly. If data provenance and recomputability are to be achieved, the manipulation and agglomeration of data in the pre-processing of numerical simulation initialisation data for geophysical models should be integrated into GIS. A new approach to the discretisation of geophysical domains is presented, and introduced with a verified implementation. This brings together the technologies of geospatial analysis, meshing and numerical simulation models. This platform enables us to combine and build up features, quickly drafting and updating mesh descriptions with the rigour that established GIS tools provide. This, combined with the systematic workflow, supports a strong provenance for model initialisation and encourages the convergence of standards.
NASA Astrophysics Data System (ADS)
Lenkiewicz, Przemyslaw; Pereira, Manuela; Freire, Mário M.; Fernandes, José
2013-12-01
In this article, we propose a novel image segmentation method called the whole mesh deformation (WMD) model, which aims at addressing the problems of modern medical imaging. Such problems have raised from the combination of several factors: (1) significant growth of medical image volumes sizes due to increasing capabilities of medical acquisition devices; (2) the will to increase the complexity of image processing algorithms in order to explore new functionality; (3) change in processor development and turn towards multi processing units instead of growing bus speeds and the number of operations per second of a single processing unit. Our solution is based on the concept of deformable models and is characterized by a very effective and precise segmentation capability. The proposed WMD model uses a volumetric mesh instead of a contour or a surface to represent the segmented shapes of interest, which allows exploiting more information in the image and obtaining results in shorter times, independently of image contents. The model also offers a good ability for topology changes and allows effective parallelization of workflow, which makes it a very good choice for large datasets. We present a precise model description, followed by experiments on artificial images and real medical data.
An Interpreted Language and System for the Visualization of Unstructured Meshes
NASA Technical Reports Server (NTRS)
Moran, Patrick J.; Gerald-Yamasaki, Michael (Technical Monitor)
1998-01-01
We present an interpreted language and system supporting the visualization of unstructured meshes and the manipulation of shapes defined in terms of mesh subsets. The language features primitives inspired by geometric modeling, mathematical morphology and algebraic topology. The adaptation of the topology ideas to an interpreted environment, along with support for programming constructs such, as user function definition, provide a flexible system for analyzing a mesh and for calculating with shapes defined in terms of the mesh. We present results demonstrating some of the capabilities of the language, based on an implementation called the Shape Calculator, for tetrahedral meshes in R^3.
TRANSL8GDECIM8. Data Translation and Filtering for Large 3D Triangle Mesh Models
Janucik, F.X.; Ross, D.M.
1993-09-01
The TRANSL8GDECIM8 system consists of two programs: TRANSL8G and DECIM8. The TRANSL8G program facilitates the interchange, topology generation, error checking, and enhancement of large 3D triangle meshes. Such data is frequently used to represent conceptual designs, scientific visualization volume modeling, or discrete sample data. Interchange is provided between several popular commercial and defacto standard geometry formats. Error checking is included to identify duplicate and zero area triangles. Model enhancement features include common vertex joining, consistent triangle vertex ordering, vertex normal vector averaging, and triangle strip generation. Many of the traditional O(n squared) algorithms required to provide the above features have been recast and are O(n) which support large mesh sizes. The DECIM8 program is based on a data filter algorithm that significantly reduces the number of triangles required to represent three dimensional (3D) models of geometry, scientific visualization results, and discretely sampled data. The algorithm uses a combined incremental and iterative strategy. It eliminates local patches of triangles whose geometries are not appreciably different and replaces them with fewer larger triangles. The algorithm has been used to reduce triangles in large conceptual design models to facilitate virtual walk throughs and to enable interactive viewing of large 3D iso-surface volume visualizations.
Simple method for model reference adaptive control
NASA Technical Reports Server (NTRS)
Seraji, H.
1989-01-01
A simple method is presented for combined signal synthesis and parameter adaptation within the framework of model reference adaptive control theory. The results are obtained using a simple derivation based on an improved Liapunov function.
Generation of the 30 M-Mesh Global Digital Surface Model by Alos Prism
NASA Astrophysics Data System (ADS)
Tadono, T.; Nagai, H.; Ishida, H.; Oda, F.; Naito, S.; Minakawa, K.; Iwamoto, H.
2016-06-01
Topographical information is fundamental to many geo-spatial related information and applications on Earth. Remote sensing satellites have the advantage in such fields because they are capable of global observation and repeatedly. Several satellite-based digital elevation datasets were provided to examine global terrains with medium resolutions e.g. the Shuttle Radar Topography Mission (SRTM), the global digital elevation model by the Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER GDEM). A new global digital surface model (DSM) dataset using the archived data of the Panchromatic Remote-sensing Instrument for Stereo Mapping (PRISM) onboard the Advanced Land Observing Satellite (ALOS, nicknamed "Daichi") has been completed on March 2016 by Japan Aerospace Exploration Agency (JAXA) collaborating with NTT DATA Corp. and Remote Sensing Technology Center, Japan. This project is called "ALOS World 3D" (AW3D), and its dataset consists of the global DSM dataset with 0.15 arcsec. pixel spacing (approx. 5 m mesh) and ortho-rectified PRISM image with 2.5 m resolution. JAXA is also processing the global DSM with 1 arcsec. spacing (approx. 30 m mesh) based on the AW3D DSM dataset, and partially releasing it free of charge, which calls "ALOS World 3D 30 m mesh" (AW3D30). The global AW3D30 dataset will be released on May 2016. This paper describes the processing status, a preliminary validation result of the AW3D30 DSM dataset, and its public release status. As a summary of the preliminary validation of AW3D30 DSM, 4.40 m (RMSE) of the height accuracy of the dataset was confirmed using 5,121 independent check points distributed in the world.
An adaptive multigrid model for hurricane track prediction
NASA Technical Reports Server (NTRS)
Fulton, Scott R.
1993-01-01
This paper describes a simple numerical model for hurricane track prediction which uses a multigrid method to adapt the model resolution as the vortex moves. The model is based on the modified barotropic vorticity equation, discretized in space by conservative finite differences and in time by a Runge-Kutta scheme. A multigrid method is used to solve an elliptic problem for the streamfunction at each time step. Nonuniform resolution is obtained by superimposing uniform grids of different spatial extent; these grids move with the vortex as it moves. Preliminary numerical results indicate that the local mesh refinement allows accurate prediction of the hurricane track with substantially less computer time than required on a single uniform grid.
Documentation for MeshKit - Reactor Geometry (&mesh) Generator
Jain, Rajeev; Mahadevan, Vijay
2015-09-30
This report gives documentation for using MeshKit’s Reactor Geometry (and mesh) Generator (RGG) GUI and also briefly documents other algorithms and tools available in MeshKit. RGG is a program designed to aid in modeling and meshing of complex/large hexagonal and rectilinear reactor cores. RGG uses Argonne’s SIGMA interfaces, Qt and VTK to produce an intuitive user interface. By integrating a 3D view of the reactor with the meshing tools and combining them into one user interface, RGG streamlines the task of preparing a simulation mesh and enables real-time feedback that reduces accidental scripting mistakes that could waste hours of meshing. RGG interfaces with MeshKit tools to consolidate the meshing process, meaning that going from model to mesh is as easy as a button click. This report is designed to explain RGG v 2.0 interface and provide users with the knowledge and skills to pilot RGG successfully. Brief documentation of MeshKit source code, tools and other algorithms available are also presented for developers to extend and add new algorithms to MeshKit. RGG tools work in serial and parallel and have been used to model complex reactor core models consisting of conical pins, load pads, several thousands of axially varying material properties of instrumentation pins and other interstices meshes.
A Bayesian Model of Sensory Adaptation
Sato, Yoshiyuki; Aihara, Kazuyuki
2011-01-01
Recent studies reported two opposite types of adaptation in temporal perception. Here, we propose a Bayesian model of sensory adaptation that exhibits both types of adaptation. We regard adaptation as the adaptive updating of estimations of time-evolving variables, which determine the mean value of the likelihood function and that of the prior distribution in a Bayesian model of temporal perception. On the basis of certain assumptions, we can analytically determine the mean behavior in our model and identify the parameters that determine the type of adaptation that actually occurs. The results of our model suggest that we can control the type of adaptation by controlling the statistical properties of the stimuli presented. PMID:21541346
Lafarge, Florent; Keriven, Renaud; Brédif, Mathieu
2010-07-01
We propose an original hybrid modeling process of urban scenes that represents 3-D models as a combination of mesh-based surfaces and geometric 3-D-primitives. Meshes describe details such as ornaments and statues, whereas 3-D-primitives code for regular shapes such as walls and columns. Starting from an 3-D-surface obtained by multiview stereo techniques, these primitives are inserted into the surface after being detected. This strategy allows the introduction of semantic knowledge, the simplification of the modeling, and even correction of errors generated by the acquisition process. We design a hierarchical approach exploring different scales of an observed scene. Each level consists first in segmenting the surface using a multilabel energy model optimized by -expansion and then in fitting 3-D-primitives such as planes, cylinders or tori on the obtained partition where relevant. Experiments on real meshes, depth maps and synthetic surfaces show good potential for the proposed approach. PMID:20236893
Greenough, Jeffrey A.; de Supinski, Bronis R.; Yates, Robert K.; Rendleman, Charles A.; Skinner, David; Beckner, Vince; Lijewski, Mike; Bell, John; Sexton, James C.
2005-04-25
We describe the performance of the block-structured Adaptive Mesh Refinement (AMR) code Raptor on the 32k node IBM BlueGene/L computer. This machine represents a significant step forward towards petascale computing. As such, it presents Raptor with many challenges for utilizing the hardware efficiently. In terms of performance, Raptor shows excellent weak and strong scaling when running in single level mode (no adaptivity). Hardware performance monitors show Raptor achieves an aggregate performance of 3:0 Tflops in the main integration kernel on the 32k system. Results from preliminary AMR runs on a prototype astrophysical problem demonstrate the efficiency of the current software when running at large scale. The BG/L system is enabling a physics problem to be considered that represents a factor of 64 increase in overall size compared to the largest ones of this type computed to date. Finally, we provide a description of the development work currently underway to address our inefficiencies.
Computerized Adaptive Testing under Nonparametric IRT Models
ERIC Educational Resources Information Center
Xu, Xueli; Douglas, Jeff
2006-01-01
Nonparametric item response models have been developed as alternatives to the relatively inflexible parametric item response models. An open question is whether it is possible and practical to administer computerized adaptive testing with nonparametric models. This paper explores the possibility of computerized adaptive testing when using…
An Adaptive Code for Radial Stellar Model Pulsations
NASA Astrophysics Data System (ADS)
Buchler, J. Robert; Kolláth, Zoltán; Marom, Ariel
1997-09-01
We describe an implicit 1-D adaptive mesh hydrodynamics code that is specially tailored for radial stellar pulsations. In the Lagrangian limit the code reduces to the well tested Fraley scheme. The code has the useful feature that unwanted, long lasting transients can be avoided by smoothly switching on the adaptive mesh features starting from the Lagrangean code. Thus, a limit cycle pulsation that can readily be computed with the relaxation method of Stellingwerf will converge in a few tens of pulsation cycles when put into the adaptive mesh code. The code has been checked with two shock problems, viz. Noh and Sedov, for which analytical solutions are known, and it has been found to be both accurate and stable. Superior results were obtained through the solution of the total energy (gravitational + kinetic + internal) equation rather than that of the internal energy only.
Toward An Unstructured Mesh Database
NASA Astrophysics Data System (ADS)
Rezaei Mahdiraji, Alireza; Baumann, Peter Peter
2014-05-01
Unstructured meshes are used in several application domains such as earth sciences (e.g., seismology), medicine, oceanography, cli- mate modeling, GIS as approximate representations of physical objects. Meshes subdivide a domain into smaller geometric elements (called cells) which are glued together by incidence relationships. The subdivision of a domain allows computational manipulation of complicated physical structures. For instance, seismologists model earthquakes using elastic wave propagation solvers on hexahedral meshes. The hexahedral con- tains several hundred millions of grid points and millions of hexahedral cells. Each vertex node in the hexahedrals stores a multitude of data fields. To run simulation on such meshes, one needs to iterate over all the cells, iterate over incident cells to a given cell, retrieve coordinates of cells, assign data values to cells, etc. Although meshes are used in many application domains, to the best of our knowledge there is no database vendor that support unstructured mesh features. Currently, the main tool for querying and manipulating unstructured meshes are mesh libraries, e.g., CGAL and GRAL. Mesh li- braries are dedicated libraries which includes mesh algorithms and can be run on mesh representations. The libraries do not scale with dataset size, do not have declarative query language, and need deep C++ knowledge for query implementations. Furthermore, due to high coupling between the implementations and input file structure, the implementations are less reusable and costly to maintain. A dedicated mesh database offers the following advantages: 1) declarative querying, 2) ease of maintenance, 3) hiding mesh storage structure from applications, and 4) transparent query optimization. To design a mesh database, the first challenge is to define a suitable generic data model for unstructured meshes. We proposed ImG-Complexes data model as a generic topological mesh data model which extends incidence graph model to multi
NASA Technical Reports Server (NTRS)
Turon, A.; Davila, C. G.; Camanho, P. P.; Costa, J.
2007-01-01
This paper presents a methodology to determine the parameters to be used in the constitutive equations of Cohesive Zone Models employed in the simulation of delamination in composite materials by means of decohesion finite elements. A closed-form expression is developed to define the stiffness of the cohesive layer. A novel procedure that allows the use of coarser meshes of decohesion elements in large-scale computations is also proposed. The procedure ensures that the energy dissipated by the fracture process is computed correctly. It is shown that coarse-meshed models defined using the approach proposed here yield the same results as the models with finer meshes normally used for the simulation of fracture processes.
Generating meshes for finite-difference analysis using a solid modeler
NASA Astrophysics Data System (ADS)
Laguna, G. W.; White, W. T.; Cabral, B. K.
1987-09-01
One tool used by the Engineering Research Division of LLNL to help analyze the behavior of electronic systems in hostile environments is 3D finite-difference time-domain (FDTD) computation. FDTD codes solve Maxwell's equations,the differential equations of electromagnetism, on a uniform lattice of points. It is this uniform lattice, or mesh, that distinguishes finite-difference codes from other codes. The simple mesh makes FDTD codes computationally more efficient than other codes, which enables them to run larger problems and to run faster (up to thirty times faster than finite-element codes, for example). Therefore, within the Engineering Department at LLNL, Electronics Engineering (EE) has initiated a project to develop a mesh generator that will provide meshes suitable for FDTD analysis. This report describes the results of the first year of EE's FDTD Mesh Generation Project. During this year a preliminary version of an automated mesh generator was built and used to create a mesh of an object of interest to the High-Power Microwave Program, namely an electrically detonatable land mine. The code was verified by meshing basic solids such as spheres and cylinders. Because of the design of the code, there is no software limitation to the size of meshes that can be accommodated. The algorithm with a mesh space of approximately 500,000 cells has been demonstrated. The mesh generator can detect certain objects with walls that are thinner than the width of a cell. The code has internal graphics for viewing objects as they appear prior to being converted to a finite-difference representation. Additionally, via data files, the code is coupled to two external graphics packages for visually checking the meshes, namely TAURUS on the Cray and a new code, IMAGE, on the Silicon Graphics IRIS workstation.
Generating meshes for finite-difference analysis using a solid modeler
Laguna, G.W.; White, W.T.; Cabral, B.K.
1987-09-01
One tool used by the Engineering Research Division of LLNL to help analyze the behavior of electronic systems in hostile environments is 3D finite-difference time-domain (FDTD) computation. FDTD codes solve Maxwell's equations,the differential equations of electromagnetism, on a uniform lattice of points. It is this uniform lattice, or ''mesh,'' that distinguishes finite-difference codes from other codes. The simple mesh makes FDTD codes computationally more efficient than other codes, which enables them to run larger problems and to run faster (up to thirty times faster than finite-element codes, for example). Therefore, within the Engineering Department at LLNL, Electronics Engineering (EE) has initiated a project to develop a mesh generator that will provide meshes suitable for FDTD analysis. This report describes the results of the first year of EE's FDTD Mesh Generation Project. During this year a preliminary version of an automated mesh generator was built and used to create a mesh of an object of interest to the High-Power Microwave Program, namely an electrically detonatable land mine. The code was verified by meshing basic solids such as spheres and cylinders. Because of the design of the code, there is no software limitation to the size of meshes that can be accommodated. The algorithm with a mesh space of approximately 500,000 cells has been demonstrated. The mesh generator can detect certain objects with walls that are thinner than the width of a cell. The code has internal graphics for viewing objects as they appear prior to being converted to a finite-difference representation. Additionally, via data files, the code is coupled to two external graphics packages for visually checking the meshes, namely TAURUS on the Cray and a new code, IMAGE, on the Silicon Graphics IRIS workstation.
An Adaptive Critic Approach to Reference Model Adaptation
NASA Technical Reports Server (NTRS)
Krishnakumar, K.; Limes, G.; Gundy-Burlet, K.; Bryant, D.
2003-01-01
Neural networks have been successfully used for implementing control architectures for different applications. In this work, we examine a neural network augmented adaptive critic as a Level 2 intelligent controller for a C- 17 aircraft. This intelligent control architecture utilizes an adaptive critic to tune the parameters of a reference model, which is then used to define the angular rate command for a Level 1 intelligent controller. The present architecture is implemented on a high-fidelity non-linear model of a C-17 aircraft. The goal of this research is to improve the performance of the C-17 under degraded conditions such as control failures and battle damage. Pilot ratings using a motion based simulation facility are included in this paper. The benefits of using an adaptive critic are documented using time response comparisons for severe damage situations.
NASA Astrophysics Data System (ADS)
Gansen, A.; Hachemi, M. El; Belouettar, S.; Hassan, O.; Morgan, K.
2016-09-01
The standard Yee algorithm is widely used in computational electromagnetics because of its simplicity and divergence free nature. A generalization of the classical Yee scheme to 3D unstructured meshes is adopted, based on the use of a Delaunay primal mesh and its high quality Voronoi dual. This allows the problem of accuracy losses, which are normally associated with the use of the standard Yee scheme and a staircased representation of curved material interfaces, to be circumvented. The 3D dual mesh leapfrog-scheme which is presented has the ability to model both electric and magnetic anisotropic lossy materials. This approach enables the modelling of problems, of current practical interest, involving structured composites and metamaterials.
Towards Spherical Mesh Gravity and Magnetic Modelling in an HPC Environment
NASA Astrophysics Data System (ADS)
Lane, R. J.; Brodie, R. C.; de Hoog, M.; Navin, J.; Chen, C.; Du, J.; Liang, Q.; Wang, H.; Li, Y.
2013-12-01
Staff at Geoscience Australia (GA), Australia's Commonwealth Government geoscientific agency, have routinely performed 3D gravity and magnetic modelling as part of geoscience investigations. For this work, we have used software programs that have been based on a Cartesian mesh spatial framework. These programs have come as executable files that were compiled to operate in a Windows environment on single core personal computers (PCs). To cope with models with higher resolution and larger extents, we developed an approach whereby a large problem could be broken down into a number of overlapping smaller models (';tiles') that could be modelled separately, with the results combined back into a single output model. To speed up the processing, we established a Condor distributed network from existing desktop PCs. A number of factors have caused us to consider a new approach to this modelling work. The drivers for change include; 1) models with very large lateral extents where the effects of Earth curvature are a consideration, 2) a desire to ensure that the modelling of separate regions is carried out in a consistent and managed fashion, 3) migration of scientific computing to off-site High Performance Computing (HPC) facilities, and 4) development of virtual globe environments for integration and visualization of 3D spatial objects. Some of the more surprising realizations to emerge have been that; 1) there aren't any readily available commercial software packages for modelling gravity and magnetic data in a spherical mesh spatial framework, 2) there are many different types of HPC environments, 3) no two HPC environments are the same, and 4) the most common virtual globe environment (i.e., Google Earth) doesn't allow spatial objects to be displayed below the topographic/bathymetric surface. Our response has been to do the following; 1) form a collaborative partnership with researchers at the Colorado School of Mines (CSM) and the China University of Geosciences (CUG
Multigrid solution of compressible turbulent flow on unstructured meshes using a two-equation model
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.; Martinelli, L.
1991-01-01
The system of equations consisting of the full Navier-Stokes equations and two turbulence equations was solved for in the steady state using a multigrid strategy on unstructured meshes. The flow equations and turbulence equations are solved in a loosely coupled manner. The flow equations are advanced in time using a multistage Runge-Kutta time stepping scheme with a stability bound local time step, while the turbulence equations are advanced in a point-implicit scheme with a time step which guarantees stability and positively. Low Reynolds number modifications to the original two equation model are incorporated in a manner which results in well behaved equations for arbitrarily small wall distances. A variety of aerodynamic flows are solved for, initializing all quantities with uniform freestream values, and resulting in rapid and uniform convergence rates for the flow and turbulence equations.
NASA Astrophysics Data System (ADS)
Franceries, X.; Doyon, B.; Chauveau, N.; Rigaud, B.; Celsis, P.; Morucci, J.-P.
2003-03-01
In electroencephalography (EEG) and event related potentials (ERP), localizing the electrical sources at the origin of scalp potentials (inverse problem) imposes, in a first step, the computation of scalp potential distribution from the simulation of sources (forward problem). This article proposes an alternative method for mimicing both the electrical and geometrical properties of the head, including brain, skull, and scalp tissue with resistors. Two resistor mesh models have been designed to reproduce the three-sphere reference model (analytical model). The first one (spherical resistor mesh) closely mimics the geometrical and electrical properties of the analytical model. The second one (cubic resistor mesh) is designed to conveniently handle anatomical data from magnetic resonance imaging. Both models have been validated, in reference to the analytical solution calculated on the three-sphere model, by computing the magnification factor and the relative difference measure. Results suggest that the mesh models can be used as robust and user-friendly simulation or exploration tools in EEG/ERP.
The Subelement Sweeping method for radiation transport modeling on polygonal meshes
NASA Astrophysics Data System (ADS)
Gleicher, Frederick
A new memory efficient way of obtaining numerical solutions to the radiation transport equation on random polygon meshes is developed and analyzed. This method is called the Subelement Sweeping method, and is applied to the discrete ordinates form of the 1-D and 2-D mono-energetic transport equations. In the Subelement Sweeping method, the coarse mesh is first subdivided into triangular subelements, and the subelement mesh is then swept to obtain subelement angular flux solutions. As the subelements are swept the scalar flux for the scattering source is interpolated from the coarse mesh. Numerical solutions on the subelements are obtained with the linear discontinuous finite element method, and the resulting angular fluxes are projected back onto the coarse mesh and accumulated into new scalar flux values. The old subelement information is then thrown away allowing the method to be memory efficient. Formulas for the interpolation from the coarse mesh fluxes to the subelement mesh fluxes and for the projection from the subelement fluxes to the coarse mesh fluxes are derived by minimizing the squared error norm between coarse mesh and subelement scalar fluxes. Asymptotic analysis is carried out in 1-D, and the Subelement Sweeping method is shown to yield a valid diffusion discretization on the coarse mesh. Asymptotic analysis is also carried out in 2-D, and the Subelement Sweeping method is shown to have the diffusion limit for orthogonal quadrilateral meshes with some simplifying assumptions. The method was implemented for the 1-D slab geometry, and numerical experiments in 1-D show that the Subelement Sweeping method is at least third order accurate and has the diffusion limit. The method was also implemented in the Capsaicin framework within the Anaheim package for 2-D polygonal meshes at Los Alamos National Laboratory. Numerical experiments in 2-D show that the Subelement Sweeping method is at least second order accurate, and suggest strongly that the diffusion
NASA Astrophysics Data System (ADS)
Horritt, M. S.; Bates, P. D.; Mattinson, M. J.
2006-09-01
SummaryThe effects of mesh resolution and topographic data quality on the predictions of a 2D finite volume model of channel flow are investigated. 25 cm resolution side scan sonar swath bathymetry of a 7 km reach of the river Thames, UK, provides topography for a series of finite volume models with resolutions ranging from 2.5 to 50 m. Results from the coarser meshes are compared with the 2.5 m simulation which is used as a benchmark. The model shows greater sensitivity to mesh resolution than topographic sampling. Sensitivity to mesh resolution is attributed to two effects of roughly equal magnitude. Small elements are able to represent hydraulic features such as recirculation zones, and a more accurate representation of the domain boundary helps to drive these flow features. In practical terms, a models at a resolution of 20 and 50 m require 50 m cross-sections, whereas the 10 m model predictions are improved by using all the bathymetry data.
Multigrid techniques for unstructured meshes
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1995-01-01
An overview of current multigrid techniques for unstructured meshes is given. The basic principles of the multigrid approach are first outlined. Application of these principles to unstructured mesh problems is then described, illustrating various different approaches, and giving examples of practical applications. Advanced multigrid topics, such as the use of algebraic multigrid methods, and the combination of multigrid techniques with adaptive meshing strategies are dealt with in subsequent sections. These represent current areas of research, and the unresolved issues are discussed. The presentation is organized in an educational manner, for readers familiar with computational fluid dynamics, wishing to learn more about current unstructured mesh techniques.
Li, Jian; Li, Danxun; Wang, Xingkui
2012-01-01
The Xiangxi River is one of the main tributaries in the Three Gorges reservoir, with the shortest distance to the Three Gorges Project Dam. Severe and frequent algal bloom events have occurred frequently in the Xiangxi River in recent years. Therefore, the current study develops a three-dimensional unstructured-mesh model to investigate the dynamic process of algal bloom. The developed model comprises three modules, namely, hydrodynamics, nutrient cycles, and phytoplankton ecological dynamics. A number of factors, including hydrodynamic condition, nutrient concentration, temperature, and light illumination, that would affect the evolution of phytoplankton were considered. Moreover, the wave equation was used to solve the free surface fluctuations and vertical Z-coordinates with adjustable layered thicknesses. These values, in turn, are suitable for solving the algal bloom problems that occurred in the river style reservoir that has a complex boundary and dramatically changing hydrodynamic conditions. The comparisons between the modeling results and field data of years 2007 and 2008 indicate that the developed model is capable of simulating the algal bloom process in the Xiangxi River with reasonable accuracy. However, hydrodynamic force and external pollution loads affect the concentrations of nutrients, which, along with the underwater light intensity, could consequently affect phytoplankton evolution. Thus, flow velocity cannot be ignored in the analysis of river algal bloom. Based on the modeling results, building an impounding reservoir and increasing the releasing discharge at appropriate times are effective ways for controlling algal bloom. PMID:23520863
Unstructured-mesh modeling of the Congo river-to-sea continuum
NASA Astrophysics Data System (ADS)
Bars, Yoann Le; Vallaeys, Valentin; Deleersnijder, Éric; Hanert, Emmanuel; Carrere, Loren; Channelière, Claire
2016-04-01
With the second largest outflow in the world and one of the widest hydrological basins, the Congo River is of a major importance both locally and globally. However, relatively few studies have been conducted on its hydrology, as compared to other great rivers such as the Amazon, Nile, Yangtze, or Mississippi. The goal of this study is therefore to help fill this gap and provide the first high-resolution simulation of the Congo river-estuary-coastal sea continuum. To this end, we are using a discontinuous-Galerkin finite element marine model that solves the two-dimensional depth-averaged shallow water equations on an unstructured mesh. To ensure a smooth transition from river to coastal sea, we have considered a model that encompasses both hydrological and coastal ocean processes. An important difficulty in setting up this model was to find data to parameterize and validate it, as it is a rather remote and understudied area. Therefore, an important effort in this study has been to establish a methodology to take advantage of all the data sources available including nautical charts that had to be digitalized. The model surface elevation has then been validated with respect to an altimetric database. Model results suggest the existence of gyres in the vicinity of the river mouth that have never been documented before. The effect of those gyres on the Congo River dynamics has been further investigated by simulating the transport of Lagrangian particles and computing the water age.
NASA Astrophysics Data System (ADS)
Valdivia, Valeska; Hennebelle, Patrick
2014-11-01
Context. Ultraviolet radiation plays a crucial role in molecular clouds. Radiation and matter are tightly coupled and their interplay influences the physical and chemical properties of gas. In particular, modeling the radiation propagation requires calculating column densities, which can be numerically expensive in high-resolution multidimensional simulations. Aims: Developing fast methods for estimating column densities is mandatory if we are interested in the dynamical influence of the radiative transfer. In particular, we focus on the effect of the UV screening on the dynamics and on the statistical properties of molecular clouds. Methods: We have developed a tree-based method for a fast estimate of column densities, implemented in the adaptive mesh refinement code RAMSES. We performed numerical simulations using this method in order to analyze the influence of the screening on the clump formation. Results: We find that the accuracy for the extinction of the tree-based method is better than 10%, while the relative error for the column density can be much more. We describe the implementation of a method based on precalculating the geometrical terms that noticeably reduces the calculation time. To study the influence of the screening on the statistical properties of molecular clouds we present the probability distribution function of gas and the associated temperature per density bin and the mass spectra for different density thresholds. Conclusions: The tree-based method is fast and accurate enough to be used during numerical simulations since no communication is needed between CPUs when using a fully threaded tree. It is then suitable to parallel computing. We show that the screening for far UV radiation mainly affects the dense gas, thereby favoring low temperatures and affecting the fragmentation. We show that when we include the screening, more structures are formed with higher densities in comparison to the case that does not include this effect. We
NASA Technical Reports Server (NTRS)
Malek, Miroslaw; Ozden, Banu
1990-01-01
Efficient testing techniques for two-dimensional mesh interconnection networks are presented. The tests cover faults in the arbitration logic of the switches; this includes an examination of fault detection in the data paths, routing, and control circuitry, including the conflict resolution capabilities of mesh interconnection networks using topological test methods. The proposed methods are not implementation specific and can be applied to any design with a mesh topology. The topology and behavior of the network are described and definitions are presented. The fault model is defined and parallel testing methods for the entire network are given.
On Fractional Model Reference Adaptive Control
Shi, Bao; Dong, Chao
2014-01-01
This paper extends the conventional Model Reference Adaptive Control systems to fractional ones based on the theory of fractional calculus. A control law and an incommensurate fractional adaptation law are designed for the fractional plant and the fractional reference model. The stability and tracking convergence are analyzed using the frequency distributed fractional integrator model and Lyapunov theory. Moreover, numerical simulations of both linear and nonlinear systems are performed to exhibit the viability and effectiveness of the proposed methodology. PMID:24574897
On fractional Model Reference Adaptive Control.
Shi, Bao; Yuan, Jian; Dong, Chao
2014-01-01
This paper extends the conventional Model Reference Adaptive Control systems to fractional ones based on the theory of fractional calculus. A control law and an incommensurate fractional adaptation law are designed for the fractional plant and the fractional reference model. The stability and tracking convergence are analyzed using the frequency distributed fractional integrator model and Lyapunov theory. Moreover, numerical simulations of both linear and nonlinear systems are performed to exhibit the viability and effectiveness of the proposed methodology. PMID:24574897
... Device Safety Safety Communications Surgical Mesh: FDA Safety Communication Share Tweet Linkedin Pin it More sharing options ... Prolapse and Stress Urinary Incontinence More in Safety Communications Information About Heparin Preventing Tubing and Luer Misconnections ...
Mesh-based Monte Carlo code for fluorescence modeling in complex tissues with irregular boundaries
NASA Astrophysics Data System (ADS)
Wilson, Robert H.; Chen, Leng-Chun; Lloyd, William; Kuo, Shiuhyang; Marcelo, Cynthia; Feinberg, Stephen E.; Mycek, Mary-Ann
2011-07-01
There is a growing need for the development of computational models that can account for complex tissue morphology in simulations of photon propagation. We describe the development and validation of a user-friendly, MATLAB-based Monte Carlo code that uses analytically-defined surface meshes to model heterogeneous tissue geometry. The code can use information from non-linear optical microscopy images to discriminate the fluorescence photons (from endogenous or exogenous fluorophores) detected from different layers of complex turbid media. We present a specific application of modeling a layered human tissue-engineered construct (Ex Vivo Produced Oral Mucosa Equivalent, EVPOME) designed for use in repair of oral tissue following surgery. Second-harmonic generation microscopic imaging of an EVPOME construct (oral keratinocytes atop a scaffold coated with human type IV collagen) was employed to determine an approximate analytical expression for the complex shape of the interface between the two layers. This expression can then be inserted into the code to correct the simulated fluorescence for the effect of the irregular tissue geometry.
A stochastic model to describe the design and operation of knitted mesh mist eliminators
Wilcock, E.; Davies, G.A.
1996-12-31
Methods used to design and predict the performance of mist eliminators are deficient in their inability to predict the outlet concentration of dispersed liquid in the outlet gas stream. This is a requirement specified in many process applications especially in gas treatment prior to atmospheric dispersion. In this paper a model is proposed, based on determining the trajectory of droplets through a knitted mesh mist eliminator, from which both the outlet gas concentration and drop size distribution in this gas stream can be specified. The model predicts the expected variations in performance, a decrease in the separation efficiency with decreasing both drop size in the feed mixture and gas velocity. A limiting velocity is predicted at which the drainage of liquid from the pad decreases. This corresponds to the flooding condition. Quantitative predictions of this limiting velocity compare well with the experimental results measured for air-water systems. The model offers the prospect of optimizing the pad construction to maximize the separation efficiency at a target pressure drop or designing to a maximum pressure drop. For all designs, the conditions of the outlet gas steam can be defined.
Incorporation of detailed eye model into polygon-mesh versions of ICRP-110 reference phantoms
NASA Astrophysics Data System (ADS)
Tat Nguyen, Thang; Yeom, Yeon Soo; Kim, Han Sung; Wang, Zhao Jun; Han, Min Cheol; Kim, Chan Hyeong; Lee, Jai Ki; Zankl, Maria; Petoussi-Henss, Nina; Bolch, Wesley E.; Lee, Choonsik; Chung, Beom Sun
2015-11-01
The dose coefficients for the eye lens reported in ICRP 2010 Publication 116 were calculated using both a stylized model and the ICRP-110 reference phantoms, according to the type of radiation, energy, and irradiation geometry. To maintain consistency of lens dose assessment, in the present study we incorporated the ICRP-116 detailed eye model into the converted polygon-mesh (PM) version of the ICRP-110 reference phantoms. After the incorporation, the dose coefficients for the eye lens were calculated and compared with those of the ICRP-116 data. The results showed generally a good agreement between the newly calculated lens dose coefficients and the values of ICRP 2010 Publication 116. Significant differences were found for some irradiation cases due mainly to the use of different types of phantoms. Considering that the PM version of the ICRP-110 reference phantoms preserve the original topology of the ICRP-110 reference phantoms, it is believed that the PM version phantoms, along with the detailed eye model, provide more reliable and consistent dose coefficients for the eye lens.
NASA Astrophysics Data System (ADS)
Chen, Ying; Shen, Jie
2016-03-01
In this paper we develop a fully adaptive energy stable scheme for Cahn-Hilliard Navier-Stokes system, which is a phase-field model for two-phase incompressible flows, consisting a Cahn-Hilliard-type diffusion equation and a Navier-Stokes equation. This scheme, which is decoupled and unconditionally energy stable based on stabilization, involves adaptive mesh, adaptive time and a nonlinear multigrid finite difference method. Numerical experiments are carried out to validate the scheme for problems with matched density and non-matched density, and also demonstrate that CPU time can be significantly reduced with our adaptive approach.
NASA Astrophysics Data System (ADS)
Nissen-Meyer, T.; Luo, Y.; Morency, C.; Tromp, J.
2008-12-01
Seismic-wave propagation in exploration-industry settings has seen major research and development efforts for decades, yet large-scale applications have often been limited to 2D or 3D finite-difference, (visco- )acoustic wave propagation due to computational limitations. We explore the possibility of including all relevant physical signatures in the wavefield using the spectral- element method (SPECFEM3D, SPECFEM2D), thereby accounting for acoustic, (visco-)elastic, poroelastic, anisotropic wave propagation in meshes which honor all crucial discontinuities. Mesh design is the crux of the problem, and we use CUBIT (Sandia Laboratories) to generate unstructured quadrilateral 2D and hexahedral 3D meshes for these complex background models. While general hexahedral mesh generation is an unresolved problem, we are able to accommodate most of the relevant settings (e.g., layer-cake models, salt bodies, overthrusting faults, and strong topography) with respectively tailored workflows. 2D simulations show localized, characteristic wave effects due to these features that shall be helpful in designing survey acquisition geometries in a relatively economic fashion. We address some of the fundamental issues this comprehensive modeling approach faces regarding its feasibility: Assessing geological structures in terms of the necessity to honor the major structural units, appropriate velocity model interpolation, quality control of the resultant mesh, and computational cost for realistic settings up to frequencies of 40 Hz. The solution to this forward problem forms the basis for subsequent 2D and 3D adjoint tomography within this context, which is the subject of a companion paper.
Numerical simulation of deformation of dynamic mesh in the human vocal tract model
NASA Astrophysics Data System (ADS)
Řidký, Václav; Šidlof, Petr
2015-05-01
Numerical simulation of the acoustic signal generation in the human vocal tract is a very complex problem. The computational mesh is not static; it is deformed due to vibration of vocal folds. Movement of vocal folds is in this case prescribed as function of translation and rotation. A new boundary condition for the 2DOF motion of the vocal folds was implemented in OpenFOAM, an open-source software package based on finite volume method Work is focused on the dynamic mesh and deformation of structured meshes in the computation a package OpenFOAM. These methods are compared with focus onquality of the mesh (non-orthogonality, aspect ratio and skewness).
Examples of Mesh and NURBS modelling for in vivo lung counting studies.
Farah, Jad; Broggio, David; Franck, Didier
2011-03-01
Realistic calibration coefficients for in vivo counting installations are assessed using voxel phantoms and Monte Carlo calculations. However, voxel phantoms construction is time consuming and their flexibility extremely limited. This paper involves Mesh and non-uniform rational B-splines graphical formats, of greater flexibility, to optimise the calibration of in vivo counting installations. Two studies validating the use of such phantoms and involving geometry deformation and modelling were carried out to study the morphologic effect on lung counting efficiency. The created 3D models fitted with the reference ones, with volumetric differences of <5 %. Moreover, it was found that counting efficiency varies with the inverse of lungs' volume and that the latter primes when compared with chest wall thickness. Finally, a series of different thoracic female phantoms of various cup sizes, chest girths and internal organs' volumes were created starting from the International Commission on Radiological Protection (ICRP) adult female reference computational phantom to give correction factors for the lung monitoring of female workers. PMID:21030397
Resistor mesh model of a spherical head: part 1: applications to scalp potential interpolation.
Chauveau, N; Morucci, J P; Franceries, X; Celsis, P; Rigaud, B
2005-11-01
A resistor mesh model (RMM) has been implemented to describe the electrical properties of the head and the configuration of the intracerebral current sources by simulation of forward and inverse problems in electroencephalogram/event related potential (EEG/ERP) studies. For this study, the RMM representing the three basic tissues of the human head (brain, skull and scalp) was superimposed on a spherical volume mimicking the head volume: it included 43 102 resistances and 14 123 nodes. The validation was performed with reference to the analytical model by consideration of a set of four dipoles close to the cortex. Using the RMM and the chosen dipoles, four distinct families of interpolation technique (nearest neighbour, polynomial, splines and lead fields) were tested and compared so that the scalp potentials could be recovered from the electrode potentials. The 3D spline interpolation and the inverse forward technique (IFT) gave the best results. The IFT is very easy to use when the lead-field matrix between scalp electrodes and cortex nodes has been calculated. By simple application of the Moore-Penrose pseudo inverse matrix to the electrode cap potentials, a set of current sources on the cortex is obtained. Then, the forward problem using these cortex sources renders all the scalp potentials. PMID:16594294
Numerical Modeling of Long Bone Adaptation due to Mechanical Loading: Correlation with Experiments
Kumar, Natarajan Chennimalai; Dantzig, Jonathan A.; Jasiuk, Iwona M.; Robling, Alex G.; Turner, Charles H.
2011-01-01
The process of external bone adaptation in cortical bone is modeled mathematically using finite element (FE) stress analysis coupled with an evolution model, in which adaptation response is triggered by mechanical stimulus represented by strain energy density. The model is applied to experiments in which a rat ulna is subjected to cyclic loading, and the results demonstrate the ability of the model to predict the bone adaptation response. The FE mesh is generated from micro-computed tomography (μCT) images of the rat ulna, and the stress analysis is carried out using boundary and loading conditions on the rat ulna obtained from the experiments [Robling, A. G., F. M. Hinant, D. B. Burr, and C. H. Turner. J. Bone Miner. Res. 17:1545–1554, 2002]. The external adaptation process is implemented in the model by moving the surface nodes of the FE mesh based on an evolution law characterized by two parameters: one that captures the rate of the adaptation process (referred to as gain); and the other characterizing the threshold value of the mechanical stimulus required for adaptation (referred to as threshold-sensitivity). A parametric study is carried out to evaluate the effect of these two parameters on the adaptation response. We show, following comparison of results from the simulations to the experimental observations of Robling et al. (J. Bone Miner. Res. 17:1545–1554, 2002), that splitting the loading cycles into different number of bouts affects the threshold-sensitivity but not the rate of adaptation. We also show that the threshold-sensitivity parameter can quantify the mechanosensitivity of the osteocytes. PMID:20013156
Graphical Models and Computerized Adaptive Testing.
ERIC Educational Resources Information Center
Almond, Russell G.; Mislevy, Robert J.
1999-01-01
Considers computerized adaptive testing from the perspective of graphical modeling (GM). GM provides methods for making inferences about multifaceted skills and knowledge and for extracting data from complex performances. Provides examples from language-proficiency assessment. (SLD)
Guzik, S; McCorquodale, P; Colella, P
2011-12-16
A fourth-order accurate finite-volume method is presented for solving time-dependent hyperbolic systems of conservation laws on mapped grids that are adaptively refined in space and time. Novel considerations for formulating the semi-discrete system of equations in computational space combined with detailed mechanisms for accommodating the adapting grids ensure that conservation is maintained and that the divergence of a constant vector field is always zero (freestream-preservation property). Advancement in time is achieved with a fourth-order Runge-Kutta method.
Development and validation of a three-dimensional, wave-current coupled model on unstructured meshes
NASA Astrophysics Data System (ADS)
Wang, JinHua; Shen, YongMing
2011-01-01
Using unstructured meshes provides great flexibility for modeling the flow in complex geomorphology of tidal creeks, barriers and islands, with refined grid resolution in regions of interest and not elsewhere. In this paper, an unstructured three-dimensional fully coupled wave-current model is developed. Firstly, a parallel, unstructured wave module is developed. Variations in wave properties are governed by a wave energy equation that includes wave-current interactions and dissipation representative of wave breaking. Then, the existing Finite-Volume Coastal Ocean Model (FVCOM) is modified to couple with the wave module. The couple procedure includes depth dependent wave radiation stress terms, Stokes drift, vertical transfer of wave-generated pressure transfer to the mean momentum equation, wave dissipation as a source term in the turbulence kinetic energy equation, and mean current advection and refraction of wave energy. Several applications are presented to evaluate the developed model. In particular the wind and wave-induced storm surge generated by Hurricane Katrina is investigated. The obtained results have been compared to the in situ measurements with respect to the wave heights and water level elevations revealing good accuracy of the model in reproduction of the investigated events. In a comparison to water level measurements at Dauphin Island, inclusion of the wave induced water level setup reduced the normalized root mean square error from 0.301 to 0.257 m and increased the correlation coefficient from 0.860 to 0.929. Several runs were carried out to analyze the effects of waves. The experiments show that among the processes that represent wave effects, radiation stress and wave-induced surface stress are more important than wave-induced bottom stress in affecting the water level. The Hurricane Katrina simulations showed the importance of the inclusion of the wave effects for the hindcast of the water levels during the storm surge.
Tautges, Timothy J.
2004-04-01
MOAB is a component for representing and evaluating mesh data. MOAB can store stuctured and unstructured mesh, consisting of elements in the finite element "zoo". The functional interface to MOAB is simple yet powerful, allowing the representation of many types of metadata commonly found on the mesh. MOAB is optimized for efficiency in space and time, based on access to mesh in chunks rather than through individual entities, while also versatile enough to support individual entity access. The MOAB data model consists of a mesh interface instance, mesh entities (vertices and elements), sets, and tags. Entities are addressed through handles rather than pointers, to allow the underlying representation of an entity to change without changing the handle to that entity. Sets are arbitrary groupings of mesh entities and other sets. Sets also support parent/child relationships as a relation distinct from sets containing other sets. The directed-graph provided by set parent/child relationships is useful for modeling topological relations from a geometric model or other metadata. Tags are named data which can be assigned to the mesh as a whole, individual entities, or sets. Tags are a mechanism for attaching data to individual entities and sets are a mechanism for describing relations between entities; the combination of these two mechanisms isa powerful yet simple interface for representing metadata or application-specific data. For example, sets and tags can be used together to describe geometric topology, boundary condition, and inter-processor interface groupings in a mesh. MOAB is used in several ways in various applications. MOAB serves as the underlying mesh data representation in the VERDE mesh verification code. MOAB can also be used as a mesh input mechanism, using mesh readers induded with MOAB, or as a tanslator between mesh formats, using readers and writers included with MOAB.
2004-04-01
MOAB is a component for representing and evaluating mesh data. MOAB can store stuctured and unstructured mesh, consisting of elements in the finite element "zoo". The functional interface to MOAB is simple yet powerful, allowing the representation of many types of metadata commonly found on the mesh. MOAB is optimized for efficiency in space and time, based on access to mesh in chunks rather than through individual entities, while also versatile enough to support individualmore » entity access. The MOAB data model consists of a mesh interface instance, mesh entities (vertices and elements), sets, and tags. Entities are addressed through handles rather than pointers, to allow the underlying representation of an entity to change without changing the handle to that entity. Sets are arbitrary groupings of mesh entities and other sets. Sets also support parent/child relationships as a relation distinct from sets containing other sets. The directed-graph provided by set parent/child relationships is useful for modeling topological relations from a geometric model or other metadata. Tags are named data which can be assigned to the mesh as a whole, individual entities, or sets. Tags are a mechanism for attaching data to individual entities and sets are a mechanism for describing relations between entities; the combination of these two mechanisms isa powerful yet simple interface for representing metadata or application-specific data. For example, sets and tags can be used together to describe geometric topology, boundary condition, and inter-processor interface groupings in a mesh. MOAB is used in several ways in various applications. MOAB serves as the underlying mesh data representation in the VERDE mesh verification code. MOAB can also be used as a mesh input mechanism, using mesh readers induded with MOAB, or as a tanslator between mesh formats, using readers and writers included with MOAB.« less
Geometrical and topological issues in octree based automatic meshing
NASA Technical Reports Server (NTRS)
Saxena, Mukul; Perucchio, Renato
1987-01-01
Finite element meshes derived automatically from solid models through recursive spatial subdivision schemes (octrees) can be made to inherit the hierarchical structure and the spatial addressability intrinsic to the underlying grid. These two properties, together with the geometric regularity that can also be built into the mesh, make octree based meshes ideally suited for efficient analysis and self-adaptive remeshing and reanalysis. The element decomposition of the octal cells that intersect the boundary of the domain is discussed. The problem, central to octree based meshing, is solved by combining template mapping and element extraction into a procedure that utilizes both constructive solid geometry and boundary representation techniques. Boundary cells that are not intersected by the edge of the domain boundary are easily mapped to predefined element topology. Cells containing edges (and vertices) are first transformed into a planar polyhedron and then triangulated via element extractor. The modeling environments required for the derivation of planar polyhedra and for element extraction are analyzed.
NASA Astrophysics Data System (ADS)
Wang, Hairen; Lou, Zheng; Qian, Yuan; Zheng, Xianzhong; Zuo, Yingxi
2016-03-01
The optimization of a primary mirror support system is one of the most critical problems in the design of large telescopes. Here, we propose a hybrid optimization methodology of variable densities mesh model (HOMVDMM) for the axial supporting design, which has three key steps: (1) creating a variable densities mesh model, which will partition the mirror into several sparse mesh areas and several dense mesh areas; (2) global optimization based on the zero-order optimization method for the support of primary mirror with a large tolerance; (3) based on the optimization results of the second step, further optimization with first-order optimization method in dense mesh areas by a small tolerance. HOMVDMM exploits the complementary merits of both the zero- and first-order optimizations, with the former in global scale and the latter in small scale. As an application, the axial support of the primary mirror of the 2.5-m wide-field survey telescope (WFST) is optimized by HOMVDMM. These three designs are obtained via a comparative study of different supporting points including 27 supporting points, 39 supporting points, and 54 supporting points. Their residual half-path length errors are 28.78, 9.32, and 5.29 nm. The latter two designs both meet the specification of WFST. In each of the three designs, a global optimization value with high accuracy will be obtained in an hour on an ordinary PC. As the results suggest, the overall performance of HOMVDMM is superior to the first-order optimization method as well as the zero-order optimization method.
Mesh versus bathtub - effects of flood models on exposure analysis in Switzerland
NASA Astrophysics Data System (ADS)
Röthlisberger, Veronika; Zischg, Andreas; Keiler, Margreth
2016-04-01
In Switzerland, mainly two types of maps that indicate potential flood zones are available for flood exposure analyses: 1) Aquaprotect, a nationwide overview provided by the Federal Office for the Environment and 2) communal flood hazard maps available from the 26 cantons. The model used to produce Aquaprotect can be described as a bathtub approach or linear superposition method with three main parameters, namely the horizontal and vertical distance of a point to water features and the size of the river sub-basin. Whereas the determination of flood zones in Aquaprotect is based on a uniform, nationwide model, the communal flood hazard maps are less homogenous, as they have been elaborated either at communal or cantonal levels. Yet their basic content (i.e. indication of potential flood zones for three recurrence periods, with differentiation of at least three inundation depths) is described in national directives and the vast majority of communal flood hazard maps are based on 2D inundation simulations using meshes. Apart from the methodical differences between Aquaprotect and the communal flood hazard maps (and among different communal flood hazard maps), all of these maps include a layer with a similar recurrence period (i.e. Aquaprotect 250 years, flood hazard maps 300 years) beyond the intended protection level of installed structural systems. In our study, we compare the resulting exposure by overlaying the two types of flood maps with a complete, harmonized, and nationwide dataset of building polygons. We assess the different exposure at the national level, and also consider differences among the 26 cantons and the six biogeographically unique regions, respectively. It was observed that while the nationwide exposure rates for both types of flood maps are similar, the differences within certain cantons and biogeographical regions are remarkable. We conclude that flood maps based on bathtub models are appropriate for assessments at national levels, while maps
Parallel goal-oriented adaptive finite element modeling for 3D electromagnetic exploration
NASA Astrophysics Data System (ADS)
Zhang, Y.; Key, K.; Ovall, J.; Holst, M.
2014-12-01
We present a parallel goal-oriented adaptive finite element method for accurate and efficient electromagnetic (EM) modeling of complex 3D structures. An unstructured tetrahedral mesh allows this approach to accommodate arbitrarily complex 3D conductivity variations and a priori known boundaries. The total electric field is approximated by the lowest order linear curl-conforming shape functions and the discretized finite element equations are solved by a sparse LU factorization. Accuracy of the finite element solution is achieved through adaptive mesh refinement that is performed iteratively until the solution converges to the desired accuracy tolerance. Refinement is guided by a goal-oriented error estimator that uses a dual-weighted residual method to optimize the mesh for accurate EM responses at the locations of the EM receivers. As a result, the mesh refinement is highly efficient since it only targets the elements where the inaccuracy of the solution corrupts the response at the possibly distant locations of the EM receivers. We compare the accuracy and efficiency of two approaches for estimating the primary residual error required at the core of this method: one uses local element and inter-element residuals and the other relies on solving a global residual system using a hierarchical basis. For computational efficiency our method follows the Bank-Holst algorithm for parallelization, where solutions are computed in subdomains of the original model. To resolve the load-balancing problem, this approach applies a spectral bisection method to divide the entire model into subdomains that have approximately equal error and the same number of receivers. The finite element solutions are then computed in parallel with each subdomain carrying out goal-oriented adaptive mesh refinement independently. We validate the newly developed algorithm by comparison with controlled-source EM solutions for 1D layered models and with 2D results from our earlier 2D goal oriented
Tsunami modelling with adaptively refined finite volume methods
LeVeque, R.J.; George, D.L.; Berger, M.J.
2011-01-01
Numerical modelling of transoceanic tsunami propagation, together with the detailed modelling of inundation of small-scale coastal regions, poses a number of algorithmic challenges. The depth-averaged shallow water equations can be used to reduce this to a time-dependent problem in two space dimensions, but even so it is crucial to use adaptive mesh refinement in order to efficiently handle the vast differences in spatial scales. This must be done in a 'wellbalanced' manner that accurately captures very small perturbations to the steady state of the ocean at rest. Inundation can be modelled by allowing cells to dynamically change from dry to wet, but this must also be done carefully near refinement boundaries. We discuss these issues in the context of Riemann-solver-based finite volume methods for tsunami modelling. Several examples are presented using the GeoClaw software, and sample codes are available to accompany the paper. The techniques discussed also apply to a variety of other geophysical flows. ?? 2011 Cambridge University Press.
Toward Interoperable Mesh, Geometry and Field Components for PDE Simulation Development
Chand, K K; Diachin, L F; Li, X; Ollivier-Gooch, C; Seol, E S; Shephard, M; Tautges, T; Trease, H
2005-07-11
Mesh-based PDE simulation codes are becoming increasingly sophisticated and rely on advanced meshing and discretization tools. Unfortunately, it is still difficult to interchange or interoperate tools developed by different communities to experiment with various technologies or to develop new capabilities. To address these difficulties, we have developed component interfaces designed to support the information flow of mesh-based PDE simulations. We describe this information flow and discuss typical roles and services provided by the geometry, mesh, and field components of the simulation. Based on this delineation for the roles of each component, we give a high-level description of the abstract data model and set of interfaces developed by the Department of Energy's Interoperable Tools for Advanced Petascale Simulation (ITAPS) center. These common interfaces are critical to our interoperability goal, and we give examples of several services based upon these interfaces including mesh adaptation and mesh improvement.
Gamra: Simple meshing for complex earthquakes
NASA Astrophysics Data System (ADS)
Landry, Walter; Barbot, Sylvain
2016-05-01
The static offsets caused by earthquakes are well described by elastostatic models with a discontinuity in the displacement along the fault. A traditional approach to model this discontinuity is to align the numerical mesh with the fault and solve the equations using finite elements. However, this distorted mesh can be difficult to generate and update. We present a new numerical method, inspired by the Immersed Interface Method (Leveque and Li, 1994), for solving the elastostatic equations with embedded discontinuities. This method has been carefully designed so that it can be used on parallel machines on an adapted finite difference grid. We have implemented this method in Gamra, a new code for earth modeling. We demonstrate the correctness of the method with analytic tests, and we demonstrate its practical performance by solving a realistic earthquake model to extremely high precision.
Tangle-Free Mesh Motion for Ablation Simulations
NASA Technical Reports Server (NTRS)
Droba, Justin
2016-01-01
Problems involving mesh motion-which should not be mistakenly associated with moving mesh methods, a class of adaptive mesh redistribution techniques-are of critical importance in numerical simulations of the thermal response of melting and ablative materials. Ablation is the process by which material vaporizes or otherwise erodes due to strong heating. Accurate modeling of such materials is of the utmost importance in design of passive thermal protection systems ("heatshields") for spacecraft, the layer of the vehicle that ensures survival of crew and craft during re-entry. In an explicit mesh motion approach, a complete thermal solve is first performed. Afterwards, the thermal response is used to determine surface recession rates. These values are then used to generate boundary conditions for an a posteriori correction designed to update the location of the mesh nodes. Most often, linear elastic or biharmonic equations are used to model this material response, traditionally in a finite element framework so that complex geometries can be simulated. A simple scheme for moving the boundary nodes involves receding along the surface normals. However, for all but the simplest problem geometries, evolution in time following such a scheme will eventually bring the mesh to intersect and "tangle" with itself, inducing failure. This presentation demonstrates a comprehensive and sophisticated scheme that analyzes the local geometry of each node with help from user-provided clues to eliminate the tangle and enable simulations on a wide-class of difficult problem geometries. The method developed is demonstrated for linear elastic equations but is general enough that it may be adapted to other modeling equations. The presentation will explicate the inner workings of the tangle-free mesh motion algorithm for both two and three-dimensional meshes. It will show abstract examples of the method's success, including a verification problem that demonstrates its accuracy and
Adaptive Modeling Language and Its Derivatives
NASA Technical Reports Server (NTRS)
Chemaly, Adel
2006-01-01
Adaptive Modeling Language (AML) is the underlying language of an object-oriented, multidisciplinary, knowledge-based engineering framework. AML offers an advanced modeling paradigm with an open architecture, enabling the automation of the entire product development cycle, integrating product configuration, design, analysis, visualization, production planning, inspection, and cost estimation.
Parallel Adaptive Multi-Mechanics Simulations using Diablo
Parsons, D; Solberg, J
2004-12-03
Coupled multi-mechanics simulations (such as thermal-stress and fluidstructure interaction problems) are of substantial interest to engineering analysts. In addition, adaptive mesh refinement techniques present an attractive alternative to current mesh generation procedures and provide quantitative error bounds that can be used for model verification. This paper discusses spatially adaptive multi-mechanics implicit simulations using the Diablo computer code. (U)
Winny, M; Grethe, L; Maegel, L; Jonigk, D; Lippmann, T; Klempnauer, J; Poehnert, D
2016-01-01
Background: Meshes implanted intraperitoneally are known to cause adhesions potentially resulting in complications such as chronic pain, enterocutaneous fistula, or mesh infection. This study introduces a model for investigation of intestine-to-mesh adhesions and evaluates as to whether missing of visceral peritoneum is causative. Methods: In 18 rats, rectangular 1.5 x 2 cm patches of an uncoated polypropylene mesh (Ultrapro®) were sewn to the inner abdominal wall next to the cecum. Additionally, a meso-suture ensured contact between cecum and mesh. Rats were assigned to 2 groups: in 8 rats the peritoneum was left intact, in 10 the cecum was depleted from peritoneum with abrasion. Sacrifice was on day 7. Macroscopic evaluation used two adhesion scores. Specimens were evaluated microscopically, statistical analyses employed student's t-test. Results: On day 7, rats with mesh implantation combined with locally de-peritonealization by cecal abrasion mostly showed severe cecum-to-mesh agglutination (mean Lauder score 92%, mean total Hoffmann score 90%), whereas meshes of most animals without cecal abrasion only had some coverage with intraabdominal fat (33%, 24%; p = 0.0002). Histological work-up showed adequate wall ingrowth of mesh in all rats. In animals with cecal abrasion, meshes were mostly adhesive with cecal wall. However, when the peritoneum of cecum was unimpaired, abdominal wall above the mesh as well as cecum usually revealed sub-peritoneal tissue and a mono-layer cell coverage as seen in normal peritoneum. Conclusion: This study introduces a model mimicking a clinical situation of e.g. hernia repair by intraperitoneally implanted meshes when mesh has contact with normal and with de-peritonealized intestine. The model might be useful for testing mesh types and coatings as well as other devices for their efficacy in adhesion prevention. The high adhesion scores of rats with local de-peritonealization compared with the low scores of animals with intact
Interoperable mesh and geometry tools for advanced petascale simulations
Diachin, L; Bauer, A; Fix, B; Kraftcheck, J; Jansen, K; Luo, X; Miller, M; Ollivier-Gooch, C; Shephard, M; Tautges, T; Trease, H
2007-07-04
SciDAC applications have a demonstrated need for advanced software tools to manage the complexities associated with sophisticated geometry, mesh, and field manipulation tasks, particularly as computer architectures move toward the petascale. The Center for Interoperable Technologies for Advanced Petascale Simulations (ITAPS) will deliver interoperable and interchangeable mesh, geometry, and field manipulation services that are of direct use to SciDAC applications. The premise of our technology development goal is to provide such services as libraries that can be used with minimal intrusion into application codes. To develop these technologies, we focus on defining a common data model and datastructure neutral interfaces that unify a number of different services such as mesh generation and improvement, front tracking, adaptive mesh refinement, shape optimization, and solution transfer operations. We highlight the use of several ITAPS services in SciDAC applications.
Ziegelwanger, Harald; Majdak, Piotr; Kreuzer, Wolfgang
2015-01-01
Head-related transfer functions (HRTFs) can be numerically calculated by applying the boundary element method on the geometry of a listener’s head and pinnae. The calculation results are defined by geometrical, numerical, and acoustical parameters like the microphone used in acoustic measurements. The scope of this study was to estimate requirements on the size and position of the microphone model and on the discretization of the boundary geometry as triangular polygon mesh for accurate sound localization. The evaluation involved the analysis of localization errors predicted by a sagittal-plane localization model, the comparison of equivalent head radii estimated by a time-of-arrival model, and the analysis of actual localization errors obtained in a sound-localization experiment. While the average edge length (AEL) of the mesh had a negligible effect on localization performance in the lateral dimension, the localization performance in sagittal planes, however, degraded for larger AELs with the geometrical error as dominant factor. A microphone position at an arbitrary position at the entrance of the ear canal, a microphone size of 1 mm radius, and a mesh with 1 mm AEL yielded a localization performance similar to or better than observed with acoustically measured HRTFs. PMID:26233020
Adaptive System Modeling for Spacecraft Simulation
NASA Technical Reports Server (NTRS)
Thomas, Justin
2011-01-01
This invention introduces a methodology and associated software tools for automatically learning spacecraft system models without any assumptions regarding system behavior. Data stream mining techniques were used to learn models for critical portions of the International Space Station (ISS) Electrical Power System (EPS). Evaluation on historical ISS telemetry data shows that adaptive system modeling reduces simulation error anywhere from 50 to 90 percent over existing approaches. The purpose of the methodology is to outline how someone can create accurate system models from sensor (telemetry) data. The purpose of the software is to support the methodology. The software provides analysis tools to design the adaptive models. The software also provides the algorithms to initially build system models and continuously update them from the latest streaming sensor data. The main strengths are as follows: Creates accurate spacecraft system models without in-depth system knowledge or any assumptions about system behavior. Automatically updates/calibrates system models using the latest streaming sensor data. Creates device specific models that capture the exact behavior of devices of the same type. Adapts to evolving systems. Can reduce computational complexity (faster simulations).
Hybrid adaptive control of a dragonfly model
NASA Astrophysics Data System (ADS)
Couceiro, Micael S.; Ferreira, Nuno M. F.; Machado, J. A. Tenreiro
2012-02-01
Dragonflies show unique and superior flight performances than most of other insect species and birds. They are equipped with two pairs of independently controlled wings granting an unmatchable flying performance and robustness. In this paper, it is presented an adaptive scheme controlling a nonlinear model inspired in a dragonfly-like robot. It is proposed a hybrid adaptive ( HA) law for adjusting the parameters analyzing the tracking error. At the current stage of the project it is considered essential the development of computational simulation models based in the dynamics to test whether strategies or algorithms of control, parts of the system (such as different wing configurations, tail) as well as the complete system. The performance analysis proves the superiority of the HA law over the direct adaptive ( DA) method in terms of faster and improved tracking and parameter convergence.
NASA Astrophysics Data System (ADS)
Kim, Kyeong Ok; Choi, Byung Ho; Jung, Kyung Tae
2016-04-01
The performance of an integrally coupled wave-tide-surge model using the unstructured mesh system has been tested for the typhoon Bolaven which is regarded as the most powerful storm to strike the Korean Peninsula in nearly a decade with wind gusts measured up to 50 m/s, causing serious damages with 19 victims. Use of the unstructured mesh in coastal sea regions of marginal scale allows all energy from deep to shallow waters to be seamlessly followed; the physics of wave-circulation interactions can be then correctly resolved. The model covers the whole Yellow and East China Seas with locally refined meshes near the regions of Gageo Island (offshore southwestern corner of the Korean Peninsula) and south of Jeju Island (Gangjeong and Seogwipo ports). The wind and pressure fields during the passage of typhoon Bolaven are generated by the blending method. Generally the numerical atmospheric model cannot satisfactorily reproduce the strength of typhoons due to dynamic and resolution restrictions. In this study we could achieve an improved conservation of the typhoon strength by blending the Holland typhoon model result by the empirical formula onto the ambient meteorological fields of NCEP dataset. The model results are compared with the observations and the model performance is then evaluated. The computed wave spectrums for one and two dimensions are compared with the observation in Ieodo station. Results show that the wind wave significantly enhances the current intensity and surge elevation, addressing that to incorporate the wave-current interaction effect in the wave-tide-surge coupled model is important for the accurate prediction of current and sea surface elevation as well as extreme waves in shallow coastal sea regions. The resulting modeling system can be used for hindcasting and forecasting the wave-tide-surges in marine environments with complex coastlines, shallow water depth and fine sediment.
The Stratified Ocean Model with Adaptive Refinement (SOMAR)
NASA Astrophysics Data System (ADS)
Santilli, Edward; Scotti, Alberto
2015-06-01
A computational framework for the evolution of non-hydrostatic, baroclinic flows encountered in regional and coastal ocean simulations is presented, which combines the flexibility of Adaptive Mesh Refinement (AMR) with a suite of numerical tools specifically developed to deal with the high degree of anisotropy of oceanic flows and their attendant numerical challenges. This framework introduces a semi-implicit update of the terms that give rise to buoyancy oscillations, which permits a stable integration of the Navier-Stokes equations when a background density stratification is present. The lepticity of each grid in the AMR hierarchy, which serves as a useful metric for anisotropy, is used to select one of several different efficient Poisson-solving techniques. In this way, we compute the pressure over the entire set of AMR grids without resorting to the hydrostatic approximation, which can degrade the structure of internal waves whose dynamics may have large-scale significance. We apply the modeling framework to three test cases, for which numerical or analytical solutions are known that can be used to benchmark the results. In all the cases considered, the model achieves an excellent degree of congruence with the benchmark, while at the same time achieving a substantial reduction of the computational resources needed.
Adaptive deformable model for mouth boundary detection
NASA Astrophysics Data System (ADS)
Mirhosseini, Ali R.; Yan, Hong; Lam, Kin-Man
1998-03-01
A new generalized algorithm is proposed to automatically extract a mouth boundary model form human face images. Such an algorithm can contribute to human face recognition and lip-reading-assisted speech recognition systems, in particular, and multimodal human computer interaction system, in general. The new model is an iterative algorithm based on a hierarchical model adaptation scheme using deformable templates, as a generalization of some of the previous works. The role of prior knowledge is essential for perceptual organization in the algorithm. The prior knowledge about the mouth shape is used to define and initialize a primary deformable mode. Each primary boundary curve of a mouth is formed on three control points, including two mouth corners, whose locations are optimized using a primary energy functional. This energy functional essentially captures the knowledge of the mouth shape to perceptually organize image information. The primary model is finely tuned in the second stage of optimization algorithm using a generalized secondary energy functional. Basically each boundary curve is finely tuned using more control points. The primary model is replaced by an adapted model if there is an increase in the secondary energy functional. The results indicate that the new model adaptation technique satisfactorily generalizes the mouth boundary model extraction in an automated fashion.
NASA Astrophysics Data System (ADS)
Gansen, A.; El Hachemi, M.; Belouettar, S.; Hassan, O.; Morgan, K.
2015-12-01
In computational electromagnetics, the advantages of the standard Yee algorithm are its simplicity and its low computational costs. However, because of the accuracy losses resulting from the staircased representation of curved interfaces, it is normally not the method of choice for modelling electromagnetic interactions with objects of arbitrary shape. For these problems, an unstructured mesh finite volume time domain method is often employed, although the scheme does not satisfy the divergence free condition at the discrete level. In this paper, we generalize the standard Yee algorithm for use on unstructured meshes and solve the problem concerning the loss of accuracy linked to staircasing, while preserving the divergence free nature of the algorithm. The scheme is implemented on high quality primal Delaunay and dual Voronoi meshes. The performance of the approach was validated in previous work by simulating the scattering of electromagnetic waves by spherical 3D PEC objects in free space. In this paper we demonstrate the performance of this scheme for penetration problems in lossy dielectrics using a new averaging technique for Delaunay and Voronoi edges at the interface. A detailed explanation of the implementation of the method, and a demonstration of the quality of the results obtained for transmittance and scattering simulations by 3D objects of arbitrary shapes, are presented.
NASA Astrophysics Data System (ADS)
Blatov, I. A.; Dobrobog, N. V.; Kitaeva, E. V.
2016-07-01
The Galerkin finite element method is applied to nonself-adjoint singularly perturbed boundary value problems on Shishkin meshes. The Galerkin projection method is used to obtain conditionally ɛ-uniform a priori error estimates and to prove the convergence of a sequence of meshes in the case of an unknown boundary layer edge.
Automated adaptive inference of phenomenological dynamical models
Daniels, Bryan C.; Nemenman, Ilya
2015-01-01
Dynamics of complex systems is often driven by large and intricate networks of microscopic interactions, whose sheer size obfuscates understanding. With limited experimental data, many parameters of such dynamics are unknown, and thus detailed, mechanistic models risk overfitting and making faulty predictions. At the other extreme, simple ad hoc models often miss defining features of the underlying systems. Here we develop an approach that instead constructs phenomenological, coarse-grained models of network dynamics that automatically adapt their complexity to the available data. Such adaptive models produce accurate predictions even when microscopic details are unknown. The approach is computationally tractable, even for a relatively large number of dynamical variables. Using simulated data, it correctly infers the phase space structure for planetary motion, avoids overfitting in a biological signalling system and produces accurate predictions for yeast glycolysis with tens of data points and over half of the interacting species unobserved. PMID:26293508
Automated adaptive inference of phenomenological dynamical models
NASA Astrophysics Data System (ADS)
Daniels, Bryan C.; Nemenman, Ilya
2015-08-01
Dynamics of complex systems is often driven by large and intricate networks of microscopic interactions, whose sheer size obfuscates understanding. With limited experimental data, many parameters of such dynamics are unknown, and thus detailed, mechanistic models risk overfitting and making faulty predictions. At the other extreme, simple ad hoc models often miss defining features of the underlying systems. Here we develop an approach that instead constructs phenomenological, coarse-grained models of network dynamics that automatically adapt their complexity to the available data. Such adaptive models produce accurate predictions even when microscopic details are unknown. The approach is computationally tractable, even for a relatively large number of dynamical variables. Using simulated data, it correctly infers the phase space structure for planetary motion, avoids overfitting in a biological signalling system and produces accurate predictions for yeast glycolysis with tens of data points and over half of the interacting species unobserved.
Model reference adaptive systems some examples.
NASA Technical Reports Server (NTRS)
Landau, I. D.; Sinner, E.; Courtiol, B.
1972-01-01
A direct design method is derived for several single-input single-output model reference adaptive systems (M.R.A.S.). The approach used helps to clarify the various steps involved in a design, which utilizes the hyperstability concept. An example of a multiinput, multioutput M.R.A.S. is also discussed. Attention is given to the problem of a series compensator. It is pointed out that a series compensator which contains derivative terms must generally be introduced in the adaptation mechanism in order to assure asymptotic hyperstability. Results obtained by the simulation of a M.R.A.S. on an analog computer are also presented.
Maximizing Adaptivity in Hierarchical Topological Models Using Cancellation Trees
Bremer, P; Pascucci, V; Hamann, B
2008-12-08
We present a highly adaptive hierarchical representation of the topology of functions defined over two-manifold domains. Guided by the theory of Morse-Smale complexes, we encode dependencies between cancellations of critical points using two independent structures: a traditional mesh hierarchy to store connectivity information and a new structure called cancellation trees to encode the configuration of critical points. Cancellation trees provide a powerful method to increase adaptivity while using a simple, easy-to-implement data structure. The resulting hierarchy is significantly more flexible than the one previously reported. In particular, the resulting hierarchy is guaranteed to be of logarithmic height.
Hip Joint Replacement Using Monofilament Polypropylene Surgical Mesh: An Animal Model
Białecki, Jacek; Klimowicz-Bodys, Małgorzata Dorota; Wierzchoś, Edward; Kołomecki, Krzysztof
2014-01-01
Hip joint dysplasia is a deformation of the articular elements (pelvic acetabulum, head of the femur, and/or ligament of the head of the femur) leading to laxity of the hip components and dislocation of the femoral head from the pelvic acetabulum. Diagnosis is based on symptoms observed during clinical and radiological examinations. There are two treatment options: conservative and surgical. The classic surgical procedures are juvenile pubic symphysiodesis (JPS), triple pelvic osteotomy (TPO), total hip replacement (THR), and femoral head and neck resection (FHNE). The aim of this experiment was to present an original technique of filling the acetabulum with a polypropylene implant, resting the femoral neck directly on the mesh. The experiment was performed on eight sheep. The clinical value of the new surgical technique was evaluated using clinical, radiological, and histological methods. This technique helps decrease the loss of limb length by supporting the femoral neck on the mesh equivalent to the femoral head. It also reduces joint pain and leads to the formation of stable and mobile pseudarthrosis. The mesh manifested osteoprotective properties and enabled the formation of a stiff-elastic connection within the hip joint. The method is very cost-effective and the technique itself is simple to perform. PMID:24987672
Hip joint replacement using monofilament polypropylene surgical mesh: an animal model.
Białecki, Jacek; Majchrzycki, Marian; Szymczak, Antoni; Klimowicz-Bodys, Małgorzata Dorota; Wierzchoś, Edward; Kołomecki, Krzysztof
2014-01-01
Hip joint dysplasia is a deformation of the articular elements (pelvic acetabulum, head of the femur, and/or ligament of the head of the femur) leading to laxity of the hip components and dislocation of the femoral head from the pelvic acetabulum. Diagnosis is based on symptoms observed during clinical and radiological examinations. There are two treatment options: conservative and surgical. The classic surgical procedures are juvenile pubic symphysiodesis (JPS), triple pelvic osteotomy (TPO), total hip replacement (THR), and femoral head and neck resection (FHNE). The aim of this experiment was to present an original technique of filling the acetabulum with a polypropylene implant, resting the femoral neck directly on the mesh. The experiment was performed on eight sheep. The clinical value of the new surgical technique was evaluated using clinical, radiological, and histological methods. This technique helps decrease the loss of limb length by supporting the femoral neck on the mesh equivalent to the femoral head. It also reduces joint pain and leads to the formation of stable and mobile pseudarthrosis. The mesh manifested osteoprotective properties and enabled the formation of a stiff-elastic connection within the hip joint. The method is very cost-effective and the technique itself is simple to perform. PMID:24987672
NASA Astrophysics Data System (ADS)
Burman, E.; Jacot, A.; Picasso, M.
2004-03-01
A multiphase-field model for the description of coalescence in a binary alloy is solved numerically using adaptive finite elements with high aspect ratio. The unknown of the multiphase-field model are the three phase fields (solid phase 1, solid phase 2, and liquid phase), a Lagrange multiplier and the concentration field. An Euler implicit scheme is used for time discretization, together with continuous, piecewise linear finite elements. At each time step, a linear system corresponding to the three phases plus the Lagrange multiplier has to be solved. Then, the linear system pertaining to concentration is solved. An adaptive finite element algorithm is proposed. In order to reduce the number of mesh vertices, the generated meshes contain elements with high aspect ratio. The refinement and coarsening criteria are based on an error indicator which has already been justified theoretically for simpler problems. Numerical results on two test cases show the efficiency of the method.
Adaptive Behaviour Assessment System: Indigenous Australian Adaptation Model (ABAS: IAAM)
ERIC Educational Resources Information Center
du Plessis, Santie
2015-01-01
The study objectives were to develop, trial and evaluate a cross-cultural adaptation of the Adaptive Behavior Assessment System-Second Edition Teacher Form (ABAS-II TF) ages 5-21 for use with Indigenous Australian students ages 5-14. This study introduced a multiphase mixed-method design with semi-structured and informal interviews, school…
Adapting overcomplete wavelet models to natural images
NASA Astrophysics Data System (ADS)
Sallee, Phil; Olshausen, Bruno A.
2003-11-01
Overcomplete wavelet representations have become increasingly popular for their ability to provide highly sparse and robust descriptions of natural signals. We describe a method for incorporating an overcomplete wavelet representation as part of a statistical model of images which includes a sparse prior distribution over the wavelet coefficients. The wavelet basis functions are parameterized by a small set of 2-D functions. These functions are adapted to maximize the average log-likelihood of the model for a large database of natural images. When adapted to natural images, these functions become selective to different spatial orientations, and they achieve a superior degree of sparsity on natural images as compared with traditional wavelet bases. The learned basis is similar to the Steerable Pyramid basis, and yields slightly higher SNR for the same number of active coefficients. Inference with the learned model is demonstrated for applications such as denoising, with results that compare favorably with other methods.
Numerical simulation of H2/air detonation using unstructured mesh
NASA Astrophysics Data System (ADS)
Togashi, Fumiya; Löhner, Rainald; Tsuboi, Nobuyuki
2009-06-01
To explore the capability of unstructured mesh to simulate detonation wave propagation phenomena, numerical simulation of H2/air detonation using unstructured mesh was conducted. The unstructured mesh has several adv- antages such as easy mesh adaptation and flexibility to the complicated configurations. To examine the resolution dependency of the unstructured mesh, several simulations varying the mesh size were conducted and compared with a computed result using a structured mesh. The results show that the unstructured mesh solution captures the detailed structure of detonation wave, as well as the structured mesh solution. To capture the detailed detonation cell structure, the unstructured mesh simulations required at least twice, ideally 5times the resolution of structured mesh solution.
Quadrilateral/hexahedral finite element mesh coarsening
Staten, Matthew L; Dewey, Mark W; Scott, Michael A; Benzley, Steven E
2012-10-16
A technique for coarsening a finite element mesh ("FEM") is described. This technique includes identifying a coarsening region within the FEM to be coarsened. Perimeter chords running along perimeter boundaries of the coarsening region are identified. The perimeter chords are redirected to create an adaptive chord separating the coarsening region from a remainder of the FEM. The adaptive chord runs through mesh elements residing along the perimeter boundaries of the coarsening region. The adaptive chord is then extracted to coarsen the FEM.
6th International Meshing Roundtable '97
White, D.
1997-09-01
The goal of the 6th International Meshing Roundtable is to bring together researchers and developers from industry, academia, and government labs in a stimulating, open environment for the exchange of technical information related to the meshing process. In the pas~ the Roundtable has enjoyed significant participation born each of these groups from a wide variety of countries. The Roundtable will consist of technical presentations from contributed papers and abstracts, two invited speakers, and two invited panels of experts discussing topics related to the development and use of automatic mesh generation tools. In addition, this year we will feature a "Bring Your Best Mesh" competition and poster session to encourage discussion and participation from a wide variety of mesh generation tool users. The schedule and evening social events are designed to provide numerous opportunities for informal dialog. A proceedings will be published by Sandia National Laboratories and distributed at the Roundtable. In addition, papers of exceptionally high quaIity will be submitted to a special issue of the International Journal of Computational Geometry and Applications. Papers and one page abstracts were sought that present original results on the meshing process. Potential topics include but are got limited to: Unstructured triangular and tetrahedral mesh generation Unstructured quadrilateral and hexahedral mesh generation Automated blocking and structured mesh generation Mixed element meshing Surface mesh generation Geometry decomposition and clean-up techniques Geometry modification techniques related to meshing Adaptive mesh refinement and mesh quality control Mesh visualization Special purpose meshing algorithms for particular applications Theoretical or novel ideas with practical potential Technical presentations from industrial researchers.
Multi-Resolution Dynamic Meshes with Arbitrary Deformations
Shamir, A.; Pascucci, V.; Bajaj, C.
2000-07-10
Multi-resolution techniques and models have been shown to be effective for the display and transmission of large static geometric object. Dynamic environments with internally deforming models and scientific simulations using dynamic meshes pose greater challenges in terms of time and space, and need the development of similar solutions. In this paper we introduce the T-DAG, an adaptive multi-resolution representation for dynamic meshes with arbitrary deformations including attribute, position, connectivity and topology changes. T-DAG stands for Time-dependent Directed Acyclic Graph which defines the structure supporting this representation. We also provide an incremental algorithm (in time) for constructing the T-DAG representation of a given input mesh. This enables the traversal and use of the multi-resolution dynamic model for partial playback while still constructing new time-steps.
Augmentation mentoplasty using Mersilene mesh.
McCollough, E G; Hom, D B; Weigel, M T; Anderson, J R
1990-10-01
Many different materials are available for augmentation mentoplasty. However, the optimal implant material for chin implantation has yet to be found. During the past several years, a number of experienced surgeons have turned to the use of Mersilene mesh. Mersilene mesh is a non-absorbable Dacron polyester fiber that can be conformed easily into layers to achieve tailored dimensions and shape. At the McCollough Plastic Surgery Clinic PA, Birmingham, Ala, 277 patients over a 10-year period underwent chin augmentation with Mersilene mesh implants. The material provides excellent tensile strength, durability, and surgical adaptability. The overall complication rate was 3.2% (nine patients); infection rate, 2.5% (seven patients); and removal secondary to infection, 1.7% (five patients). Based on this 10-year experience, Mersilene mesh remains our material of choice for chin augmentation. PMID:2206500
Dynamic Load Balancing for Adaptive Unstructured Grids
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Saini, Subhash (Technical Monitor)
1998-01-01
Dynamic mesh adaptation on unstructured grids is a powerful tool for computing unsteady three-dimensional problems that require grid modifications to efficiently resolve solution features. By locally refining and coarsening the mesh to capture phenomena of interest, such procedures make standard computational methods more cost effective. Highly refined meshes are required to accurately capture shock waves, contact discontinuities, vortices, and shear layers in fluid flow problems. Adaptive meshes have also proved to be useful in several other areas of computational science and engineering like computer vision and graphics, semiconductor device modeling, and structural mechanics. Local mesh adaptation provides the opportunity to obtain solutions that are comparable to those obtained on globally-refined grids but at a much lower cost. Additional information is contained in the original extended abstract.
Adaptive Control with Reference Model Modification
NASA Technical Reports Server (NTRS)
Stepanyan, Vahram; Krishnakumar, Kalmanje
2012-01-01
This paper presents a modification of the conventional model reference adaptive control (MRAC) architecture in order to improve transient performance of the input and output signals of uncertain systems. A simple modification of the reference model is proposed by feeding back the tracking error signal. It is shown that the proposed approach guarantees tracking of the given reference command and the reference control signal (one that would be designed if the system were known) not only asymptotically but also in transient. Moreover, it prevents generation of high frequency oscillations, which are unavoidable in conventional MRAC systems for large adaptation rates. The provided design guideline makes it possible to track a reference commands of any magnitude from any initial position without re-tuning. The benefits of the method are demonstrated with a simulation example
Adaptive cyber-attack modeling system
NASA Astrophysics Data System (ADS)
Gonsalves, Paul G.; Dougherty, Edward T.
2006-05-01
The pervasiveness of software and networked information systems is evident across a broad spectrum of business and government sectors. Such reliance provides an ample opportunity not only for the nefarious exploits of lone wolf computer hackers, but for more systematic software attacks from organized entities. Much effort and focus has been placed on preventing and ameliorating network and OS attacks, a concomitant emphasis is required to address protection of mission critical software. Typical software protection technique and methodology evaluation and verification and validation (V&V) involves the use of a team of subject matter experts (SMEs) to mimic potential attackers or hackers. This manpower intensive, time-consuming, and potentially cost-prohibitive approach is not amenable to performing the necessary multiple non-subjective analyses required to support quantifying software protection levels. To facilitate the evaluation and V&V of software protection solutions, we have designed and developed a prototype adaptive cyber attack modeling system. Our approach integrates an off-line mechanism for rapid construction of Bayesian belief network (BN) attack models with an on-line model instantiation, adaptation and knowledge acquisition scheme. Off-line model construction is supported via a knowledge elicitation approach for identifying key domain requirements and a process for translating these requirements into a library of BN-based cyber-attack models. On-line attack modeling and knowledge acquisition is supported via BN evidence propagation and model parameter learning.
NASA Astrophysics Data System (ADS)
Dahdouh, S.; Varsier, N.; Serrurier, A.; De la Plata, J.-P.; Anquez, J.; Angelini, E. D.; Wiart, J.; Bloch, I.
2014-08-01
Fetal dosimetry studies require the development of accurate numerical 3D models of the pregnant woman and the fetus. This paper proposes a 3D articulated fetal growth model covering the main phases of pregnancy and a pregnant woman model combining the utero-fetal structures and a deformable non-pregnant woman body envelope. The structures of interest were automatically or semi-automatically (depending on the stage of pregnancy) segmented from a database of images and surface meshes were generated. By interpolating linearly between fetal structures, each one can be generated at any age and in any position. A method is also described to insert the utero-fetal structures in the maternal body. A validation of the fetal models is proposed, comparing a set of biometric measurements to medical reference charts. The usability of the pregnant woman model in dosimetry studies is also investigated, with respect to the influence of the abdominal fat layer.
Adaptive human behavior in epidemiological models.
Fenichel, Eli P; Castillo-Chavez, Carlos; Ceddia, M G; Chowell, Gerardo; Parra, Paula A Gonzalez; Hickling, Graham J; Holloway, Garth; Horan, Richard; Morin, Benjamin; Perrings, Charles; Springborn, Michael; Velazquez, Leticia; Villalobos, Cristina
2011-04-12
The science and management of infectious disease are entering a new stage. Increasingly public policy to manage epidemics focuses on motivating people, through social distancing policies, to alter their behavior to reduce contacts and reduce public disease risk. Person-to-person contacts drive human disease dynamics. People value such contacts and are willing to accept some disease risk to gain contact-related benefits. The cost-benefit trade-offs that shape contact behavior, and hence the course of epidemics, are often only implicitly incorporated in epidemiological models. This approach creates difficulty in parsing out the effects of adaptive behavior. We use an epidemiological-economic model of disease dynamics to explicitly model the trade-offs that drive person-to-person contact decisions. Results indicate that including adaptive human behavior significantly changes the predicted course of epidemics and that this inclusion has implications for parameter estimation and interpretation and for the development of social distancing policies. Acknowledging adaptive behavior requires a shift in thinking about epidemiological processes and parameters. PMID:21444809
Adaptive human behavior in epidemiological models
Fenichel, Eli P.; Castillo-Chavez, Carlos; Ceddia, M. G.; Chowell, Gerardo; Parra, Paula A. Gonzalez; Hickling, Graham J.; Holloway, Garth; Horan, Richard; Morin, Benjamin; Perrings, Charles; Springborn, Michael; Velazquez, Leticia; Villalobos, Cristina
2011-01-01
The science and management of infectious disease are entering a new stage. Increasingly public policy to manage epidemics focuses on motivating people, through social distancing policies, to alter their behavior to reduce contacts and reduce public disease risk. Person-to-person contacts drive human disease dynamics. People value such contacts and are willing to accept some disease risk to gain contact-related benefits. The cost–benefit trade-offs that shape contact behavior, and hence the course of epidemics, are often only implicitly incorporated in epidemiological models. This approach creates difficulty in parsing out the effects of adaptive behavior. We use an epidemiological–economic model of disease dynamics to explicitly model the trade-offs that drive person-to-person contact decisions. Results indicate that including adaptive human behavior significantly changes the predicted course of epidemics and that this inclusion has implications for parameter estimation and interpretation and for the development of social distancing policies. Acknowledging adaptive behavior requires a shift in thinking about epidemiological processes and parameters. PMID:21444809
Modeling and adaptive control of acoustic noise
NASA Astrophysics Data System (ADS)
Venugopal, Ravinder
Active noise control is a problem that receives significant attention in many areas including aerospace and manufacturing. The advent of inexpensive high performance processors has made it possible to implement real-time control algorithms to effect active noise control. Both fixed-gain and adaptive methods may be used to design controllers for this problem. For fixed-gain methods, it is necessary to obtain a mathematical model of the system to design controllers. In addition, models help us gain phenomenological insights into the dynamics of the system. Models are also necessary to perform numerical simulations. However, models are often inadequate for the purpose of controller design because they involve parameters that are difficult to determine and also because there are always unmodeled effects. This fact motivates the use of adaptive algorithms for control since adaptive methods usually require significantly less model information than fixed-gain methods. The first part of this dissertation deals with derivation of a state space model of a one-dimensional acoustic duct. Two types of actuation, namely, a side-mounted speaker (interior control) and an end-mounted speaker (boundary control) are considered. The techniques used to derive the model of the acoustic duct are extended to the problem of fluid surface wave control. A state space model of small amplitude surfaces waves of a fluid in a rectangular container is derived and two types of control methods, namely, surface pressure control and map actuator based control are proposed and analyzed. The second part of this dissertation deals with the development of an adaptive disturbance rejection algorithm that is applied to the problem of active noise control. ARMARKOV models which have the same structure as predictor models are used for system representation. The algorithm requires knowledge of only one path of the system, from control to performance, and does not require a measurement of the disturbance nor
A new spectral finite volume method for elastic wave modelling on unstructured meshes
NASA Astrophysics Data System (ADS)
Zhang, Wensheng; Zhuang, Yuan; Chung, Eric T.
2016-04-01
In this paper, we consider a new spectral finite volume method for the elastic wave equations. Our new finite volume method is based on a piecewise constant approximation on a fine mesh and a high-order polynomial reconstruction on a coarser mesh. Our new method is constructed based on two existing techniques, the high-order finite volume method and the spectral finite volume method. In fact, we will construct a new method to take advantage of both methods. More precisely, our method has two distinctive features. The first one is that the local polynomial reconstructions are performed on the coarse triangles, and the reconstruction matrices for all the coarse triangles are the same. This fact enhances the parallelization of our algorithm. We will present a parallel implementation of our method and show excellent efficiency results. The second one is that, by using a suitable number of finer triangles with a coarse triangle, we obtain an over-determined reconstruction system, which can enhance the robustness of the reconstruction process. To derive our scheme, standard finite volume technique is applied to each fine triangle, and the high-order reconstructed polynomials, computed on coarse triangles, are used to compute numerical fluxes. We will present numerical results to show the performance of our method. Our method is presented for 2D problems, but the same methodology can be applied to 3D.
A new spectral finite volume method for elastic wave modelling on unstructured meshes
NASA Astrophysics Data System (ADS)
Zhang, Wensheng; Zhuang, Yuan; Chung, Eric T.
2016-07-01
In this paper, we consider a new spectral finite volume method (FVM) for the elastic wave equations. Our new FVM is based on a piecewise constant approximation on a fine mesh and a high-order polynomial reconstruction on a coarser mesh. Our new method is constructed based on two existing techniques, the high-order FVM and the spectral FVM. In fact, we will construct a new method to take advantage of both methods. More precisely, our method has two distinctive features. The first one is that the local polynomial reconstructions are performed on the coarse triangles and the reconstruction matrices for all the coarse triangles are the same. This fact enhances the parallelization of our algorithm. We will present a parallel implementation of our method and show excellent efficiency results. The second one is that, by using a suitable number of finer triangles with a coarse triangle, we obtain an overdetermined reconstruction system, which can enhance the robustness of the reconstruction process. To derive our scheme, standard finite volume technique is applied to each fine triangle, and the high-order reconstructed polynomials, computed on coarse triangles, are used to compute numerical fluxes. We will present numerical results to show the performance of our method. Our method is presented for 2-D problems, but the same methodology can be applied to 3-D.
Turing pattern dynamics and adaptive discretization for a super-diffusive Lotka-Volterra model.
Bendahmane, Mostafa; Ruiz-Baier, Ricardo; Tian, Canrong
2016-05-01
In this paper we analyze the effects of introducing the fractional-in-space operator into a Lotka-Volterra competitive model describing population super-diffusion. First, we study how cross super-diffusion influences the formation of spatial patterns: a linear stability analysis is carried out, showing that cross super-diffusion triggers Turing instabilities, whereas classical (self) super-diffusion does not. In addition we perform a weakly nonlinear analysis yielding a system of amplitude equations, whose study shows the stability of Turing steady states. A second goal of this contribution is to propose a fully adaptive multiresolution finite volume method that employs shifted Grünwald gradient approximations, and which is tailored for a larger class of systems involving fractional diffusion operators. The scheme is aimed at efficient dynamic mesh adaptation and substantial savings in computational burden. A numerical simulation of the model was performed near the instability boundaries, confirming the behavior predicted by our analysis. PMID:26219250
Criman, Erik T.; Kurata, Wendy E.; Matsumoto, Karen W.; Aubin, Harry T.; Campbell, Carmen E.
2016-01-01
Background: The reported incidence of mesh infection in contaminated operative fields is as high as 30% regardless of the material used. Recently, mesenchymal stem cells (MSCs) have been shown to possess favorable immunomodulatory properties and improve tissue incorporation when seeded onto bioprosthetics. The aim of this study was to evaluate whether seeding noncrosslinked bovine pericardium (Veritas Collagen Matrix) with allogeneic bone marrow–derived MSCs improves infection resistance in vivo after inoculation with Escherichia coli (E. coli). Methods: Rat bone marrow–derived MSCs at passage 3 were seeded onto bovine pericardium and cultured for 7 days before implantation. Additional rats (n = 24) were implanted subcutaneously with MSC-seeded or unseeded mesh and inoculated with 7 × 105 colony-forming units of E. coli or saline before wound closure (group 1, unseeded mesh/saline; group 2, unseeded mesh/E. coli; group 3, MSC-seeded mesh/E. coli; 8 rats per group). Meshes were explanted at 4 weeks and underwent microbiologic and histologic analyses. Results: MSC-seeded meshes inoculated with E. coli demonstrated superior bacterial clearance and preservation of mesh integrity compared with E. coli–inoculated unseeded meshes (87.5% versus 0% clearance; p = 0.001). Complete mesh degradation concurrent with abscess formation was observed in 100% of rats in the unseeded/E. coli group, which is in contrast to 12.5% of rats in the MSC-seeded/E. coli group. Histologic evaluation determined that remodeling characteristics of E. coli–inoculated MSC-seeded meshes were similar to those of uninfected meshes 4 weeks after implantation. Conclusions: Augmenting a bioprosthetic material with stem cells seems to markedly enhance resistance to bacterial infection in vivo and preserve mesh integrity. PMID:27482490
Controlling Reflections from Mesh Refinement Interfaces in Numerical Relativity
NASA Technical Reports Server (NTRS)
Baker, John G.; Van Meter, James R.
2005-01-01
A leading approach to improving the accuracy on numerical relativity simulations of black hole systems is through fixed or adaptive mesh refinement techniques. We describe a generic numerical error which manifests as slowly converging, artificial reflections from refinement boundaries in a broad class of mesh-refinement implementations, potentially limiting the effectiveness of mesh- refinement techniques for some numerical relativity applications. We elucidate this numerical effect by presenting a model problem which exhibits the phenomenon, but which is simple enough that its numerical error can be understood analytically. Our analysis shows that the effect is caused by variations in finite differencing error generated across low and high resolution regions, and that its slow convergence is caused by the presence of dramatic speed differences among propagation modes typical of 3+1 relativity. Lastly, we resolve the problem, presenting a class of finite-differencing stencil modifications which eliminate this pathology in both our model problem and in numerical relativity examples.
Computing The Compliances Of Gear Meshes
NASA Technical Reports Server (NTRS)
Lewicki, D. G.; Savage, M.; Caldwell, R. J.; Wisor, G. D.
1988-01-01
Computer model simulates compliance and sharing of loads in spur-gear mesh. Use of solid-body analysis as lower bound and rim analysis as upper bound for mesh compliance, reasonable approximations obtained for compliance in spur-gear mesh.
NASA Astrophysics Data System (ADS)
Wong, M.; Skamarock, W. C.
2015-12-01
Global numerical weather forecast tests were performed using the global nonhydrostatic atmospheric model, Model for Prediction Across Scales (MPAS), for the NOAA Storm Prediction Center 2015 Spring Forecast Experiment (May 2015) and the Plains Elevated Convection at Night (PECAN) field campaign (June to mid-July 2015). These two sets of forecasts were performed on 50-to-3 km and 15-to-3 km smoothly-varying horizontal meshes, respectively. Both variable-resolution meshes have nominal convection-permitting 3-km grid spacing over the entire continental US. Here we evaluate the limited-area (vs. global) spectra from these NWP simulations. We will show the simulated spectral characteristics of total kinetic energy, vertical velocity variance, and precipitation during these spring and summer periods when diurnal continental convection is most active over central US. Spectral characteristics of a high-resolution global 3-km simulation (essentially no nesting) from the 20 May 2013 Moore, OK tornado case are also shown. These characteristics include spectral scaling, shape, and anisotropy, as well as the effective resolution of continental convection representation in MPAS.
Hybrid mesh for nasal airflow studies.
Zubair, Mohammed; Abdullah, Mohammed Zulkifly; Ahmad, Kamarul Arifin
2013-01-01
The accuracy of the numerical result is closely related to mesh density as well as its distribution. Mesh plays a very significant role in the outcome of numerical simulation. Many nasal airflow studies have employed unstructured mesh and more recently hybrid mesh scheme has been utilized considering the complexity of anatomical architecture. The objective of this study is to compare the results of hybrid mesh with unstructured mesh and study its effect on the flow parameters inside the nasal cavity. A three-dimensional nasal cavity model is reconstructed based on computed tomographic images of a healthy Malaysian adult nose. Navier-Stokes equation for steady airflow is solved numerically to examine inspiratory nasal flow. The pressure drop obtained using the unstructured computational grid is about 22.6 Pa for a flow rate of 20 L/min, whereas the hybrid mesh resulted in 17.8 Pa for the same flow rate. The maximum velocity obtained at the nasal valve using unstructured grid is 4.18 m/s and that with hybrid mesh is around 4.76 m/s. Hybrid mesh reported lower grid convergence index (GCI) than the unstructured mesh. Significant differences between unstructured mesh and hybrid mesh are determined highlighting the usefulness of hybrid mesh for nasal airflow studies. PMID:23983811
Hybrid Mesh for Nasal Airflow Studies
Zubair, Mohammed; Abdullah, Mohammed Zulkifly; Ahmad, Kamarul Arifin
2013-01-01
The accuracy of the numerical result is closely related to mesh density as well as its distribution. Mesh plays a very significant role in the outcome of numerical simulation. Many nasal airflow studies have employed unstructured mesh and more recently hybrid mesh scheme has been utilized considering the complexity of anatomical architecture. The objective of this study is to compare the results of hybrid mesh with unstructured mesh and study its effect on the flow parameters inside the nasal cavity. A three-dimensional nasal cavity model is reconstructed based on computed tomographic images of a healthy Malaysian adult nose. Navier-Stokes equation for steady airflow is solved numerically to examine inspiratory nasal flow. The pressure drop obtained using the unstructured computational grid is about 22.6 Pa for a flow rate of 20 L/min, whereas the hybrid mesh resulted in 17.8 Pa for the same flow rate. The maximum velocity obtained at the nasal valve using unstructured grid is 4.18 m/s and that with hybrid mesh is around 4.76 m/s. Hybrid mesh reported lower grid convergence index (GCI) than the unstructured mesh. Significant differences between unstructured mesh and hybrid mesh are determined highlighting the usefulness of hybrid mesh for nasal airflow studies. PMID:23983811
An adaptive contextual quantum language model
NASA Astrophysics Data System (ADS)
Li, Jingfei; Zhang, Peng; Song, Dawei; Hou, Yuexian
2016-08-01
User interactions in search system represent a rich source of implicit knowledge about the user's cognitive state and information need that continuously evolves over time. Despite massive efforts that have been made to exploiting and incorporating this implicit knowledge in information retrieval, it is still a challenge to effectively capture the term dependencies and the user's dynamic information need (reflected by query modifications) in the context of user interaction. To tackle these issues, motivated by the recent Quantum Language Model (QLM), we develop a QLM based retrieval model for session search, which naturally incorporates the complex term dependencies occurring in user's historical queries and clicked documents with density matrices. In order to capture the dynamic information within users' search session, we propose a density matrix transformation framework and further develop an adaptive QLM ranking model. Extensive comparative experiments show the effectiveness of our session quantum language models.