NASA Astrophysics Data System (ADS)
Leboeuf, Jean-Noel; Decyk, Viktor; Newman, David; Sanchez, Raul
2012-03-01
The massively parallel, nonlinear, 3D, toroidal, electrostatic, gyrokinetic, PIC, Cartesian geometry UCAN code, with particle ions and adiabatic electrons, has been successfully exercised to identify non-diffusive transport characteristics in DIII-D-like tokamak discharges. The limitation in applying UCAN to larger scale discharges is the 1D domain decomposition in the toroidal (or z-) direction for massively parallel implementation using MPI which has restricted the calculations to a few hundred ion Larmor radii per minor radius. To exceed these sizes, we have implemented 2D domain decomposition in UCAN with the addition of the y-direction to the processor mix. This has been facilitated by use of relevant components in the 2D domain decomposed PLIB2 library of field and particle management routines developed for UCLA's UPIC framework of conventional PIC codes. The gyro-averaging in gyrokinetic codes has necessitated the use of replicated arrays for efficient charge accumulation and particle push. The 2D domain-decomposed UCAN2 code reproduces the original 1D domain results within roundoff. Production calculations at large system sizes have been performed with UCAN2 on 131072 processors of the Cray XE6 at NERSC.
NASA Astrophysics Data System (ADS)
Sidler, Rolf; Carcione, José M.; Holliger, Klaus
2013-02-01
We present a novel numerical approach for the comprehensive, flexible, and accurate simulation of poro-elastic wave propagation in 2D polar coordinates. An important application of this method and its extensions will be the modeling of complex seismic wave phenomena in fluid-filled boreholes, which represents a major, and as of yet largely unresolved, computational problem in exploration geophysics. In view of this, we consider a numerical mesh, which can be arbitrarily heterogeneous, consisting of two or more concentric rings representing the fluid in the center and the surrounding porous medium. The spatial discretization is based on a Chebyshev expansion in the radial direction and a Fourier expansion in the azimuthal direction and a Runge-Kutta integration scheme for the time evolution. A domain decomposition method is used to match the fluid-solid boundary conditions based on the method of characteristics. This multi-domain approach allows for significant reductions of the number of grid points in the azimuthal direction for the inner grid domain and thus for corresponding increases of the time step and enhancements of computational efficiency. The viability and accuracy of the proposed method has been rigorously tested and verified through comparisons with analytical solutions as well as with the results obtained with a corresponding, previously published, and independently benchmarked solution for 2D Cartesian coordinates. Finally, the proposed numerical solution also satisfies the reciprocity theorem, which indicates that the inherent singularity associated with the origin of the polar coordinate system is adequately handled.
Sidler, Rolf; Carcione, José M.; Holliger, Klaus
2013-02-15
We present a novel numerical approach for the comprehensive, flexible, and accurate simulation of poro-elastic wave propagation in 2D polar coordinates. An important application of this method and its extensions will be the modeling of complex seismic wave phenomena in fluid-filled boreholes, which represents a major, and as of yet largely unresolved, computational problem in exploration geophysics. In view of this, we consider a numerical mesh, which can be arbitrarily heterogeneous, consisting of two or more concentric rings representing the fluid in the center and the surrounding porous medium. The spatial discretization is based on a Chebyshev expansion in the radial direction and a Fourier expansion in the azimuthal direction and a Runge–Kutta integration scheme for the time evolution. A domain decomposition method is used to match the fluid–solid boundary conditions based on the method of characteristics. This multi-domain approach allows for significant reductions of the number of grid points in the azimuthal direction for the inner grid domain and thus for corresponding increases of the time step and enhancements of computational efficiency. The viability and accuracy of the proposed method has been rigorously tested and verified through comparisons with analytical solutions as well as with the results obtained with a corresponding, previously published, and independently benchmarked solution for 2D Cartesian coordinates. Finally, the proposed numerical solution also satisfies the reciprocity theorem, which indicates that the inherent singularity associated with the origin of the polar coordinate system is adequately handled.
NASA Astrophysics Data System (ADS)
Tønning, Erik; Polders, Daniel; Callaghan, Paul T.; Engelsen, Søren B.
2007-09-01
This paper demonstrates how the multi-linear PARAFAC model can with advantage be used to decompose 2D diffusion-relaxation correlation NMR spectra prior to 2D-Laplace inversion to the T2- D domain. The decomposition is advantageous for better interpretation of the complex correlation maps as well as for the quantification of extracted T2- D components. To demonstrate the new method seventeen mixtures of wheat flour, starch, gluten, oil and water were prepared and measured with a 300 MHz nuclear magnetic resonance (NMR) spectrometer using a pulsed gradient stimulated echo (PGSTE) pulse sequence followed by a Carr-Purcell-Meiboom-Gill (CPMG) pulse echo train. By varying the gradient strength, 2D diffusion-relaxation data were recorded for each sample. From these double exponentially decaying relaxation data the PARAFAC algorithm extracted two unique diffusion-relaxation components, explaining 99.8% of the variation in the data set. These two components were subsequently transformed to the T2- D domain using 2D-inverse Laplace transformation and quantitatively assigned to the oil and water components of the samples. The oil component was one distinct distribution with peak intensity at D = 3 × 10 -12 m 2 s -1 and T2 = 180 ms. The water component consisted of two broad populations of water molecules with diffusion coefficients and relaxation times centered around correlation pairs: D = 10 -9 m 2 s -1, T2 = 10 ms and D = 3 × 10 -13 m 2 s -1, T2 = 13 ms. Small spurious peaks observed in the inverse Laplace transformation of original complex data were effectively filtered by the PARAFAC decomposition and thus considered artefacts from the complex Laplace transformation. The oil-to-water ratio determined by PARAFAC followed by 2D-Laplace inversion was perfectly correlated with known oil-to-water ratio of the samples. The new method of using PARAFAC prior to the 2D-Laplace inversion proved to have superior potential in analysis of diffusion-relaxation spectra, as it
Domain decomposition for the SPN solver MINOS
Jamelot, Erell; Baudron, Anne-Marie; Lautard, Jean-Jacques
2012-07-01
In this article we present a domain decomposition method for the mixed SPN equations, discretized with Raviart-Thomas-Nedelec finite elements. This domain decomposition is based on the iterative Schwarz algorithm with Robin interface conditions to handle communications. After having described this method, we give details on how to optimize the convergence. Finally, we give some numerical results computed in a realistic 3D domain. The computations are done with the MINOS solver of the APOLLO3 (R) code. (authors)
Multilevel domain decomposition for electronic structure calculations
Barrault, M. . E-mail: maxime.barrault@edf.fr; Cances, E. . E-mail: cances@cermics.enpc.fr; Hager, W.W. . E-mail: hager@math.ufl.edu; Le Bris, C. . E-mail: lebris@cermics.enpc.fr
2007-03-01
We introduce a new multilevel domain decomposition method (MDD) for electronic structure calculations within semi-empirical and density functional theory (DFT) frameworks. This method iterates between local fine solvers and global coarse solvers, in the spirit of domain decomposition methods. Using this approach, calculations have been successfully performed on several linear polymer chains containing up to 40,000 atoms and 200,000 atomic orbitals. Both the computational cost and the memory requirement scale linearly with the number of atoms. Additional speed-up can easily be obtained by parallelization. We show that this domain decomposition method outperforms the density matrix minimization (DMM) method for poor initial guesses. Our method provides an efficient preconditioner for DMM and other linear scaling methods, variational in nature, such as the orbital minimization (OM) procedure.
Convergence Analysis of a Domain Decomposition Paradigm
Bank, R E; Vassilevski, P S
2006-06-12
We describe a domain decomposition algorithm for use in several variants of the parallel adaptive meshing paradigm of Bank and Holst. This algorithm has low communication, makes extensive use of existing sequential solvers, and exploits in several important ways data generated as part of the adaptive meshing paradigm. We show that for an idealized version of the algorithm, the rate of convergence is independent of both the global problem size N and the number of subdomains p used in the domain decomposition partition. Numerical examples illustrate the effectiveness of the procedure.
Layer tracking, asymptotics, and domain decomposition
NASA Technical Reports Server (NTRS)
Brown, D. L.; Chin, R. C. Y.; Hedstrom, G. W.; Manteuffel, T. A.
1991-01-01
A preliminary report is presented on the work on the tracking of internal layers in a singularly-perturbed convection-diffusion equation. It is shown why such tracking may be desirable, and it is also shown how to do it using domain decomposition based on asymptotic analysis.
Domain decomposition methods for mortar finite elements
Widlund, O.
1996-12-31
In the last few years, domain decomposition methods, previously developed and tested for standard finite element methods and elliptic problems, have been extended and modified to work for mortar and other nonconforming finite element methods. A survey will be given of work carried out jointly with Yves Achdou, Mario Casarin, Maksymilian Dryja and Yvon Maday. Results on the p- and h-p-version finite elements will also be discussed.
Domain decomposition multigrid for unstructured grids
Shapira, Yair
1997-01-01
A two-level preconditioning method for the solution of elliptic boundary value problems using finite element schemes on possibly unstructured meshes is introduced. It is based on a domain decomposition and a Galerkin scheme for the coarse level vertex unknowns. For both the implementation and the analysis, it is not required that the curves of discontinuity in the coefficients of the PDE match the interfaces between subdomains. Generalizations to nonmatching or overlapping grids are made.
Protein Domain Decomposition Using a Graph-Theoretic Approach
Xu, Y.; Xu, D.; Gabow, H.N.
2000-08-20
This paper presents a new algorithm for the decomposition of a multi-domain protein into individual structural domains. The underlying principle used is that residue-residue contacts are denser within a domain than between domains.
Combinatorial geometry domain decomposition strategies for Monte Carlo simulations
Li, G.; Zhang, B.; Deng, L.; Mo, Z.; Liu, Z.; Shangguan, D.; Ma, Y.; Li, S.; Hu, Z.
2013-07-01
Analysis and modeling of nuclear reactors can lead to memory overload for a single core processor when it comes to refined modeling. A method to solve this problem is called 'domain decomposition'. In the current work, domain decomposition algorithms for a combinatorial geometry Monte Carlo transport code are developed on the JCOGIN (J Combinatorial Geometry Monte Carlo transport INfrastructure). Tree-based decomposition and asynchronous communication of particle information between domains are described in the paper. Combination of domain decomposition and domain replication (particle parallelism) is demonstrated and compared with that of MERCURY code. A full-core reactor model is simulated to verify the domain decomposition algorithms using the Monte Carlo particle transport code JMCT (J Monte Carlo Transport Code), which has being developed on the JCOGIN infrastructure. Besides, influences of the domain decomposition algorithms to tally variances are discussed. (authors)
Layout decomposition of self-aligned double patterning for 2D random logic patterning
NASA Astrophysics Data System (ADS)
Ban, Yongchan; Miloslavsky, Alex; Lucas, Kevin; Choi, Soo-Han; Park, Chul-Hong; Pan, David Z.
2011-04-01
Self-aligned double pattering (SADP) has been adapted as a promising solution for sub-30nm technology nodes due to its lower overlay problem and better process tolerance. SADP is in production use for 1D dense patterns with good pitch control such as NAND Flash memory applications, but it is still challenging to apply SADP to 2D random logic patterns. The favored type of SADP for complex logic interconnects is a two mask approach using a core mask and a trim mask. In this paper, we first describe layout decomposition methods of spacer-type double patterning lithography, then report a type of SADP compliant layouts, and finally report SADP applications on Samsung 22nm SRAM layout. For SADP decomposition, we propose several SADP-aware layout coloring algorithms and a method of generating lithography-friendly core mask patterns. Experimental results on 22nm node designs show that our proposed layout decomposition for SADP effectively decomposes any given layouts.
Domain decomposition methods in computational fluid dynamics
NASA Technical Reports Server (NTRS)
Gropp, William D.; Keyes, David E.
1992-01-01
The divide-and-conquer paradigm of iterative domain decomposition, or substructuring, has become a practical tool in computational fluid dynamic applications because of its flexibility in accommodating adaptive refinement through locally uniform (or quasi-uniform) grids, its ability to exploit multiple discretizations of the operator equations, and the modular pathway it provides towards parallelism. These features are illustrated on the classic model problem of flow over a backstep using Newton's method as the nonlinear iteration. Multiple discretizations (second-order in the operator and first-order in the preconditioner) and locally uniform mesh refinement pay dividends separately, and they can be combined synergistically. Sample performance results are included from an Intel iPSC/860 hypercube implementation.
Domain decomposition methods in computational fluid dynamics
NASA Technical Reports Server (NTRS)
Gropp, William D.; Keyes, David E.
1991-01-01
The divide-and-conquer paradigm of iterative domain decomposition, or substructuring, has become a practical tool in computational fluid dynamic applications because of its flexibility in accommodating adaptive refinement through locally uniform (or quasi-uniform) grids, its ability to exploit multiple discretizations of the operator equations, and the modular pathway it provides towards parallelism. These features are illustrated on the classic model problem of flow over a backstep using Newton's method as the nonlinear iteration. Multiple discretizations (second-order in the operator and first-order in the preconditioner) and locally uniform mesh refinement pay dividends separately, and they can be combined synergistically. Sample performance results are included from an Intel iPSC/860 hypercube implementation.
Domain decomposition algorithms and computation fluid dynamics
NASA Technical Reports Server (NTRS)
Chan, Tony F.
1988-01-01
In the past several years, domain decomposition was a very popular topic, partly motivated by the potential of parallelization. While a large body of theory and algorithms were developed for model elliptic problems, they are only recently starting to be tested on realistic applications. The application of some of these methods to two model problems in computational fluid dynamics are investigated. Some examples are two dimensional convection-diffusion problems and the incompressible driven cavity flow problem. The construction and analysis of efficient preconditioners for the interface operator to be used in the iterative solution of the interface solution is described. For the convection-diffusion problems, the effect of the convection term and its discretization on the performance of some of the preconditioners is discussed. For the driven cavity problem, the effectiveness of a class of boundary probe preconditioners is discussed.
A computational model of the short-cut rule for 2D shape decomposition.
Luo, Lei; Shen, Chunhua; Liu, Xinwang; Zhang, Chunyuan
2015-01-01
We propose a new 2D shape decomposition method based on the short-cut rule. The short-cut rule originates from cognition research, and states that the human visual system prefers to partition an object into parts using the shortest possible cuts. We propose and implement a computational model for the short-cut rule and apply it to the problem of shape decomposition. The model we proposed generates a set of cut hypotheses passing through the points on the silhouette, which represent the negative minima of curvature. We then show that most part-cut hypotheses can be eliminated by analysis of local properties of each. Finally, the remaining hypotheses are evaluated in ascending length order, which guarantees that of any pair of conflicting cuts only the shortest will be accepted. We demonstrate that, compared with state-of-the-art shape decomposition methods, the proposed approach achieves decomposition results, which better correspond to human intuition as revealed in psychological experiments. PMID:25438318
Domain Decomposition By the Advancing-Partition Method
NASA Technical Reports Server (NTRS)
Pirzadeh, Shahyar Z.
2008-01-01
A new method of domain decomposition has been developed for generating unstructured grids in subdomains either sequentially or using multiple computers in parallel. Domain decomposition is a crucial and challenging step for parallel grid generation. Prior methods are generally based on auxiliary, complex, and computationally intensive operations for defining partition interfaces and usually produce grids of lower quality than those generated in single domains. The new technique, referred to as "Advancing Partition," is based on the Advancing-Front method, which partitions a domain as part of the volume mesh generation in a consistent and "natural" way. The benefits of this approach are: 1) the process of domain decomposition is highly automated, 2) partitioning of domain does not compromise the quality of the generated grids, and 3) the computational overhead for domain decomposition is minimal. The new method has been implemented in NASA's unstructured grid generation code VGRID.
Dynamic load balancing algorithm for molecular dynamics based on Voronoi cells domain decompositions
Fattebert, J.-L.; Richards, D.F.; Glosli, J.N.
2012-12-01
We present a new algorithm for automatic parallel load balancing in classical molecular dynamics. It assumes a spatial domain decomposition of particles into Voronoi cells. It is a gradient method which attempts to minimize a cost function by displacing Voronoi sites associated with each processor/sub-domain along steepest descent directions. Excellent load balance has been obtained for quasi-2D and 3D practical applications, with up to 440·10^{6} particles on 65,536 MPI tasks.
Interface conditions for domain decomposition with radical grid refinement
NASA Technical Reports Server (NTRS)
Scroggs, Jeffrey S.
1991-01-01
Interface conditions for coupling the domains in a physically motivated domain decomposition method are discussed. The domain decomposition is based on an asymptotic-induced method for the numerical solution of hyperbolic conservation laws with small viscosity. The method consists of multiple stages. The first stage is to obtain a first approximation using a first-order method, such as the Godunov scheme. Subsequent stages of the method involve solving internal-layer problem via a domain decomposition. The method is derived and justified via singular perturbation techniques.
Domain decomposition: A bridge between nature and parallel computers
NASA Technical Reports Server (NTRS)
Keyes, David E.
1992-01-01
Domain decomposition is an intuitive organizing principle for a partial differential equation (PDE) computation, both physically and architecturally. However, its significance extends beyond the readily apparent issues of geometry and discretization, on one hand, and of modular software and distributed hardware, on the other. Engineering and computer science aspects are bridged by an old but recently enriched mathematical theory that offers the subject not only unity, but also tools for analysis and generalization. Domain decomposition induces function-space and operator decompositions with valuable properties. Function-space bases and operator splittings that are not derived from domain decompositions generally lack one or more of these properties. The evolution of domain decomposition methods for elliptically dominated problems has linked two major algorithmic developments of the last 15 years: multilevel and Krylov methods. Domain decomposition methods may be considered descendants of both classes with an inheritance from each: they are nearly optimal and at the same time efficiently parallelizable. Many computationally driven application areas are ripe for these developments. A progression is made from a mathematically informal motivation for domain decomposition methods to a specific focus on fluid dynamics applications. To be introductory rather than comprehensive, simple examples are provided while convergence proofs and algorithmic details are left to the original references; however, an attempt is made to convey their most salient features, especially where this leads to algorithmic insight.
Domain decomposition for aerodynamic and aeroacoustic analyses, and optimization
NASA Technical Reports Server (NTRS)
Baysal, Oktay
1995-01-01
The overarching theme was the domain decomposition, which intended to improve the numerical solution technique for the partial differential equations at hand; in the present study, those that governed either the fluid flow, or the aeroacoustic wave propagation, or the sensitivity analysis for a gradient-based optimization. The role of the domain decomposition extended beyond the original impetus of discretizing geometrical complex regions or writing modular software for distributed-hardware computers. It induced function-space decompositions and operator decompositions that offered the valuable property of near independence of operator evaluation tasks. The objectives have gravitated about the extensions and implementations of either the previously developed or concurrently being developed methodologies: (1) aerodynamic sensitivity analysis with domain decomposition (SADD); (2) computational aeroacoustics of cavities; and (3) dynamic, multibody computational fluid dynamics using unstructured meshes.
Parallel Domain Decomposition Preconditioning for Computational Fluid Dynamics
NASA Technical Reports Server (NTRS)
Barth, Timothy J.; Chan, Tony F.; Tang, Wei-Pai; Kutler, Paul (Technical Monitor)
1998-01-01
This viewgraph presentation gives an overview of the parallel domain decomposition preconditioning for computational fluid dynamics. Details are given on some difficult fluid flow problems, stabilized spatial discretizations, and Newton's method for solving the discretized flow equations. Schur complement domain decomposition is described through basic formulation, simplifying strategies (including iterative subdomain and Schur complement solves, matrix element dropping, localized Schur complement computation, and supersparse computations), and performance evaluation.
A Load Balanced Domain Decomposition Method Using Wavelet Analysis
Jameson, L; Johnson, J; Hesthaven, J
2001-05-31
Wavelet Analysis provides an orthogonal basis set which is localized in both the physical space and the Fourier transform space. We present here a domain decomposition method that uses wavelet analysis to maintain roughly uniform error throughout the computation domain while keeping the computational work balanced in a parallel computing environment.
A New Domain Decomposition Approach for the Gust Response Problem
NASA Technical Reports Server (NTRS)
Scott, James R.; Atassi, Hafiz M.; Susan-Resiga, Romeo F.
2002-01-01
A domain decomposition method is developed for solving the aerodynamic/aeroacoustic problem of an airfoil in a vortical gust. The computational domain is divided into inner and outer regions wherein the governing equations are cast in different forms suitable for accurate computations in each region. Boundary conditions which ensure continuity of pressure and velocity are imposed along the interface separating the two regions. A numerical study is presented for reduced frequencies ranging from 0.1 to 3.0. It is seen that the domain decomposition approach in providing robust and grid independent solutions.
A Domain Decomposition Parallelization of the Fast Marching Method
NASA Technical Reports Server (NTRS)
Herrmann, M.
2003-01-01
In this paper, the first domain decomposition parallelization of the Fast Marching Method for level sets has been presented. Parallel speedup has been demonstrated in both the optimal and non-optimal domain decomposition case. The parallel performance of the proposed method is strongly dependent on load balancing separately the number of nodes on each side of the interface. A load imbalance of nodes on either side of the domain leads to an increase in communication and rollback operations. Furthermore, the amount of inter-domain communication can be reduced by aligning the inter-domain boundaries with the interface normal vectors. In the case of optimal load balancing and aligned inter-domain boundaries, the proposed parallel FMM algorithm is highly efficient, reaching efficiency factors of up to 0.98. Future work will focus on the extension of the proposed parallel algorithm to higher order accuracy. Also, to further enhance parallel performance, the coupling of the domain decomposition parallelization to the G(sub 0)-based parallelization will be investigated.
A multilevel preconditioner for domain decomposition boundary systems
Bramble, J.H.; Pasciak, J.E.; Xu, Jinchao.
1991-12-11
In this note, we consider multilevel preconditioning of the reduced boundary systems which arise in non-overlapping domain decomposition methods. It will be shown that the resulting preconditioned systems have condition numbers which be bounded in the case of multilevel spaces on the whole domain and grow at most proportional to the number of levels in the case of multilevel boundary spaces without multilevel extensions into the interior.
Dynamic-domain-decomposition parallel molecular dynamics
NASA Astrophysics Data System (ADS)
Srinivasan, S. G.; Ashok, I.; Jônsson, Hannes; Kalonji, Gretchen; Zahorjan, John
1997-05-01
Parallel molecular dynamics with short-range forces can suffer from load-imbalance problems and attendant performance degradation due to density variations in the simulated system. In this paper, we describe an approach to dynamical load balancing, enabled by the Ādhāra runtime system. The domain assigned to each processor is automatically and dynamically resized so as to evenly distribute the molecular dynamics computations across all the processors. The algorithm was tested on an Intel Paragon parallel computer for two and three-dimensional Lennard-Jones systems containing 99 458 and 256000 atoms, respectively, and using up to 256 processors. In these benchmarks, the overhead for carrying out the load-balancing operations was found to be small and the total computation time was reduced by as much as 50%.
Implementation and performance of a domain decomposition algorithm in Sisal
DeBoni, T.; Feo, J.; Rodrigue, G.; Muller, J.
1993-09-23
Sisal is a general-purpose functional language that hides the complexity of parallel processing, expedites parallel program development, and guarantees determinacy. Parallelism and management of concurrent tasks are realized automatically by the compiler and runtime system. Spatial domain decomposition is a widely-used method that focuses computational resources on the most active, or important, areas of a domain. Many complex programming issues are introduced in paralleling this method including: dynamic spatial refinement, dynamic grid partitioning and fusion, task distribution, data distribution, and load balancing. In this paper, we describe a spatial domain decomposition algorithm programmed in Sisal. We explain the compilation process, and present the execution performance of the resultant code on two different multiprocessor systems: a multiprocessor vector supercomputer, and cache-coherent scalar multiprocessor.
3D versus 2D domain wall interaction in ideal and rough nanowires
NASA Astrophysics Data System (ADS)
Pivano, A.; Dolocan, Voicu O.
2015-11-01
The interaction between transverse magnetic domain walls (TDWs) in planar (2D) and cylindrical (3D) nanowires is examined using micromagnetic simulations. We show that in perfect and surface deformed wires the free TDWs behave differently, as the 3D TDWs combine into metastable states with average lifetimes of 300 ns depending on roughness, while the 2D TDWs do not due to 2D shape anisotropy. When the 2D and 3D TDWs are pinned at artificial constrictions, they behave similarly as they interact mainly through the dipolar field. This magnetostatic interaction is well described by the point charge model with multipole expansion. In surface deformed wires with artificial constrictions, the interaction becomes more complex as the depinning field decreases and dynamical pinning can lead to local resonances. This can strongly influence the control of TDWs in DW-based devices.
Separation Surfaces in the Spectral TV Domain for Texture Decomposition
NASA Astrophysics Data System (ADS)
Horesh, Dikla; Gilboa, Guy
2016-09-01
In this paper we introduce a novel notion of separation surfaces for image decomposition. A surface is embedded in the spectral total-variation (TV) three dimensional domain and encodes a spatially-varying separation scale. The method allows good separation of textures with gradually varying pattern-size, pattern-contrast or illumination. The recently proposed total variation spectral framework is used to decompose the image into a continuum of textural scales. A desired texture, within a scale range, is found by fitting a surface to the local maximal responses in the spectral domain. A band above and below the surface, referred to as the \\textit{Texture Stratum}, defines for each pixel the adaptive scale-range of the texture. Based on the decomposition an application is proposed which can attenuate or enhance textures in the image in a very natural and visually convincing manner.
Domain decomposition methods for solving an image problem
Tsui, W.K.; Tong, C.S.
1994-12-31
The domain decomposition method is a technique to break up a problem so that ensuing sub-problems can be solved on a parallel computer. In order to improve the convergence rate of the capacitance systems, pre-conditioned conjugate gradient methods are commonly used. In the last decade, most of the efficient preconditioners are based on elliptic partial differential equations which are particularly useful for solving elliptic partial differential equations. In this paper, the authors apply the so called covering preconditioner, which is based on the information of the operator under investigation. Therefore, it is good for various kinds of applications, specifically, they shall apply the preconditioned domain decomposition method for solving an image restoration problem. The image restoration problem is to extract an original image which has been degraded by a known convolution process and additive Gaussian noise.
Adaptive domain decomposition methods for advection-diffusion problems
Carlenzoli, C.; Quarteroni, A.
1995-12-31
Domain decomposition methods can perform poorly on advection-diffusion equations if diffusion is dominated by advection. Indeed, the hyperpolic part of the equations could affect the behavior of iterative schemes among subdomains slowing down dramatically their rate of convergence. Taking into account the direction of the characteristic lines we introduce suitable adaptive algorithms which are stable with respect to the magnitude of the convective field in the equations and very effective on bear boundary value problems.
A domain decomposition scheme for Eulerian shock physics codes
Bell, R.L.; Hertel, E.S. Jr.
1994-08-01
A new algorithm which allows for complex domain decomposition in Eulerian codes was developed at Sandia National Laboratories. This new feature allows a user to customize the zoning for each portion of a calculation and to refine volumes of the computational space of particular interest This option is available in one, two, and three dimensions. The new technique will be described in detail and several examples of the effectiveness of this technique will also be discussed.
Iterative image-domain decomposition for dual-energy CT
Niu, Tianye; Dong, Xue; Petrongolo, Michael; Zhu, Lei
2014-04-15
Purpose: Dual energy CT (DECT) imaging plays an important role in advanced imaging applications due to its capability of material decomposition. Direct decomposition via matrix inversion suffers from significant degradation of image signal-to-noise ratios, which reduces clinical values of DECT. Existing denoising algorithms achieve suboptimal performance since they suppress image noise either before or after the decomposition and do not fully explore the noise statistical properties of the decomposition process. In this work, the authors propose an iterative image-domain decomposition method for noise suppression in DECT, using the full variance-covariance matrix of the decomposed images. Methods: The proposed algorithm is formulated in the form of least-square estimation with smoothness regularization. Based on the design principles of a best linear unbiased estimator, the authors include the inverse of the estimated variance-covariance matrix of the decomposed images as the penalty weight in the least-square term. The regularization term enforces the image smoothness by calculating the square sum of neighboring pixel value differences. To retain the boundary sharpness of the decomposed images, the authors detect the edges in the CT images before decomposition. These edge pixels have small weights in the calculation of the regularization term. Distinct from the existing denoising algorithms applied on the images before or after decomposition, the method has an iterative process for noise suppression, with decomposition performed in each iteration. The authors implement the proposed algorithm using a standard conjugate gradient algorithm. The method performance is evaluated using an evaluation phantom (Catphan©600) and an anthropomorphic head phantom. The results are compared with those generated using direct matrix inversion with no noise suppression, a denoising method applied on the decomposed images, and an existing algorithm with similar formulation as the
NASA Technical Reports Server (NTRS)
Barth, Timothy J.; Chan, Tony F.; Tang, Wei-Pai
1998-01-01
This paper considers an algebraic preconditioning algorithm for hyperbolic-elliptic fluid flow problems. The algorithm is based on a parallel non-overlapping Schur complement domain-decomposition technique for triangulated domains. In the Schur complement technique, the triangulation is first partitioned into a number of non-overlapping subdomains and interfaces. This suggests a reordering of triangulation vertices which separates subdomain and interface solution unknowns. The reordering induces a natural 2 x 2 block partitioning of the discretization matrix. Exact LU factorization of this block system yields a Schur complement matrix which couples subdomains and the interface together. The remaining sections of this paper present a family of approximate techniques for both constructing and applying the Schur complement as a domain-decomposition preconditioner. The approximate Schur complement serves as an algebraic coarse space operator, thus avoiding the known difficulties associated with the direct formation of a coarse space discretization. In developing Schur complement approximations, particular attention has been given to improving sequential and parallel efficiency of implementations without significantly degrading the quality of the preconditioner. A computer code based on these developments has been tested on the IBM SP2 using MPI message passing protocol. A number of 2-D calculations are presented for both scalar advection-diffusion equations as well as the Euler equations governing compressible fluid flow to demonstrate performance of the preconditioning algorithm.
Applications of domain decomposition methods to turbomachinery flows
NASA Astrophysics Data System (ADS)
Rai, M. M.
Domain decomposition techniques can be used to great advantage by computational fluid dynamicists in computing flows about complex geometries and adapting the grid to the solution. These techniques are particularly useful in computing flows about several bodies that are in relative motion such as rotor/stator configurations in turbomachinery or helicopter rotor/fuselage configurations. This paper discusses some of the basic ideas involved in transferring information between subdomains in a multidomain calculation and presents results for a simple rotor/stator configuration.
Domain decomposition approach to flexible multibody dynamics simulation
NASA Astrophysics Data System (ADS)
Kwak, JunYoung; Chun, TaeYoung; Shin, SangJoon; Bauchau, Olivier A.
2014-01-01
Finite element based formulations for flexible multibody systems are becoming increasingly popular and as the complexity of the configurations to be treated increases, so does the computational cost. It seems natural to investigate the applicability of parallel processing to this type of problems; domain decomposition techniques have been used extensively for this purpose. In this approach, the computational domain is divided into non-overlapping sub-domains, and the continuity of the displacement field across sub-domain boundaries is enforced via the Lagrange multiplier technique. In the finite element literature, this approach is presented as a mathematical algorithm that enables parallel processing. In this paper, the divided system is viewed as a flexible multibody system, and the sub-domains are connected by kinematic constraints. Consequently, all the techniques applicable to the enforcement of constraints in multibody systems become applicable to the present problem. In particular, it is shown that a combination of the localized Lagrange multiplier technique with the augmented Lagrange formulation leads to interesting solution strategies. The proposed algorithm is compared with the well-known FETI approach with regards to convergence and efficiency characteristics. The present algorithm is relatively simple and leads to improved convergence and efficiency characteristics. Finally, implementation on a parallel computer was conducted for the proposed approach.
Numerical computation of 2D Sommerfeld integrals - Decomposition of the angular integral
NASA Astrophysics Data System (ADS)
Dvorak, Steven L.; Kuester, Edward F.
1992-02-01
The computational efficiency of the 2D Sommerfeld integrals is shown to undergo improvement through the discovery of novel ways to compute the inner angular integral in polar representations. It is shown that the angular integral can be decomposed into a finite number of incomplete Lipschitz-Hankel integrals; these can in turn be calculated through a series of expansions, so that the angular integral can be computed by summing a series rather than applying a standard numerical integration algorithm. The technique is most efficient and accurate when piecewise-sinusoidal basis functions are employed to analyze a printed strip-dipole antenna in a layered medium.
Segmented Domain Decomposition Multigrid For 3-D Turbomachinery Flows
NASA Technical Reports Server (NTRS)
Celestina, M. L.; Adamczyk, J. J.; Rubin, S. G.
2001-01-01
A Segmented Domain Decomposition Multigrid (SDDMG) procedure was developed for three-dimensional viscous flow problems as they apply to turbomachinery flows. The procedure divides the computational domain into a coarse mesh comprised of uniformly spaced cells. To resolve smaller length scales such as the viscous layer near a surface, segments of the coarse mesh are subdivided into a finer mesh. This is repeated until adequate resolution of the smallest relevant length scale is obtained. Multigrid is used to communicate information between the different grid levels. To test the procedure, simulation results will be presented for a compressor and turbine cascade. These simulations are intended to show the ability of the present method to generate grid independent solutions. Comparisons with data will also be presented. These comparisons will further demonstrate the usefulness of the present work for they allow an estimate of the accuracy of the flow modeling equations independent of error attributed to numerical discretization.
NASA Astrophysics Data System (ADS)
Tucciarelli, T.
2012-12-01
A new methodology for the solution of irrotational 2D flow problems in domains with strongly unstructured meshes is presented. A fractional time step procedure is applied to the original governing equations, solving consecutively a convective prediction system and a diffusive corrective system. The non linear components of the problem are concentrated in the prediction step, while the correction step leads to the solution of a linear system, of the order of the number of computational cells. A MArching in Space and Time (MAST) approach is applied for the solution of the convective prediction step. The major advantages of the model, as well as its ability to maintain the solution monotonicity even in strongly irregular meshes, are briefly described. The algorithm is applied to the solution of diffusive shallow water equations in a simple domain.
Feng, Xiaobing
1996-12-31
A non-overlapping domain decomposition iterative method is proposed and analyzed for mixed finite element methods for a sequence of noncoercive elliptic systems with radiation boundary conditions. These differential systems describe the motion of a nearly elastic solid in the frequency domain. The convergence of the iterative procedure is demonstrated and the rate of convergence is derived for the case when the domain is decomposed into subdomains in which each subdomain consists of an individual element associated with the mixed finite elements. The hybridization of mixed finite element methods plays a important role in the construction of the discrete procedure.
Simplified approaches to some nonoverlapping domain decomposition methods
Xu, Jinchao
1996-12-31
An attempt will be made in this talk to present various domain decomposition methods in a way that is intuitively clear and technically coherent and concise. The basic framework used for analysis is the {open_quotes}parallel subspace correction{close_quotes} or {open_quotes}additive Schwarz{close_quotes} method, and other simple technical tools include {open_quotes}local-global{close_quotes} and {open_quotes}global-local{close_quotes} techniques, the formal one is for constructing subspace preconditioner based on a preconditioner on the whole space whereas the later one for constructing preconditioner on the whole space based on a subspace preconditioner. The domain decomposition methods discussed in this talk fall into two major categories: one, based on local Dirichlet problems, is related to the {open_quotes}substructuring method{close_quotes} and the other, based on local Neumann problems, is related to the {open_quotes}Neumann-Neumann method{close_quotes} and {open_quotes}balancing method{close_quotes}. All these methods will be presented in a systematic and coherent manner and the analysis for both two and three dimensional cases are carried out simultaneously. In particular, some intimate relationship between these algorithms are observed and some new variants of the algorithms are obtained.
Recursive anisotropic 2-D Gaussian filtering based on a triple-axis decomposition.
Lam, Stanley Yiu Man; Shi, Bertram E
2007-07-01
We describe a recursive algorithm for anisotropic 2-D Gaussian filtering, based on separating the filter into the cascade of three, rather two, 1-D filters. The filters operate along axes obtained by integer horizontal and/or vertical pixel shifts. This eliminates interpolation, which removes spatial inhomogeneity in the filter, and produces more elliptically shaped kernels. It also results in a more regular filter structure, which facilitates implementation in DSP chips. Finally, it improves matching between filters with the same eccentricity and width, but different orientations. Our analysis and experiments indicate that the computational complexity is similar to an algorithm that operates along two axes (<11 ms for a 512 x 512 image using a 3.2-GHz Pentium 4 PC). On the other hand, given a limited set of basis filter axes, there is an orientation dependent lower bound on the achievable aspect ratios. PMID:17605390
2D Seismic Imaging of Elastic Parameters by Frequency Domain Full Waveform Inversion
NASA Astrophysics Data System (ADS)
Brossier, R.; Virieux, J.; Operto, S.
2008-12-01
Thanks to recent advances in parallel computing, full waveform inversion is today a tractable seismic imaging method to reconstruct physical parameters of the earth interior at different scales ranging from the near- surface to the deep crust. We present a massively parallel 2D frequency-domain full-waveform algorithm for imaging visco-elastic media from multi-component seismic data. The forward problem (i.e. the resolution of the frequency-domain 2D PSV elastodynamics equations) is based on low-order Discontinuous Galerkin (DG) method (P0 and/or P1 interpolations). Thanks to triangular unstructured meshes, the DG method allows accurate modeling of both body waves and surface waves in case of complex topography for a discretization of 10 to 15 cells per shear wavelength. The frequency-domain DG system is solved efficiently for multiple sources with the parallel direct solver MUMPS. The local inversion procedure (i.e. minimization of residuals between observed and computed data) is based on the adjoint-state method which allows to efficiently compute the gradient of the objective function. Applying the inversion hierarchically from the low frequencies to the higher ones defines a multiresolution imaging strategy which helps convergence towards the global minimum. In place of expensive Newton algorithm, the combined use of the diagonal terms of the approximate Hessian matrix and optimization algorithms based on quasi-Newton methods (Conjugate Gradient, LBFGS, ...) allows to improve the convergence of the iterative inversion. The distribution of forward problem solutions over processors driven by a mesh partitioning performed by METIS allows to apply most of the inversion in parallel. We shall present the main features of the parallel modeling/inversion algorithm, assess its scalability and illustrate its performances with realistic synthetic case studies.
Parallel domain decomposition methods in fluid models with Monte Carlo transport
Alme, H.J.; Rodrigues, G.H.; Zimmerman, G.B.
1996-12-01
To examine the domain decomposition code coupled Monte Carlo-finite element calculation, it is important to use a domain decomposition that is suitable for the individual models. We have developed a code that simulates a Monte Carlo calculation ( ) on a massively parallel processor. This code is used to examine the load balancing behavior of three domain decomposition ( ) for a Monte Carlo calculation. Results are presented.
Domain decomposition methods for the parallel computation of reacting flows
NASA Technical Reports Server (NTRS)
Keyes, David E.
1988-01-01
Domain decomposition is a natural route to parallel computing for partial differential equation solvers. Subdomains of which the original domain of definition is comprised are assigned to independent processors at the price of periodic coordination between processors to compute global parameters and maintain the requisite degree of continuity of the solution at the subdomain interfaces. In the domain-decomposed solution of steady multidimensional systems of PDEs by finite difference methods using a pseudo-transient version of Newton iteration, the only portion of the computation which generally stands in the way of efficient parallelization is the solution of the large, sparse linear systems arising at each Newton step. For some Jacobian matrices drawn from an actual two-dimensional reacting flow problem, comparisons are made between relaxation-based linear solvers and also preconditioned iterative methods of Conjugate Gradient and Chebyshev type, focusing attention on both iteration count and global inner product count. The generalized minimum residual method with block-ILU preconditioning is judged the best serial method among those considered, and parallel numerical experiments on the Encore Multimax demonstrate for it approximately 10-fold speedup on 16 processors.
Traffic simulations on parallel computers using domain decomposition techniques
Hanebutte, U.R.; Tentner, A.M.
1995-12-31
Large scale simulations of Intelligent Transportation Systems (ITS) can only be achieved by using the computing resources offered by parallel computing architectures. Domain decomposition techniques are proposed which allow the performance of traffic simulations with the standard simulation package TRAF-NETSIM on a 128 nodes IBM SPx parallel supercomputer as well as on a cluster of SUN workstations. Whilst this particular parallel implementation is based on NETSIM, a microscopic traffic simulation model, the presented strategy is applicable to a broad class of traffic simulations. An outer iteration loop must be introduced in order to converge to a global solution. A performance study that utilizes a scalable test network that consist of square-grids is presented, which addresses the performance penalty introduced by the additional iteration loop.
Adaptive multigrid domain decomposition solutions for viscous interacting flows
NASA Technical Reports Server (NTRS)
Rubin, Stanley G.; Srinivasan, Kumar
1992-01-01
Several viscous incompressible flows with strong pressure interaction and/or axial flow reversal are considered with an adaptive multigrid domain decomposition procedure. Specific examples include the triple deck structure surrounding the trailing edge of a flat plate, the flow recirculation in a trough geometry, and the flow in a rearward facing step channel. For the latter case, there are multiple recirculation zones, of different character, for laminar and turbulent flow conditions. A pressure-based form of flux-vector splitting is applied to the Navier-Stokes equations, which are represented by an implicit lowest-order reduced Navier-Stokes (RNS) system and a purely diffusive, higher-order, deferred-corrector. A trapezoidal or box-like form of discretization insures that all mass conservation properties are satisfied at interfacial and outflow boundaries, even for this primitive-variable, non-staggered grid computation.
Efficient variants of the vertex space domain decomposition algorithm
Chan, T.F.; Shao, J.P. . Dept. of Mathematics); Mathew, T.P. . Dept. of Mathematics)
1994-11-01
Several variants of the vertex space algorithm of Smith for two-dimensional elliptic problems are described. The vertex space algorithm is a domain decomposition method based on nonoverlapping subregions, in which the reduced Schur complement system on the interface is solved using a generalized block Jacobi-type preconditioner, with the blocks corresponding to the vertex space, edges, and a coarse grid. Two kinds of approximations are considered for the edge and vertex space subblocks, one based on Fourier approximation, and another based on an algebraic probing technique in which sparse approximations to these subblocks are computed. The motivation is to improve the efficiency of the algorithm without sacrificing the optimal convergence rate. Numerical and theoretical results on the performance of these algorithms, including variants of an algorithm of Bramble, Pasciak, and Schatz are presented.
Analysis of generalized Schwarz alternating procedure for domain decomposition
Engquist, B.; Zhao, Hongkai
1996-12-31
The Schwartz alternating method(SAM) is the theoretical basis for domain decomposition which itself is a powerful tool both for parallel computation and for computing in complicated domains. The convergence rate of the classical SAM is very sensitive to the overlapping size between each subdomain, which is not desirable for most applications. We propose a generalized SAM procedure which is an extension of the modified SAM proposed by P.-L. Lions. Instead of using only Dirichlet data at the artificial boundary between subdomains, we take a convex combination of u and {partial_derivative}u/{partial_derivative}n, i.e. {partial_derivative}u/{partial_derivative}n + {Lambda}u, where {Lambda} is some {open_quotes}positive{close_quotes} operator. Convergence of the modified SAM without overlapping in a quite general setting has been proven by P.-L.Lions using delicate energy estimates. The important questions remain for the generalized SAM. (1) What is the most essential mechanism for convergence without overlapping? (2) Given the partial differential equation, what is the best choice for the positive operator {Lambda}? (3) In the overlapping case, is the generalized SAM superior to the classical SAM? (4) What is the convergence rate and what does it depend on? (5) Numerically can we obtain an easy to implement operator {Lambda} such that the convergence is independent of the mesh size. To analyze the convergence of the generalized SAM we focus, for simplicity, on the Poisson equation for two typical geometry in two subdomain case.
Domain decomposition methods for the parallel computation of reacting flows
NASA Astrophysics Data System (ADS)
Keyes, David E.
1989-05-01
Domain decomposition is a natural route to parallel computing for partial differential equation solvers. In this procedure, subdomains of which the original domain of definition is comprised are assigned to independent processors at the price of periodic coordination between processors to compute global parameters and maintain the requisite degree of continuity of the solution at the subdomain interfaces. In the domain-decomposed solution of steady multidimensional systems of PDEs by finite difference methods using a pseudo-transient version of Newton iteration, the only portion of the computation which generally stands in the way of efficient parallelization is the solution of the large, sparse linear systems arising at each Newton step. For some Jacobian matrices drawn from an actual two-dimensional reacting flow problem, we make comparisons between relaxation-based linear solvers and also preconditioned iterative methods of Conjugate Gradient and Chebyshev type, focusing attention on both iteration count and global inner product count. The generalized minimum residual method with block-ILU preconditioning is judged the best serial method among those considered, and parallel numerical experiments on the Encore Multimax demostrate for it approximately 10-fold speedup on 16 processsors. The three special features of reacting flow models in relation to these linear systems are: the possibly large number of degrees of freedom per gridpoint, the dominance of dense intra-point source-term coupling over inter-point convective-diffusive coupling throughout significant portions of the flow-field and strong nonlinearities which restrict the time step to small values (independent of linear algebraic considerations) throughout significant portions of the iteration history. Though these features are exploited to advantage herein, many aspects of the paper are applicable to the modeling of general convective-diffusive systems.
Composite structured mesh generation with automatic domain decomposition in complex geometries
Technology Transfer Automated Retrieval System (TEKTRAN)
This paper presents a novel automatic domain decomposition method to generate quality composite structured meshes in complex domains with arbitrary shapes, in which quality structured mesh generation still remains a challenge. The proposed decomposition algorithm is based on the analysis of an initi...
Domain Decomposition By the Advancing-Partition Method for Parallel Unstructured Grid Generation
NASA Technical Reports Server (NTRS)
Pirzadeh, Shahyar Z.; Zagaris, George
2009-01-01
A new method of domain decomposition has been developed for generating unstructured grids in subdomains either sequentially or using multiple computers in parallel. Domain decomposition is a crucial and challenging step for parallel grid generation. Prior methods are generally based on auxiliary, complex, and computationally intensive operations for defining partition interfaces and usually produce grids of lower quality than those generated in single domains. The new technique, referred to as "Advancing Partition," is based on the Advancing-Front method, which partitions a domain as part of the volume mesh generation in a consistent and "natural" way. The benefits of this approach are: 1) the process of domain decomposition is highly automated, 2) partitioning of domain does not compromise the quality of the generated grids, and 3) the computational overhead for domain decomposition is minimal. The new method has been implemented in NASA's unstructured grid generation code VGRID.
Parallel Finite Element Domain Decomposition for Structural/Acoustic Analysis
NASA Technical Reports Server (NTRS)
Nguyen, Duc T.; Tungkahotara, Siroj; Watson, Willie R.; Rajan, Subramaniam D.
2005-01-01
A domain decomposition (DD) formulation for solving sparse linear systems of equations resulting from finite element analysis is presented. The formulation incorporates mixed direct and iterative equation solving strategics and other novel algorithmic ideas that are optimized to take advantage of sparsity and exploit modern computer architecture, such as memory and parallel computing. The most time consuming part of the formulation is identified and the critical roles of direct sparse and iterative solvers within the framework of the formulation are discussed. Experiments on several computer platforms using several complex test matrices are conducted using software based on the formulation. Small-scale structural examples are used to validate thc steps in the formulation and large-scale (l,000,000+ unknowns) duct acoustic examples are used to evaluate the ORIGIN 2000 processors, and a duster of 6 PCs (running under the Windows environment). Statistics show that the formulation is efficient in both sequential and parallel computing environmental and that the formulation is significantly faster and consumes less memory than that based on one of the best available commercialized parallel sparse solvers.
Investigation on Accelerating Dust Storm Simulation via Domain Decomposition Methods
NASA Astrophysics Data System (ADS)
Yu, M.; Gui, Z.; Yang, C. P.; Xia, J.; Chen, S.
2014-12-01
Dust storm simulation is a data and computing intensive process, which requires high efficiency and adequate computing resources. To speed up the process, high performance computing is widely adopted. By partitioning a large study area into small subdomains according to their geographic location and executing them on different computing nodes in a parallel fashion, the computing performance can be significantly improved. However, it is still a question worthy of consideration that how to allocate these subdomain processes into computing nodes without introducing imbalanced task loads and unnecessary communications among computing nodes. Here we propose a domain decomposition and allocation framework that can carefully leverage the computing cost and communication cost for each computing node to minimize total execution time and reduce overall communication cost for the entire system. The framework is tested in the NMM (Nonhydrostatic Mesoscale Model)-dust model, where a 72-hour processes of the dust load are simulated. Performance result using the proposed scheduling method is compared with the one using default scheduling methods of MPI. Results demonstrate that the system improves the performance of simulation by 20% up to 80%.
2D time-domain finite-difference modeling for viscoelastic seismic wave propagation
NASA Astrophysics Data System (ADS)
Fan, Na; Zhao, Lian-Feng; Xie, Xiao-Bi; Ge, Zengxi; Yao, Zhen-Xing
2016-07-01
Real Earth media are not perfectly elastic. Instead, they attenuate propagating mechanical waves. This anelastic phenomenon in wave propagation can be modeled by a viscoelastic mechanical model consisting of several standard linear solids. Using this viscoelastic model, we approximate a constant Q over a frequency band of interest. We use a four-element viscoelastic model with a tradeoff between accuracy and computational costs to incorporate Q into 2D time-domain first-order velocity-stress wave equations. To improve the computational efficiency, we limit the Q in the model to a list of discrete values between 2 and 1000. The related stress and strain relaxation times that characterize the viscoelastic model are pre-calculated and stored in a database for use by the finite-difference calculation. A viscoelastic finite-difference scheme that is second-order in time and fourth-order in space is developed based on the MacCormack algorithm. The new method is validated by comparing the numerical result with analytical solutions that are calculated using the generalized reflection/transmission coefficient method. The synthetic seismograms exhibit greater than 95 per cent consistency in a two-layer viscoelastic model. The dispersion generated from the simulation is consistent with the Kolsky-Futterman dispersion relationship.
NASA Astrophysics Data System (ADS)
Rung-Arunwan, Tawat; Siripunvaraporn, Weerachai
2010-11-01
We use 2-D magnetotelluric (MT) problems as a feasibility study to demonstrate that 3-D MT problems can be solved with a direct solver, even on a standard single processor PC. The scheme used is a hierarchical domain decomposition (HDD) method in which a global computational domain is uniformly split into many smaller non-overlapping subdomains. To make it more efficient, two modifications are made to the standard HDD method. Instead of three levels as in the standard HDD method, we classify the unknowns into four classes: the interiors, the horizontal interfaces, the vertical interfaces and the intersections. Four sets of smaller systems of equations are successively solved with a direct method (an LU factorization). The separation significantly reduces the large memory requirements of a direct solver. It also reduces the CPU time to almost half that of the standard HDD method although it is still slower than the conventional finite difference (FD) method. To further enhance the speed of the code, a red-black ordering is applied to solve the horizontal and vertical interface reduced systems. Numerical experiments on a 2-D MT problem of a given size running on a single processor machine shows that CPU time and memory used are almost constant for any resistivity models, frequencies and modes. This is a clear advantage of our algorithm and is of particular importance if the method is applied to 3-D problems. We show that our new method results in reductions in both memory usage and CPU time for large enough domains when compared to the standard FD and HDD schemes. In addition, we also introduce a `memory minimization map', a graphical tool we can use instead of trial-and-error to pre-select the optimal size of subdomains, which yield the best performance in both CPU time and memory even running on a serial machine.
NASA Astrophysics Data System (ADS)
Liu, Hongyi; Zhou, Jun; Chen, Yijian
2015-03-01
To break through 1-D IC layout limitations, we develop computationally efficient 2-D layout decomposition and stitching techniques which combine the optical and self-aligned multiple patterning (SAMP) processes. A polynomial time algorithm is developed to decompose the target layout into two components, each containing one or multiple sets of unidirectional features that can be formed by a SAMP+cut/block process. With no need of connecting vias, the final 2-D features are formed by directly stitching two components together. This novel patterning scheme is considered as a hybrid approach as the SAMP processes offer the capability of density scaling while the stitching process creates 2-D design freedom as well as the multiple-CD/pitch capability. Its technical advantages include significant reduction of via steps and avoiding the interdigitating types of multiple patterning (for density multiplication) to improve the processing yield. The developed decomposition and synthesis algorithms are tested using 2-D layouts from NCSU open cell library. Statistical and computational characteristics of these public layout data are investigated and discussed.
Wu, Zong-Sian; Liu, Che Fu; Fu, Brian; Chou, Ruey-Hwang; Yu, Chin
2016-09-01
The extracellular portion of the human fibroblast growth factor receptor2 D2 domain (FGFR2 D2) interacts with human fibroblast growth factor 1 (hFGF1) to activate a downstream signaling cascade that ultimately affects mitosis and differentiation. Suramin is an antiparasiticdrug and a potent inhibitor of FGF-induced angiogenesis. Suramin has been shown to bind to hFGF1, and might block the interaction between hFGF1 and FGFR2 D2. Here, we titrated hFGF1 with FGFR2 D2 and suramin to elucidate their interactions using the detection of NMR. The docking results of both hFGF1-FGFR2 D2 domain and hFGF1-suramin complex were superimposed. The results indicate that suramin blocks the interaction between hFGF1 and FGFR2 D2. We used the PyMOL software to show the hydrophobic interaction of hFGF1-suramin. In addition, we used a Water-soluble Tetrazolium salts assay (WST1) to assess hFGF1 bioactivity. The results will be useful for the development of new antimitogenic activity drugs. PMID:27387234
Algebraic Nonoverlapping Domain Decomposition Methods for Stabilized FEM and FV Discretizations
NASA Technical Reports Server (NTRS)
Barth, Timothy J.; Bailey, David (Technical Monitor)
1998-01-01
We consider preconditioning methods for convection dominated fluid flow problems based on a nonoverlapping Schur complement domain decomposition procedure for arbitrary triangulated domains. The triangulation is first partitioned into a number of subdomains and interfaces which induce a natural 2 x 2 partitioning of the p.d.e. discretization matrix. We view the Schur complement induced by this partitioning as an algebraically derived coarse space approximation. This avoids the known difficulties associated with the direct formation of an effective coarse discretization for advection dominated equations. By considering various approximations of the block factorization of the 2 x 2 system, we have developed a family of robust preconditioning techniques. A computer code based on these ideas has been developed and tested on the IBM SP2 using MPI message passing protocol. A number of 2-D CFD calculations will be presented for both scalar advection-diffusion equations and the Euler equations discretized using stabilized finite element and finite volume methods. These results show very good scalability of the preconditioner for various discretizations as the number of processors is increased while the number of degrees of freedom per processor is fixed.
NASA Astrophysics Data System (ADS)
Beilina, Larisa
2016-08-01
We present domain decomposition finite element/finite difference method for the solution of hyperbolic equation. The domain decomposition is performed such that finite elements and finite differences are used in different subdomains of the computational domain: finite difference method is used on the structured part of the computational domain and finite elements on the unstructured part of the domain. Explicit discretizations for both methods are constructed such that the finite element and the finite difference schemes coincide on the common structured overlapping layer between computational subdomains. Then the resulting approach can be considered as a pure finite element scheme which avoids instabilities at the interfaces. We derive an energy estimate for the underlying hyperbolic equation with absorbing boundary conditions and illustrate efficiency of the domain decomposition method on the reconstruction of the conductivity function in three dimensions.
Preconditioned domain decomposition scheme for three-dimensional aerodynamic sensitivity analysis
NASA Technical Reports Server (NTRS)
Eleshaky, Mohammed E.; Baysal, Oktay
1993-01-01
A preconditioned domain decomposition scheme is introduced for the solution of the 3D aerodynamic sensitivity equation. This scheme uses the iterative GMRES procedure to solve the effective sensitivity equation of the boundary-interface cells in the sensitivity analysis domain-decomposition scheme. Excluding the dense matrices and the effect of cross terms between boundary-interfaces is found to produce an efficient preconditioning matrix.
Domain decomposition solvers for PDEs : some basics, practical tools, and new developments.
Dohrmann, Clark R.
2010-11-01
The first part of this talk provides a basic introduction to the building blocks of domain decomposition solvers. Specific details are given for both the classical overlapping Schwarz (OS) algorithm and a recent iterative substructuring (IS) approach called balancing domain decomposition by constraints (BDDC). A more recent hybrid OS-IS approach is also described. The success of domain decomposition solvers depends critically on the coarse space. Similarities and differences between the coarse spaces for OS and BDDC approaches are discussed, along with how they can be obtained from discrete harmonic extensions. Connections are also made between coarse spaces and multiscale modeling approaches from computational mechanics. As a specific example, details are provided on constructing coarse spaces for incompressible fluid problems. The next part of the talk deals with a variety of implementation details for domain decomposition solvers. These include mesh partitioning options, local and global solver options, reducing the coarse space dimension, dealing with constraint equations, residual weighting to accelerate the convergence of OS methods, and recycling of Krylov spaces to efficiently solve problems with multiple right hand sides. Some potential bottlenecks and remedies for domain decomposition solvers are also discussed. The final part of the talk concerns some recent theoretical advances, new algorithms, and open questions in the analysis of domain decomposition solvers. The focus will be primarily on the work of the speaker and his colleagues on elasticity, fluid mechanics, problems in H(curl), and the analysis of subdomains with irregular boundaries.
NASA Astrophysics Data System (ADS)
Korneev, V. G.
2012-09-01
BPS is a well known an efficient and rather general domain decomposition Dirichlet-Dirichlet type preconditioner, suggested in the famous series of papers Bramble, Pasciak and Schatz (1986-1989). Since then, it has been serving as the origin for the whole family of domain decomposition Dirichlet-Dirichlet type preconditioners-solvers as for h so hp discretizations of elliptic problems. For its original version, designed for h discretizations, the named authors proved the bound O(1 + log2 H/ h) for the relative condition number under some restricting conditions on the domain decomposition and finite element discretization. Here H/ h is the maximal relation of the characteristic size H of a decomposition subdomain to the mesh parameter h of its discretization. It was assumed that subdomains are images of the reference unite cube by trilinear mappings. Later similar bounds related to h discretizations were proved for more general domain decompositions, defined by means of coarse tetrahedral meshes. These results, accompanied by the development of some special tools of analysis aimed at such type of decompositions, were summarized in the book of Toselli and Widlund (2005). This paper is also confined to h discretizations. We further expand the range of admissible domain decompositions for constructing BPS preconditioners, in which decomposition subdomains can be convex polyhedrons, satisfying some conditions of shape regularity. We prove the bound for the relative condition number with the same dependence on H/ h as in the bound given above. Along the way to this result, we simplify the proof of the so called abstract bound for the relative condition number of the domain decomposition preconditioner. In the part, related to the analysis of the interface sub-problem preconditioning, our technical tools are generalization of those used by Bramble, Pasciak and Schatz.
NASA Astrophysics Data System (ADS)
Sourbier, Florent; Operto, Stéphane; Virieux, Jean; Amestoy, Patrick; L'Excellent, Jean-Yves
2009-03-01
This is the first paper in a two-part series that describes a massively parallel code that performs 2D frequency-domain full-waveform inversion of wide-aperture seismic data for imaging complex structures. Full-waveform inversion methods, namely quantitative seismic imaging methods based on the resolution of the full wave equation, are computationally expensive. Therefore, designing efficient algorithms which take advantage of parallel computing facilities is critical for the appraisal of these approaches when applied to representative case studies and for further improvements. Full-waveform modelling requires the resolution of a large sparse system of linear equations which is performed with the massively parallel direct solver MUMPS for efficient multiple-shot simulations. Efficiency of the multiple-shot solution phase (forward/backward substitutions) is improved by using the BLAS3 library. The inverse problem relies on a classic local optimization approach implemented with a gradient method. The direct solver returns the multiple-shot wavefield solutions distributed over the processors according to a domain decomposition driven by the distribution of the LU factors. The domain decomposition of the wavefield solutions is used to compute in parallel the gradient of the objective function and the diagonal Hessian, this latter providing a suitable scaling of the gradient. The algorithm allows one to test different strategies for multiscale frequency inversion ranging from successive mono-frequency inversion to simultaneous multifrequency inversion. These different inversion strategies will be illustrated in the following companion paper. The parallel efficiency and the scalability of the code will also be quantified.
Extreme value statistics of 2D Gaussian free field: effect of finite domains
NASA Astrophysics Data System (ADS)
Cao, X.; Rosso, A.; Santachiara, R.
2016-01-01
We study minima statistics of the 2D Gaussian free field (GFF) on circles in the unit disk with Dirichlet boundary condition. Free energy distributions of the associated random energy models are exactly calculated in the high temperature phase, and shown to satisfy the duality property, which enables us to predict the minima distribution by assuming the freezing scenario. Numerical tests are provided. Related questions concerning the GFF on a sphere are also considered.
Two dimensional spectroscopy of Liquids in THz-domain: THz analogue of 2D Raman spectroscopy
NASA Astrophysics Data System (ADS)
Okumura, K.; Tanimura, Y.
1998-03-01
After the initial proposal(Y. Tanimura and S. Mukamel, J. Chem. Phys. 99, 9496 (1993)), the two dimensional Raman spectroscopy in the liquid phase has been received a considerable attention. Both experimental and theoretical activity of this field has been quite high. Since we have two controllable delay times, we can obtain more information than the lower-order experiments such as OKE. The new information includes that on heterogeneous distribution in liquids. Recently, it is found that the coupling between the modes in liquids can be investigated by the technique, both experimentally and theoretically(A. Tokmakoff, M.J. Lang, D.S. Larsen, G.R. Fleming, V. Chernyak, and S. Mukamel, Phys. Rev. Lett. (in press))^,(K. Okumura and Y. Tanimura, Chem. Phys. Lett. 278, 175 (1997)) In this talk, we will emphasize that we can perform the THz analogue of the 2D Raman spectroscopy if the THz short-pulse laser becomes available, which may not be in the far future. Theoretically, we can formulate this novel THz spectroscopy on the same footing as the 2D Raman spectroscopy. We will clarify new aspects of this technique comparing with the 2D Raman spectroscopy--- the reason it worth trying the tough experiment. See
Efficient integration of a realistic two-dimensional cardiac tissue model by domain decomposition.
Quan, W; Evans, S J; Hastings, H M
1998-03-01
The size of realistic cardiac tissue models has been limited by their high computational demands. In particular, the Luo-Rudy phase II membrane model, used to simulate a thin sheet of ventricular tissue with arrays of coupled ventricular myocytes, is usually limited to 100 x 100 arrays. We introduce a new numerical method based on domain decomposition and a priority queue integration scheme which reduces the computational cost by a factor of 3-17. In the standard algorithm all the nodes advance with the same time step delta t, whose size is limited by the time scale of activation. However, at any given time, many regions may be inactive and do not require the same small delta t and consequent extensive computations. Hence, adjusting delta t locally is a key factor in improving computational efficiency, since most of the computing time is spent calculating ionic currents. This paper proposes an efficient adaptive numerical scheme for integrating a two-dimensional (2-D) propagation model, by incorporating local adjustments of delta t. In this method, alternating direction Cooley-Dodge and Rush-Larsen methods were used for numerical integration. Between consecutive integrations over the whole domain using an implicit method, the model was spatially decomposed into many subdomains, and delta t adjusted locally. The Euler method was used for numerical integration in the subdomains. Local boundary values were determined from the boundary mesh elements of the neighboring subdomains using linear interpolation. Because delta t was defined locally, a priority queue was used to store and order next update times for each subdomain. The subdomain with the earliest update time was given the highest priority and advanced first. This new method yielded stable solutions with relative errors less than 1% and reduced computation time by a factor of 3-17 and will allow much larger (e.g., 500 x 500) models based on realistic membrane kinetics and realistic dimensions to simulate
Lattice QCD with Domain Decomposition on Intel Xeon Phi Co-Processors
Heybrock, Simon; Joo, Balint; Kalamkar, Dhiraj D; Smelyanskiy, Mikhail; Vaidyanathan, Karthikeyan; Wettig, Tilo; Dubey, Pradeep
2014-12-01
The gap between the cost of moving data and the cost of computing continues to grow, making it ever harder to design iterative solvers on extreme-scale architectures. This problem can be alleviated by alternative algorithms that reduce the amount of data movement. We investigate this in the context of Lattice Quantum Chromodynamics and implement such an alternative solver algorithm, based on domain decomposition, on Intel Xeon Phi co-processor (KNC) clusters. We demonstrate close-to-linear on-chip scaling to all 60 cores of the KNC. With a mix of single- and half-precision the domain-decomposition method sustains 400-500 Gflop/s per chip. Compared to an optimized KNC implementation of a standard solver [1], our full multi-node domain-decomposition solver strong-scales to more nodes and reduces the time-to-solution by a factor of 5.
Non-conformal domain decomposition methods for time-harmonic Maxwell equations
Shao, Yang; Peng, Zhen; Lim, Kheng Hwee; Lee, Jin-Fa
2012-01-01
We review non-conformal domain decomposition methods (DDMs) and their applications in solving electrically large and multi-scale electromagnetic (EM) radiation and scattering problems. In particular, a finite-element DDM, together with a finite-element tearing and interconnecting (FETI)-like algorithm, incorporating Robin transmission conditions and an edge corner penalty term, are discussed in detail. We address in full the formulations, and subsequently, their applications to problems with significant amounts of repetitions. The non-conformal DDM approach has also been extended into surface integral equation methods. We elucidate a non-conformal integral equation domain decomposition method and a generalized combined field integral equation method for modelling EM wave scattering from non-penetrable and penetrable targets, respectively. Moreover, a plane wave scattering from a composite mockup fighter jet has been simulated using the newly developed multi-solver domain decomposition method. PMID:22870061
Parallel domain decomposition and the solution of nonlinear systems of equations
Gropp, W.D. ); Keyes, D.E. . Dept. of Mechanical Engineering)
1990-01-01
Many linear systems arise as subproblems in the solution of nonlinear equations, either as part of a simple fixed-point of a Newton's method iteration. This paper considers the use of domain decomposition techniques for the solution of these linear problems in the context of solving a multicomponent system of nonlinear equations on two types of parallel processors. One of the computations is drawn from fluid dynamics and includes locally refined grids. Such problems require great computational resources, and domain, decomposition seems to offer a way to efficiently solve these problems on computers with significant parallelism. The domain decomposition approach used is as in Gropp and Keyes, modified to achieve better parallelism and to reduce the computational work. 13 refs., 2 figs., 5 tabs.
Domain decomposition for a mixed finite element method in three dimensions
Cai, Z.; Parashkevov, R.R.; Russell, T.F.; Wilson, J.D.; Ye, X.
2003-01-01
We consider the solution of the discrete linear system resulting from a mixed finite element discretization applied to a second-order elliptic boundary value problem in three dimensions. Based on a decomposition of the velocity space, these equations can be reduced to a discrete elliptic problem by eliminating the pressure through the use of substructures of the domain. The practicality of the reduction relies on a local basis, presented here, for the divergence-free subspace of the velocity space. We consider additive and multiplicative domain decomposition methods for solving the reduced elliptic problem, and their uniform convergence is established.
Tang Shaojie; Tang Xiangyang
2012-09-15
Purposes: The suppression of noise in x-ray computed tomography (CT) imaging is of clinical relevance for diagnostic image quality and the potential for radiation dose saving. Toward this purpose, statistical noise reduction methods in either the image or projection domain have been proposed, which employ a multiscale decomposition to enhance the performance of noise suppression while maintaining image sharpness. Recognizing the advantages of noise suppression in the projection domain, the authors propose a projection domain multiscale penalized weighted least squares (PWLS) method, in which the angular sampling rate is explicitly taken into consideration to account for the possible variation of interview sampling rate in advanced clinical or preclinical applications. Methods: The projection domain multiscale PWLS method is derived by converting an isotropic diffusion partial differential equation in the image domain into the projection domain, wherein a multiscale decomposition is carried out. With adoption of the Markov random field or soft thresholding objective function, the projection domain multiscale PWLS method deals with noise at each scale. To compensate for the degradation in image sharpness caused by the projection domain multiscale PWLS method, an edge enhancement is carried out following the noise reduction. The performance of the proposed method is experimentally evaluated and verified using the projection data simulated by computer and acquired by a CT scanner. Results: The preliminary results show that the proposed projection domain multiscale PWLS method outperforms the projection domain single-scale PWLS method and the image domain multiscale anisotropic diffusion method in noise reduction. In addition, the proposed method can preserve image sharpness very well while the occurrence of 'salt-and-pepper' noise and mosaic artifacts can be avoided. Conclusions: Since the interview sampling rate is taken into account in the projection domain
Random attractors for stochastic 2D-Navier-Stokes equations in some unbounded domains
NASA Astrophysics Data System (ADS)
Brzeźniak, Z.; Caraballo, T.; Langa, J. A.; Li, Y.; Łukaszewicz, G.; Real, J.
We show that the stochastic flow generated by the 2-dimensional Stochastic Navier-Stokes equations with rough noise on a Poincaré-like domain has a unique random attractor. One of the technical problems associated with the rough noise is overcomed by the use of the corresponding Cameron-Martin (or reproducing kernel Hilbert) space. Our results complement the result by Brzeźniak and Li (2006) [10] who showed that the corresponding flow is asymptotically compact and also generalize Caraballo et al. (2006) [12] who proved existence of a unique attractor for the time-dependent deterministic Navier-Stokes equations.
NASA Astrophysics Data System (ADS)
Panov, Yu. D.; Moskvin, A. S.; Rybakov, F. N.; Borisov, A. B.
2016-01-01
We made use of a special algorithm for compute unified device architecture for NVIDIA graphics cards, a nonlinear conjugate-gradient method to minimize energy functional, and Monte-Carlo technique to directly observe the forming of the ground state configuration for the 2D hard-core bosons by lowering the temperature and its evolution with deviation away from half-filling. The novel technique allowed us to examine earlier implications and uncover novel features of the phase transitions, in particular, look upon the nucleation of the odd domain structure, emergence of filamentary superfluidity nucleated at the antiphase domain walls of the charge-ordered phase, and nucleation and evolution of different topological structures.
Domain decomposition by the advancing-partition method for parallel unstructured grid generation
NASA Technical Reports Server (NTRS)
Pirzadeh, Shahyar Z. (Inventor); Banihashemi, legal representative, Soheila (Inventor)
2012-01-01
In a method for domain decomposition for generating unstructured grids, a surface mesh is generated for a spatial domain. A location of a partition plane dividing the domain into two sections is determined. Triangular faces on the surface mesh that intersect the partition plane are identified. A partition grid of tetrahedral cells, dividing the domain into two sub-domains, is generated using a marching process in which a front comprises only faces of new cells which intersect the partition plane. The partition grid is generated until no active faces remain on the front. Triangular faces on each side of the partition plane are collected into two separate subsets. Each subset of triangular faces is renumbered locally and a local/global mapping is created for each sub-domain. A volume grid is generated for each sub-domain. The partition grid and volume grids are then merged using the local-global mapping.
NASA Astrophysics Data System (ADS)
Anistratov, Dmitriy Y.; Azmy, Yousry Y.
2015-09-01
We study convergence of the integral transport matrix method (ITMM) based on a parallel block Jacobi (PBJ) iterative strategy for solving particle transport problems. The ITMM is a spatial domain decomposition method proposed for massively parallel computations. A Fourier analysis of the PBJ-based iterations applied to SN diamond-difference equations in 1D slab and 2D Cartesian geometries is performed. It is carried out for infinite-medium problems with homogeneous material properties. To analyze the performance of the ITMM with the PBJ algorithm and evaluate its potential in scalability we consider a limiting case of one spatial cell per subdomain. The analysis shows that in such limit the spectral radius of the iteration method is one without regard to values of the scattering ratio and optical thickness of the spatial cells. This implies lack of convergence in infinite medium. Numerical results of finite-medium problems are presented. They demonstrate effects of finite size of spatial domain on the performance of the iteration algorithm as well as its asymptotic behavior when the extent of the spatial domain increases. These numerical experiments also show that for finite domains iterative convergence to a finite criterion is achievable in a multiple of the sum of number of cells in each dimension.
MAST-2D diffusive model for flood prediction on domains with triangular Delaunay unstructured meshes
NASA Astrophysics Data System (ADS)
Aricò, C.; Sinagra, M.; Begnudelli, L.; Tucciarelli, T.
2011-11-01
A new methodology for the solution of the 2D diffusive shallow water equations over Delaunay unstructured triangular meshes is presented. Before developing the new algorithm, the following question is addressed: it is worth developing and using a simplified shallow water model, when well established algorithms for the solution of the complete one do exist? The governing Partial Differential Equations are discretized using a procedure similar to the linear conforming Finite Element Galerkin scheme, with a different flux formulation and a special flux treatment that requires Delaunay triangulation but entire solution monotonicity. A simple mesh adjustment is suggested, that attains the Delaunay condition for all the triangle sides without changing the original nodes location and also maintains the internal boundaries. The original governing system is solved applying a fractional time step procedure, that solves consecutively a convective prediction system and a diffusive correction system. The non linear components of the problem are concentrated in the prediction step, while the correction step leads to the solution of a linear system of the order of the number of computational cells. A semi-analytical procedure is applied for the solution of the prediction step. The discretized formulation of the governing equations allows to handle also wetting and drying processes without any additional specific treatment. Local energy dissipations, mainly the effect of vertical walls and hydraulic jumps, can be easily included in the model. Several numerical experiments have been carried out in order to test (1) the stability of the proposed model with regard to the size of the Courant number and to the mesh irregularity, (2) its computational performance, (3) the convergence order by means of mesh refinement. The model results are also compared with the results obtained by a fully dynamic model. Finally, the application to a real field case with a Venturi channel is presented.
Gray, Elizabeth J.; Petsalaki, Evangelia; James, D. Andrew; Bagshaw, Richard D.; Stacey, Melissa M.; Rocks, Oliver; Gingras, Anne-Claude; Pawson, Tony
2014-01-01
SH2D5 is a mammalian-specific, uncharacterized adaptor-like protein that contains an N-terminal phosphotyrosine-binding domain and a C-terminal Src homology 2 (SH2) domain. We show that SH2D5 is highly enriched in adult mouse brain, particularly in Purkinjie cells in the cerebellum and the cornu ammonis of the hippocampus. Despite harboring two potential phosphotyrosine (Tyr(P)) recognition domains, SH2D5 binds minimally to Tyr(P) ligands, consistent with the absence of a conserved Tyr(P)-binding arginine residue in the SH2 domain. Immunoprecipitation coupled to mass spectrometry (IP-MS) from cultured cells revealed a prominent association of SH2D5 with breakpoint cluster region protein, a RacGAP that is also highly expressed in brain. This interaction occurred between the phosphotyrosine-binding domain of SH2D5 and an NxxF motif located within the N-terminal region of the breakpoint cluster region. siRNA-mediated depletion of SH2D5 in a neuroblastoma cell line, B35, induced a cell rounding phenotype correlated with low levels of activated Rac1-GTP, suggesting that SH2D5 affects Rac1-GTP levels. Taken together, our data provide the first characterization of the SH2D5 signaling protein. PMID:25331951
TU-F-18A-02: Iterative Image-Domain Decomposition for Dual-Energy CT
Niu, T; Dong, X; Petrongolo, M; Zhu, L
2014-06-15
Purpose: Dual energy CT (DECT) imaging plays an important role in advanced imaging applications due to its material decomposition capability. Direct decomposition via matrix inversion suffers from significant degradation of image signal-to-noise ratios, which reduces clinical value. Existing de-noising algorithms achieve suboptimal performance since they suppress image noise either before or after the decomposition and do not fully explore the noise statistical properties of the decomposition process. We propose an iterative image-domain decomposition method for noise suppression in DECT, using the full variance-covariance matrix of the decomposed images. Methods: The proposed algorithm is formulated in the form of least-square estimation with smoothness regularization. It includes the inverse of the estimated variance-covariance matrix of the decomposed images as the penalty weight in the least-square term. Performance is evaluated using an evaluation phantom (Catphan 600) and an anthropomorphic head phantom. Results are compared to those generated using direct matrix inversion with no noise suppression, a de-noising method applied on the decomposed images, and an existing algorithm with similar formulation but with an edge-preserving regularization term. Results: On the Catphan phantom, our method retains the same spatial resolution as the CT images before decomposition while reducing the noise standard deviation of decomposed images by over 98%. The other methods either degrade spatial resolution or achieve less low-contrast detectability. Also, our method yields lower electron density measurement error than direct matrix inversion and reduces error variation by over 97%. On the head phantom, it reduces the noise standard deviation of decomposed images by over 97% without blurring the sinus structures. Conclusion: We propose an iterative image-domain decomposition method for DECT. The method combines noise suppression and material decomposition into an iterative
Statistical abstraction of high-fidelity CO2 pressure histories in 2-D, uniform, cylindrical domains
Letellier, Bruce C; Sanzo, Dean L; Pawar, Rajesh J
2010-01-01
Long-term, deep, geologic sequestration of carbon dioxide (CO{sub 2}) is being evaluated as a world-wide strategy for limiting anthropogenic carbon emissions to the atmosphere. A key element of this evaluation is quantification of the ancillary risks associated with this fundamentally new linkage between the global energy economy and the subsurface ecosphere. Quantitative risk assessment methods traditionally enumerate operational scenarios and describe the multiple physical responses that may ensue from each scenario depending on the quality of information that is available to describe identified system dependencies. For example, multiplepoint injection of compressed CO{sub 2} into a geologic reservoir having a nominal stratigraphy will create a pressurized zone of liquid that migrates through the rock. Scenarios that postulate CO{sub 2} encountering previously undetected wells or natural fractures in the caprock that represent leakage paths to the surface must be treated in a probabilistic format that accommodates unknown details in the subsurface geology. Fluid pressure in the reservoir at the location of the potential transport path drives any potential leakage that might occur, so the spatial and temporal distribution of CO{sub 2} overpressure represents an important metric for numeric simulation. State-of-the-art geologic transport models like FEHM, TUFF, and ECLIPSE (Refs. 1, 2, 3) can accurately simulate multi phase gas migration in a fully characterized geologic domain. However, each simulation can require time periods ranging between minutes and hours to achieve acceptable numerical performance, so it is often impractical to link predictive physics models directly in a quantitative risk assessment that will require transport estimates for thousands of scenarios. When direct computation is not possible, a library of high-fidelity calculations can sometimes be distilled to a simplified statistical correlation that spans the variability in all relevant input
A domain decomposition algorithm for solving large elliptic problems
Nolan, M.P.
1991-01-01
AN algorithm which efficiently solves large systems of equations arising from the discretization of a single second-order elliptic partial differential equation is discussed. The global domain is partitioned into not necessarily disjoint subdomains which are traversed using the Schwarz Alternating Procedure. On each subdomain the multigrid method is used to advance the solution. The algorithm has the potential to decrease solution time when data is stored across multiple levels of a memory hierarchy. Results are presented for a virtual memory, vector multiprocessor architecture. A study of choice of inner iteration procedure and subdomain overlap is presented for a model problem, solved with two and four subdomains, sequentially and in parallel. Microtasking multiprocessing results are reported for multigrid on the Alliant FX-8 vector-multiprocessor. A convergence proof for a class of matrix splittings for the two-dimensional Helmholtz equation is given. 70 refs., 3 figs., 20 tabs.
Multi-Zone Liquid Thrust Chamber Performance Code with Domain Decomposition for Parallel Processing
NASA Technical Reports Server (NTRS)
Navaz, Homayun K.
2002-01-01
Computational Fluid Dynamics (CFD) has considerably evolved in the last decade. There are many computer programs that can perform computations on viscous internal or external flows with chemical reactions. CFD has become a commonly used tool in the design and analysis of gas turbines, ramjet combustors, turbo-machinery, inlet ducts, rocket engines, jet interaction, missile, and ramjet nozzles. One of the problems of interest to NASA has always been the performance prediction for rocket and air-breathing engines. Due to the complexity of flow in these engines it is necessary to resolve the flowfield into a fine mesh to capture quantities like turbulence and heat transfer. However, calculation on a high-resolution grid is associated with a prohibitively increasing computational time that can downgrade the value of the CFD for practical engineering calculations. The Liquid Thrust Chamber Performance (LTCP) code was developed for NASA/MSFC (Marshall Space Flight Center) to perform liquid rocket engine performance calculations. This code is a 2D/axisymmetric full Navier-Stokes (NS) solver with fully coupled finite rate chemistry and Eulerian treatment of liquid fuel and/or oxidizer droplets. One of the advantages of this code has been the resemblance of its input file to the JANNAF (Joint Army Navy NASA Air Force Interagency Propulsion Committee) standard TDK code, and its automatic grid generation for JANNAF defined combustion chamber wall geometry. These options minimize the learning effort for TDK users, and make the code a good candidate for performing engineering calculations. Although the LTCP code was developed for liquid rocket engines, it is a general-purpose code and has been used for solving many engineering problems. However, the single zone formulation of the LTCP has limited the code to be applicable to problems with complex geometry. Furthermore, the computational time becomes prohibitively large for high-resolution problems with chemistry, two
NASA Astrophysics Data System (ADS)
Lahmiri, Salim
2016-08-01
The main purpose of this work is to explore the usefulness of fractal descriptors estimated in multi-resolution domains to characterize biomedical digital image texture. In this regard, three multi-resolution techniques are considered: the well-known discrete wavelet transform (DWT) and the empirical mode decomposition (EMD), and; the newly introduced; variational mode decomposition mode (VMD). The original image is decomposed by the DWT, EMD, and VMD into different scales. Then, Fourier spectrum based fractal descriptors is estimated at specific scales and directions to characterize the image. The support vector machine (SVM) was used to perform supervised classification. The empirical study was applied to the problem of distinguishing between normal and abnormal brain magnetic resonance images (MRI) affected with Alzheimer disease (AD). Our results demonstrate that fractal descriptors estimated in VMD domain outperform those estimated in DWT and EMD domains; and also those directly estimated from the original image.
Multi-resolution flow simulations by smoothed particle hydrodynamics via domain decomposition
NASA Astrophysics Data System (ADS)
Bian, Xin; Li, Zhen; Tang, Yu-Hang; Karniadakis, George
2015-11-01
We present a methodology to concurrently couple particle-based methods via a domain decomposition (DD) technique for simulating viscous flows. In particular, we select two resolutions of the smoothed particle hydrodynamics (SPH) method as demonstration. Within the DD framework, a simulation domain is decomposed into two (or more) overlapping sub-domains, each of which has an individual particle scale determined by the local flow physics. Consistency of the two sub-domains is achieved in the overlap region by matching the two independent simulations based on Lagrangian interpolation of state variables and fluxes. The domain decomposition based SPH method (DD-SPH) employs different spatial and temporal resolutions, and hence, each sub-domain has its own smoothing length and time step. As a consequence, particle refinement and de-refinement are performed asynchronously according to individual time advancement of each sub-domain. The proposed strategy avoids SPH force interactions between different resolutions on purpose, so that coupling, in principle, can go beyond SPH - SPH, and may allow SPH to be coupled with other mesoscopic or microscopic particle methods. The DD-SPH method is validated first for a transient Couette flow, where simulation results base. US DOE Collaboratory on Mathematics for Mesoscopic Modeling of Materials (CM4).
Martino, Mikaël M.; Mochizuki, Mayumi; Rothenfluh, Dominique A.; Rempel, Sandra A.; Hubbell, Jeffrey A.; Barker, Thomas H.
2009-01-01
The extracellular matrix (ECM) exerts powerful control over many cellular phenomena, including stem cell differentiation. As such, design and modulation of ECM analogs to ligate specific integrin is a promising approach to control cellular processes in vitro and in vivo for regenerative medicine strategies. Although fibronectin (FN), a crucial ECM protein in tissue development and repair, and its RGD peptide are widely used for cell adhesion, the promiscuity with which they engage integrins leads to difficulty in control of receptor-specific interactions. Recent simulations of force-mediated unfolding of FN domains and sequences analysis of human versus mouse FN suggest that the structural stability of the FN’s central cell-binding domains (FN III9-10) affects its integrin specificity. Through production of FN III9-10 variants with variable stabilities, we obtained ligands that present different specificities for the integrin α5β1 and that can be covalently linked into fibrin matrices. Here, we demonstrate the capacity of α5β1 integrin-specific engagement to influence human mesenchymal stem cell (MSC) behavior in 2D and 3D environments. Our data indicate that α5β1 has an important role in the control of MSC osteogenic differentiation. FN fragments with increased specificity for α5β1 versus αvβ3 results in significantly enhanced osteogenic differentiation of MSCs in 2D and in a clinically relevant 3D fibrin matrix system, although attachment/spreading and proliferation were comparable with that on full-length FN. This work shows how integrin-dependant cellular interactions with the ECM can be engineered to control stem cell fate, within a system appropriate for both 3D cell culture and tissue engineering. PMID:19027948
NASA Astrophysics Data System (ADS)
Cafiero, M.; Lloberas-Valls, O.; Cante, J.; Oliver, J.
2016-04-01
A domain decomposition technique is proposed which is capable of properly connecting arbitrary non-conforming interfaces. The strategy essentially consists in considering a fictitious zero-width interface between the non-matching meshes which is discretized using a Delaunay triangulation. Continuity is satisfied across domains through normal and tangential stresses provided by the discretized interface and inserted in the formulation in the form of Lagrange multipliers. The final structure of the global system of equations resembles the dual assembly of substructures where the Lagrange multipliers are employed to nullify the gap between domains. A new approach to handle floating subdomains is outlined which can be implemented without significantly altering the structure of standard industrial finite element codes. The effectiveness of the developed algorithm is demonstrated through a patch test example and a number of tests that highlight the accuracy of the methodology and independence of the results with respect to the framework parameters. Considering its high degree of flexibility and non-intrusive character, the proposed domain decomposition framework is regarded as an attractive alternative to other established techniques such as the mortar approach.
Sapphire decomposition and inversion domains in N-polar aluminum nitride
Hussey, Lindsay White, Ryan M.; Kirste, Ronny; Bryan, Isaac; Guo, Wei; Osterman, Katherine; Haidet, Brian; Bryan, Zachary; Bobea, Milena; Collazo, Ramón; Sitar, Zlatko; Mita, Seiji
2014-01-20
Transmission electron microscopy (TEM) techniques and potassium hydroxide (KOH) etching confirmed that inversion domains in the N-polar AlN grown on c-plane sapphire were due to the decomposition of sapphire in the presence of hydrogen. The inversion domains were found to correspond to voids at the AlN and sapphire interface, and transmission electron microscopy results showed a V-shaped, columnar inversion domain with staggered domain boundary sidewalls. Voids were also observed in the simultaneously grown Al-polar AlN, however no inversion domains were present. The polarity of AlN grown above the decomposed regions of the sapphire substrate was confirmed to be Al-polar by KOH etching and TEM.
Wakeling, James M.
2014-01-01
When skeletal muscle fibres shorten, they must increase in their transverse dimensions in order to maintain a constant volume. In pennate muscle, this transverse expansion results in the fibres rotating to greater pennation angle, with a consequent reduction in their contractile velocity in a process known as gearing. Understanding the nature and extent of this transverse expansion is necessary to understand the mechanisms driving the changes in internal geometry of whole muscles during contraction. Current methodologies allow the fascicle lengths, orientations, and curvatures to be quantified, but not the transverse expansion. The purpose of this study was to develop and validate techniques for quantifying transverse strain in skeletal muscle fascicles during contraction from B-mode ultrasound images. Images were acquired from the medial and lateral gastrocnemii during cyclic contractions, enhanced using multiscale vessel enhancement filtering and the spatial frequencies resolved using 2D discrete Fourier transforms. The frequency information was resolved into the fascicle orientations that were validated against manually digitized values. The transverse fascicle strains were calculated from their wavelengths within the images. These methods showed that the transverse strain increases while the longitudinal fascicle length decreases; however, the extent of these strains was smaller than expected. PMID:25328509
Baker, Theresa; Nerle, Sujata; Pritchard, Justin; Zhao, Boyang; Rivera, Victor M.
2015-01-01
Although targeted therapies have revolutionized cancer treatment, overcoming acquired resistance remains a major clinical challenge. EZH2 inhibitors (EZH2i), EPZ-6438 and GSK126, are currently in the early stages of clinical evaluation and the first encouraging signs of efficacy have recently emerged in the clinic. To anticipate mechanisms of resistance to EZH2i, we used a forward genetic platform combining a mutagenesis screen with next generation sequencing technology and identified a hotspot of secondary mutations in the EZH2 D1 domain (Y111 and I109). Y111D mutation within the WT or A677G EZH2 allele conferred robust resistance to both EPZ-6438 and GSK126, but it only drove a partial resistance within the Y641F allele. EZH2 mutants required histone methyltransferase (HMT) catalytic activity and the polycomb repressive complex 2 (PRC2) components, SUZ12 and EED, to drive drug resistance. Furthermore, D1 domain mutations not only blocked the ability of EZH2i to bind to WT and A677G mutant, but also abrogated drug binding to the Y641F mutant. These data provide the first cellular validation of the mechanistic model underpinning the oncogenic function of WT and mutant EZH2. Importantly, our findings suggest that acquired-resistance to EZH2i may arise in WT and mutant EZH2 patients through a single mutation that remains targetable by second generation EZH2i. PMID:26360609
Lahmiri, Salim
2014-09-01
Hybrid denoising models based on combining empirical mode decomposition (EMD) and discrete wavelet transform (DWT) were found to be effective in removing additive Gaussian noise from electrocardiogram (ECG) signals. Recently, variational mode decomposition (VMD) has been proposed as a multiresolution technique that overcomes some of the limits of the EMD. Two ECG denoising approaches are compared. The first is based on denoising in the EMD domain by DWT thresholding, whereas the second is based on noise reduction in the VMD domain by DWT thresholding. Using signal-to-noise ratio and mean of squared errors as performance measures, simulation results show that the VMD-DWT approach outperforms the conventional EMD-DWT. In addition, a non-local means approach used as a reference technique provides better results than the VMD-DWT approach. PMID:26609387
Parallel domain decomposition method with non-blocking communication for flow through porous media
NASA Astrophysics Data System (ADS)
Lemmer, Andreas; Hilfer, Rudolf
2015-01-01
This paper introduces a domain decomposition method for numerically solving the Stokes equation for very large, complex geometries. Examples arise from realistic porous media. The computational method is based on the SIMPLE (Semi-Implicit Method for Pressure Linked Equations) algorithm which uses a finite-differences approach for discretizing the underlying equations. It achieves comparable speed and efficiency as lattice Boltzmann methods. The domain decomposition method splits a large three-dimensional region into slices that can be processed in parallel on multi-processor computation environments with only minimal communication between the computation nodes. With this method, the flow through a porous medium with grid sizes up to 20483 voxel has been calculated.
Domain Decomposition of a Constructive Solid Geometry Monte Carlo Transport Code
O'Brien, M J; Joy, K I; Procassini, R J; Greenman, G M
2008-12-07
Domain decomposition has been implemented in a Constructive Solid Geometry (CSG) Monte Carlo neutron transport code. Previous methods to parallelize a CSG code relied entirely on particle parallelism; but in our approach we distribute the geometry as well as the particles across processors. This enables calculations whose geometric description is larger than what could fit in memory of a single processor, thus it must be distributed across processors. In addition to enabling very large calculations, we show that domain decomposition can speed up calculations compared to particle parallelism alone. We also show results of a calculation of the proposed Laser Inertial-Confinement Fusion-Fission Energy (LIFE) facility, which has 5.6 million CSG parts.
2014-01-01
Hybrid denoising models based on combining empirical mode decomposition (EMD) and discrete wavelet transform (DWT) were found to be effective in removing additive Gaussian noise from electrocardiogram (ECG) signals. Recently, variational mode decomposition (VMD) has been proposed as a multiresolution technique that overcomes some of the limits of the EMD. Two ECG denoising approaches are compared. The first is based on denoising in the EMD domain by DWT thresholding, whereas the second is based on noise reduction in the VMD domain by DWT thresholding. Using signal-to-noise ratio and mean of squared errors as performance measures, simulation results show that the VMD-DWT approach outperforms the conventional EMD–DWT. In addition, a non-local means approach used as a reference technique provides better results than the VMD-DWT approach. PMID:26609387
A domain decomposition study of massively parallel computing in compressible gas dynamics
NASA Astrophysics Data System (ADS)
Wong, C. C.; Blottner, F. G.; Payne, J. L.; Soetrisno, M.
1995-03-01
The appropriate utilization of massively parallel computers for solving the Navier-Stokes equations is investigated and determined from an engineering perspective. The issues investigated are: (1) Should strip or patch domain decomposition of the spatial mesh be used to reduce computer time? (2) How many computer nodes should be used for a problem with a given sized mesh to reduce computer time? (3) Is the convergence of the Navier-Stokes solution procedure (LU-SGS) adversely influenced by the domain decomposition approach? The results of the paper show that the present Navier-Stokes solution technique has good performance on a massively parallel computer for transient flow problems. For steady-state problems with a large number of mesh cells, the solution procedure will require significant computer time due to an increased number of iterations to achieve a converged solution. There is an optimum number of computer nodes to use for a problem with a given global mesh size.
Bi-planar 2D-to-3D registration in Fourier domain for stereoscopic x-ray motion tracking
NASA Astrophysics Data System (ADS)
Zosso, Dominique; Le Callennec, Benoît; Bach Cuadra, Meritxell; Aminian, Kamiar; Jolles, Brigitte M.; Thiran, Jean-Philippe
2008-03-01
In this paper we present a new method to track bone movements in stereoscopic X-ray image series of the knee joint. The method is based on two different X-ray image sets: a rotational series of acquisitions of the still subject knee that allows the tomographic reconstruction of the three-dimensional volume (model), and a stereoscopic image series of orthogonal projections as the subject performs movements. Tracking the movements of bones throughout the stereoscopic image series means to determine, for each frame, the best pose of every moving element (bone) previously identified in the 3D reconstructed model. The quality of a pose is reflected in the similarity between its theoretical projections and the actual radiographs. We use direct Fourier reconstruction to approximate the three-dimensional volume of the knee joint. Then, to avoid the expensive computation of digitally rendered radiographs (DRR) for pose recovery, we develop a corollary to the 3-dimensional central-slice theorem and reformulate the tracking problem in the Fourier domain. Under the hypothesis of parallel X-ray beams, the heavy 2D-to-3D registration of projections in the signal domain is replaced by efficient slice-to-volume registration in the Fourier domain. Focusing on rotational movements, the translation-relevant phase information can be discarded and we only consider scalar Fourier amplitudes. The core of our motion tracking algorithm can be implemented as a classical frame-wise slice-to-volume registration task. Results on both synthetic and real images confirm the validity of our approach.
NASA Astrophysics Data System (ADS)
Fujishima, K.; Komasa, M.; Kitamura, S.; Tomita, M.; Kanai, A.
Proteins are a major regulatory component in complex biological systems.Among them, DNA/RNA-binding proteins, the key components of the central dogma of molecular biology, and membrane proteins, which are necessary for both signal transduction and metabolite transport, are suggested to be the most important protein families that arose in the early stage of life. In this study, we computationally analyzed the whole proteome data of six model species to overview the protein diversity in the three domains of life (Bacteria, Archaea and Eukaryota), especially focusing on the above two protein families. To compare the protein distribution among the six model species, we calculated various protein profiles: hydropathy, molecular weight, amino acid composition and periodicity for each protein. We found a domain-specific distribution of the proteome based on 2D correlation analysis of hydropathy and molecular weight. Further, the merged protein distribution of Archaea and other do mains revealed many membrane proteins localized in Bacteria-specific regions with a high ratio of hydropathy and many DNA/RNA-binding proteins localized in Eukaryota-specific regions with a low ratio of hydropathy. Since about half of the proteins encoded in the genome are still functionally unknown, we further conducted Support Vector Machine (SVM)-based functional prediction using amino acid composition (CO score) and periodicity (PD score) as feature vectors to predict the overall number of DNA/RNA-binding proteins and membrane proteins in the proteome. Our estimation indicated that two functional categories occupy approximately 60% to 80% of the proteome, and further, the proportion of the two categories varied among the three domains of life, suggesting that the proteome has gone through different selective pressure during evolution.
Luo, Wenbin; Yao, Xiaolan; Hong, Mei
2005-05-01
One of the main mechanisms of membrane protein folding is by spontaneous insertion into the lipid bilayer from the aqueous environment. The bacterial toxin, colicin Ia, is one such protein. To shed light on the conformational changes involved in this dramatic transfer from the polar to the hydrophobic milieu, we carried out 2D magic-angle spinning (13)C NMR experiments on the water-soluble and membrane-bound states of the channel-forming domain of colicin Ia. Proton-driven (13)C spin diffusion spectra of selectively (13)C-labeled protein show unequivocal attenuation of cross-peaks after membrane binding. This attenuation can be assigned to distance increases but not reduction of the diffusion coefficient. Analysis of the statistics of the interhelical and intrahelical (13)C-(13)C distances in the soluble protein structure indicates that the observed cross-peak reduction is well correlated with a high percentage of short interhelical contacts in the soluble protein. This suggests that colicin Ia channel domain becomes open and extended upon membrane binding, thus lengthening interhelical distances. In comparison, cross-peaks with similar intensities between the two states are dominated by intrahelical contacts in the soluble state. This suggests that the membrane-bound structure of colicin Ia channel domain may be described as a "molten globule", in which the helical secondary structure is retained while the tertiary structure is unfolded. This study demonstrates that (13)C spin diffusion NMR is a valuable tool for obtaining qualitative long-range distance constraints on membrane protein folding. PMID:15853348
Provatas, Nikolas; Leonard, Francois Leonard; Mahon, Jennifer; Haataja, Mikko
2005-06-01
In this letter, we examine the effects of discrete mobile dislocations on spinodal decomposition kinetics in lattice mismatched binary alloys. By employing a novel continuum model, we demonstrate that the effects of dislocation mobility on domain coarsening kinetics can be expressed in a unified manner through a scaling function, describing a crossover from t{sup 1/2} to t{sup 1/3} behavior.
Griebel, M.
1994-12-31
In recent years, it has turned out that many modern iterative algorithms (multigrid schemes, multilevel preconditioners, domain decomposition methods etc.) for solving problems resulting from the discretization of PDEs can be interpreted as additive (Jacobi-like) or multiplicative (Gauss-Seidel-like) subspace correction methods. The key to their analysis is the study of certain metric properties of the underlying splitting of the discretization space V into a sum of subspaces V{sub j}, j = 1{hor_ellipsis}, J resp. of the variational problem on V into auxiliary problems on these subspaces. Here, the author proposes a modified approach to the abstract convergence theory of these additive and multiplicative Schwarz iterative methods, that makes the relation to traditional iteration methods more explicit. To this end he introduces the enlarged Hilbert space V = V{sub 0} x {hor_ellipsis} x V{sub j} which is nothing else but the usual construction of the Cartesian product of the Hilbert spaces V{sub j} and use it now in the discretization process. This results in an enlarged, semidefinite linear system to be solved instead of the usual definite system. Then, modern multilevel methods as well as domain decomposition methods simplify to just traditional (block-) iteration methods. Now, the convergence analysis can be carried out directly for these traditional iterations on the enlarged system, making convergence proofs of multilevel and domain decomposition methods more clear, or, at least, more classical. The terms that enter the convergence proofs are exactly the ones of the classical iterative methods. It remains to estimate them properly. The convergence proof itself follow basically line by line the old proofs of the respective traditional iterative methods. Additionally, new multilevel/domain decomposition methods are constructed straightforwardly by now applying just other old and well known traditional iterative methods to the enlarged system.
Algebraic multigrid domain and range decomposition (AMG-DD / AMG-RD)*
Bank, R.; Falgout, R. D.; Jones, T.; Manteuffel, T. A.; McCormick, S. F.; Ruge, J. W.
2015-10-29
In modern large-scale supercomputing applications, algebraic multigrid (AMG) is a leading choice for solving matrix equations. However, the high cost of communication relative to that of computation is a concern for the scalability of traditional implementations of AMG on emerging architectures. This paper introduces two new algebraic multilevel algorithms, algebraic multigrid domain decomposition (AMG-DD) and algebraic multigrid range decomposition (AMG-RD), that replace traditional AMG V-cycles with a fully overlapping domain decomposition approach. While the methods introduced here are similar in spirit to the geometric methods developed by Brandt and Diskin [Multigrid solvers on decomposed domains, in Domain Decomposition Methods inmore » Science and Engineering, Contemp. Math. 157, AMS, Providence, RI, 1994, pp. 135--155], Mitchell [Electron. Trans. Numer. Anal., 6 (1997), pp. 224--233], and Bank and Holst [SIAM J. Sci. Comput., 22 (2000), pp. 1411--1443], they differ primarily in that they are purely algebraic: AMG-RD and AMG-DD trade communication for computation by forming global composite “grids” based only on the matrix, not the geometry. (As is the usual AMG convention, “grids” here should be taken only in the algebraic sense, regardless of whether or not it corresponds to any geometry.) Another important distinguishing feature of AMG-RD and AMG-DD is their novel residual communication process that enables effective parallel computation on composite grids, avoiding the all-to-all communication costs of the geometric methods. The main purpose of this paper is to study the potential of these two algebraic methods as possible alternatives to existing AMG approaches for future parallel machines. As a result, this paper develops some theoretical properties of these methods and reports on serial numerical tests of their convergence properties over a spectrum of problem parameters.« less
Algebraic multigrid domain and range decomposition (AMG-DD / AMG-RD)*
Bank, R.; Falgout, R. D.; Jones, T.; Manteuffel, T. A.; McCormick, S. F.; Ruge, J. W.
2015-10-29
In modern large-scale supercomputing applications, algebraic multigrid (AMG) is a leading choice for solving matrix equations. However, the high cost of communication relative to that of computation is a concern for the scalability of traditional implementations of AMG on emerging architectures. This paper introduces two new algebraic multilevel algorithms, algebraic multigrid domain decomposition (AMG-DD) and algebraic multigrid range decomposition (AMG-RD), that replace traditional AMG V-cycles with a fully overlapping domain decomposition approach. While the methods introduced here are similar in spirit to the geometric methods developed by Brandt and Diskin [Multigrid solvers on decomposed domains, in Domain Decomposition Methods in Science and Engineering, Contemp. Math. 157, AMS, Providence, RI, 1994, pp. 135--155], Mitchell [Electron. Trans. Numer. Anal., 6 (1997), pp. 224--233], and Bank and Holst [SIAM J. Sci. Comput., 22 (2000), pp. 1411--1443], they differ primarily in that they are purely algebraic: AMG-RD and AMG-DD trade communication for computation by forming global composite “grids” based only on the matrix, not the geometry. (As is the usual AMG convention, “grids” here should be taken only in the algebraic sense, regardless of whether or not it corresponds to any geometry.) Another important distinguishing feature of AMG-RD and AMG-DD is their novel residual communication process that enables effective parallel computation on composite grids, avoiding the all-to-all communication costs of the geometric methods. The main purpose of this paper is to study the potential of these two algebraic methods as possible alternatives to existing AMG approaches for future parallel machines. As a result, this paper develops some theoretical properties of these methods and reports on serial numerical tests of their convergence properties over a spectrum of problem parameters.
Jemcov, A.; Matovic, M.D.
1996-12-31
This paper examines the sparse representation and preconditioning of a discrete Steklov-Poincare operator which arises in domain decomposition methods. A non-overlapping domain decomposition method is applied to a second order self-adjoint elliptic operator (Poisson equation), with homogeneous boundary conditions, as a model problem. It is shown that the discrete Steklov-Poincare operator allows sparse representation with a bounded condition number in wavelet basis if the transformation is followed by thresholding and resealing. These two steps combined enable the effective use of Krylov subspace methods as an iterative solution procedure for the system of linear equations. Finding the solution of an interface problem in domain decomposition methods, known as a Schur complement problem, has been shown to be equivalent to the discrete form of Steklov-Poincare operator. A common way to obtain Schur complement matrix is by ordering the matrix of discrete differential operator in subdomain node groups then block eliminating interface nodes. The result is a dense matrix which corresponds to the interface problem. This is equivalent to reducing the original problem to several smaller differential problems and one boundary integral equation problem for the subdomain interface.
A study of domain decomposition methods applied to the discretized Helmholtz equation
NASA Astrophysics Data System (ADS)
Tramel, Robert Wallace
2001-09-01
In this work a domain decomposition based preconditioner of the additive Schwarz type is developed and tested on the linear systems which arise out of the application of the Green's Function/Wave Expansion Discretization. (GFD/WED) method to Helmholtz's equation. In order to develop the additive Schwarz preconditioner, use is made of a class of one-sided Artificial Radiation Boundary Conditions (ARBC) developed during the course of this work. These ARBCs are computationally shown to be quite accurate for use on their own. The ARBC's are used to radiatively couple the various sub-domains which are naturally part of domain decomposition based methods in such a manner as to ensure that the system matrix, when restricted to the sub-domains, is non-singular. In addition, the inter-domain ARBC is constructed such that the solution to the global linear system is unaffected by the presence of the artificial boundaries. The efficacy and efficiency of the method is demonstrated on one, two, and three-dimensional test cases.
NASA Astrophysics Data System (ADS)
Duy, Truong Vinh Truong; Ozaki, Taisuke
2014-03-01
With tens of petaflops supercomputers already in operation and exaflops machines expected to appear within the next 10 years, efficient parallel computational methods are required to take advantage of such extreme-scale machines. In this paper, we present a three-dimensional domain decomposition scheme for enabling large-scale electronic structure calculations based on density functional theory (DFT) on massively parallel computers. It is composed of two methods: (i) the atom decomposition method and (ii) the grid decomposition method. In the former method, we develop a modified recursive bisection method based on the moment of inertia tensor to reorder the atoms along a principal axis so that atoms that are close in real space are also close on the axis to ensure data locality. The atoms are then divided into sub-domains depending on their projections onto the principal axis in a balanced way among the processes. In the latter method, we define four data structures for the partitioning of grid points that are carefully constructed to make data locality consistent with that of the clustered atoms for minimizing data communications between the processes. We also propose a decomposition method for solving the Poisson equation using the three-dimensional FFT in Hartree potential calculation, which is shown to be better in terms of communication efficiency than a previously proposed parallelization method based on a two-dimensional decomposition. For evaluation, we perform benchmark calculations with our open-source DFT code, OpenMX, paying particular attention to the O(N) Krylov subspace method. The results show that our scheme exhibits good strong and weak scaling properties, with the parallel efficiency at 131,072 cores being 67.7% compared to the baseline of 16,384 cores with 131,072 atoms of the diamond structure on the K computer.
2D multi-parameter elastic seismic imaging by frequency-domain L1-norm full waveform inversion
NASA Astrophysics Data System (ADS)
Brossier, Romain; Operto, Stéphane; Virieux, Jean
2010-05-01
Full waveform inversion (FWI) is becoming a powerful and efficient tool to derive high-resolution quantitative models of the subsurface. In the frequency-domain, computationally efficient FWI algorithms can be designed for wide-aperture acquisition geometries by limiting inversion to few discrete frequencies. However, FWI remains an ill-posed and highly non-linear data-fitting procedure that is sensitive to noise, inaccuracies of the starting model and definition of multiparameter classes. The footprint of the noise in seismic imaging is conventionally mitigated by stacking highly redundant multifold data. However, when the data redundancy is decimated in the framework of efficient frequency-domain FWI, it is essential to assess the sensitivity of the inversion to noise. The impact of the noise in FWI, when applied to decimated data sets, has been marginally illustrated in the past and least-squares minimisation has remained the most popular approach. We investigate in this study the sensitivity of frequency-domain elastic FWI to noise for realistic onshore and offshore synthetic data sets contaminated by ambient random white noise. Four minimisation functionals are assessed in the framework of frequency domain FWI of decimated data: the classical least-square norm (L2), the least-absolute-values norm (L1), and some combinations of both (the Huber and the so-called Hybrid criteria). These functionals are implemented in a massively-parallel, 2D elastic frequency-domain FWI algorithm. A two-level hierarchical algorithm is implemented to mitigate the non-linearity of the inversion in complex environments. The first outer level consists of successive inversions of frequency groups of increasing high-frequency content. This level defines a multi-scale approach while preserving some data redundancy by means of simultaneous inversion of multiple frequencies. The second inner level used complex-valued frequencies for data preconditioning. This preconditioning controls the
Middleton, Beth A.
2014-01-01
A cornerstone of ecosystem ecology, decomposition was recognized as a fundamental process driving the exchange of energy in ecosystems by early ecologists such as Lindeman 1942 and Odum 1960). In the history of ecology, studies of decomposition were incorporated into the International Biological Program in the 1960s to compare the nature of organic matter breakdown in various ecosystem types. Such studies still have an important role in ecological studies of today. More recent refinements have brought debates on the relative role microbes, invertebrates and environment in the breakdown and release of carbon into the atmosphere, as well as how nutrient cycling, production and other ecosystem processes regulated by decomposition may shift with climate change. Therefore, this bibliography examines the primary literature related to organic matter breakdown, but it also explores topics in which decomposition plays a key supporting role including vegetation composition, latitudinal gradients, altered ecosystems, anthropogenic impacts, carbon storage, and climate change models. Knowledge of these topics is relevant to both the study of ecosystem ecology as well projections of future conditions for human societies.
Li,Jing; Tu, Xuemin
2008-12-10
A variant of balancing domain decomposition method by constraints (BDDC) is proposed for solving a class of indefinite system of linear equations, which arises from the finite element discretization of the Helmholtz equation of time-harmonic wave propagation in a bounded interior domain. The proposed BDDC algorithm is closely related to the dual-primal finite element tearing and interconnecting algorithm for solving Helmholtz equations (FETI-DPH). Under the condition that the diameters of the subdomains are small enough, the rate of convergence is established which depends polylogarithmically on the dimension of the individual subdomain problems and which improves with the decrease of the subdomain diameters. These results are supported by numerical experiments of solving a Helmholtz equation on a two-dimensional square domain.
Multi-resolution flow simulations by smoothed particle hydrodynamics via domain decomposition
NASA Astrophysics Data System (ADS)
Bian, Xin; Li, Zhen; Karniadakis, George Em
2015-09-01
We present a methodology to concurrently couple particle-based methods via a domain decomposition (DD) technique for simulating viscous flows. In particular, we select two resolutions of the smoothed particle hydrodynamics (SPH) method as demonstration. Within the DD framework, a simulation domain is decomposed into two (or more) overlapping sub-domains, each of which has an individual particle scale determined by the local flow physics. Consistency of the two sub-domains is achieved in the overlap region by matching the two independent simulations based on Lagrangian interpolation of state variables and fluxes. The domain decomposition based SPH method (DD-SPH) employs different spatial and temporal resolutions, and hence, each sub-domain has its own smoothing length and time step. As a consequence, particle refinement and de-refinement are performed asynchronously according to individual time advancement of each sub-domain. The proposed strategy avoids SPH force interactions between different resolutions on purpose, so that coupling, in principle, can go beyond SPH-SPH, and may allow SPH to be coupled with other mesoscopic or microscopic particle methods. The DD-SPH method is validated first for a transient Couette flow, where simulation results based on proper coupling of spatial-temporal scales agree well with analytical solutions. In particular, we find that the size of the overlap region should be at least rc,1 + 2rc,2, where rc,1 and rc,2 are cut off radii in the two sub-domains with rc,1 ≤rc,2. Subsequently, a perturbation wave is considered traveling either parallel or perpendicular to the hybrid interface. Compressibility is significant if transient behavior at short sonic-time-scale is relevant, while the fluid can be treated as quasi-incompressible at sufficiently long time scale. To this end, we propose a coupling of density fields from the two sub-domains. Finally, a steady Wannier flow is simulated, where a rotating cylinder is placed next to a
NASA Technical Reports Server (NTRS)
Shih, T. I.-P.; Bailey, R. T.; Nguyen, H. L.; Roelke, R. J.
1990-01-01
An efficient computer program, called GRID2D/3D was developed to generate single and composite grid systems within geometrically complex two- and three-dimensional (2- and 3-D) spatial domains that can deform with time. GRID2D/3D generates single grid systems by using algebraic grid generation methods based on transfinite interpolation in which the distribution of grid points within the spatial domain is controlled by stretching functions. All single grid systems generated by GRID2D/3D can have grid lines that are continuous and differentiable everywhere up to the second-order. Also, grid lines can intersect boundaries of the spatial domain orthogonally. GRID2D/3D generates composite grid systems by patching together two or more single grid systems. The patching can be discontinuous or continuous. For continuous composite grid systems, the grid lines are continuous and differentiable everywhere up to the second-order except at interfaces where different single grid systems meet. At interfaces where different single grid systems meet, the grid lines are only differentiable up to the first-order. For 2-D spatial domains, the boundary curves are described by using either cubic or tension spline interpolation. For 3-D spatial domains, the boundary surfaces are described by using either linear Coon's interpolation, bi-hyperbolic spline interpolation, or a new technique referred to as 3-D bi-directional Hermite interpolation. Since grid systems generated by algebraic methods can have grid lines that overlap one another, GRID2D/3D contains a graphics package for evaluating the grid systems generated. With the graphics package, the user can generate grid systems in an interactive manner with the grid generation part of GRID2D/3D. GRID2D/3D is written in FORTRAN 77 and can be run on any IBM PC, XT, or AT compatible computer. In order to use GRID2D/3D on workstations or mainframe computers, some minor modifications must be made in the graphics part of the program; no
Non-overlapping domain decomposition for near-wall turbulence modeling
NASA Astrophysics Data System (ADS)
Utyuzhnikov, Sergey
2016-06-01
Near-wall turbulence modeling is computationally a very expensive problem. The talk considers a novel approach based on non-overlapping domain decomposition. It allows us to avoid calculations of the region with high gradients in the vicinity of the wall while retaining sufficient overall accuracy. The technique is introduced in application to low-Reynolds number RANS models. The domain decomposition is achieved via the transfer of the boundary condition from the wall to an interface boundary. If the governing equations in the inner domain are simplified, then the interface boundary conditions are of Robin type. These boundary conditions can be obtained in an analytical form despite the fact that they are nonlinear. Possible ways to achieve a reasonable trade-off between efficiency and accuracy are discussed. The obtained interface boundary conditions are mesh-independent. They can be used to avoid the computationally expensive resolution of a high-gradient region near the wall. Moreover, once the solution is constructed in the outer region, the near-wall profile can be restored if required. In two extreme cases, if the interface boundary is too close to the wall or too far from it, the so-constructed solution to the problem automatically corresponds to low- and high-Reynolds number RANS models, respectively. Different applications are considered including unsteady problems and complex geometries. The developed approach proved to be quite robust and relatively universal. It does not contain any tuning parameters. The technique might be extended to other multiscale problems.
Parallel computing of a climate model on the dawn 1000 by domain decomposition method
NASA Astrophysics Data System (ADS)
Bi, Xunqiang
1997-12-01
In this paper the parallel computing of a grid-point nine-level atmospheric general circulation model on the Dawn 1000 is introduced. The model was developed by the Institute of Atmospheric Physics (IAP), Chinese Academy of Sciences (CAS). The Dawn 1000 is a MIMD massive parallel computer made by National Research Center for Intelligent Computer (NCIC), CAS. A two-dimensional domain decomposition method is adopted to perform the parallel computing. The potential ways to increase the speed-up ratio and exploit more resources of future massively parallel supercomputation are also discussed.
Domain decomposition methods for nonconforming finite element spaces of Lagrange-type
NASA Technical Reports Server (NTRS)
Cowsar, Lawrence C.
1993-01-01
In this article, we consider the application of three popular domain decomposition methods to Lagrange-type nonconforming finite element discretizations of scalar, self-adjoint, second order elliptic equations. The additive Schwarz method of Dryja and Widlund, the vertex space method of Smith, and the balancing method of Mandel applied to nonconforming elements are shown to converge at a rate no worse than their applications to the standard conforming piecewise linear Galerkin discretization. Essentially, the theory for the nonconforming elements is inherited from the existing theory for the conforming elements with only modest modification by constructing an isomorphism between the nonconforming finite element space and a space of continuous piecewise linear functions.
Domain Decomposition Algorithms for First-Order System Least Squares Methods
NASA Technical Reports Server (NTRS)
Pavarino, Luca F.
1996-01-01
Least squares methods based on first-order systems have been recently proposed and analyzed for second-order elliptic equations and systems. They produce symmetric and positive definite discrete systems by using standard finite element spaces, which are not required to satisfy the inf-sup condition. In this paper, several domain decomposition algorithms for these first-order least squares methods are studied. Some representative overlapping and substructuring algorithms are considered in their additive and multiplicative variants. The theoretical and numerical results obtained show that the classical convergence bounds (on the iteration operator) for standard Galerkin discretizations are also valid for least squares methods.
Adaptive dynamic load-balancing with irregular domain decomposition for particle simulations
NASA Astrophysics Data System (ADS)
Begau, Christoph; Sutmann, Godehard
2015-05-01
We present a flexible and fully adaptive dynamic load-balancing scheme, which is designed for particle simulations of three-dimensional systems with short ranged interactions. The method is based on domain decomposition with non-orthogonal non-convex domains, which are constructed based on a local repartitioning of computational work between neighbouring processors. Domains are dynamically adjusted in a flexible way under the condition that the original topology is not changed, i.e. neighbour relations between domains are retained, which guarantees a fixed communication pattern for each domain during a simulation. Extensions of this scheme are discussed and illustrated with examples, which generalise the communication patterns and do not fully restrict data exchange to direct neighbours. The proposed method relies on a linked cell algorithm, which makes it compatible with existing implementations in particle codes and does not modify the underlying algorithm for calculating the forces between particles. The method has been implemented into the molecular dynamics community code IMD and performance has been measured for various molecular dynamics simulations of systems representing realistic problems from materials science. It is found that the method proves to balance the work between processors in simulations with strongly inhomogeneous and dynamically changing particle distributions, which results in a significant increase of the efficiency of the parallel code compared both to unbalanced simulations and conventional load-balancing strategies.
NASA Astrophysics Data System (ADS)
Abuturab, Muhammad Rafiq
2014-06-01
A new color image security system based on singular value decomposition (SVD) in gyrator transform (GT) domains is proposed. In the encryption process, a color image is decomposed into red, green and blue channels. Each channel is independently modulated by random phase masks and then separately gyrator transformed at different parameters. The three gyrator spectra are joined by multiplication to get one gray ciphertext. The ciphertext is separated into U, S, and V parts by SVD. All the three parts are individually gyrator transformed at different transformation angles. The three encoded information can be assigned to different authorized users for highly secure verification. Only when all the authorized users place the U, S, and V parts in correct multiplication order in the verification system, the correct information can be obtained with all the right keys. In the proposed method, SVD offers one-way asymmetrical decomposition algorithm and it is an optimal matrix decomposition in a least-square sense. The transformation angles of GT provide very sensitive additional keys. The pre-generated keys for red, green and blue channels are served as decryption (private) keys. As all the three encrypted parts are the gray scale ciphertexts with stationary white noise distributions, which have camouflage property to some extent. These advantages enhance the security and robustness. Numerical simulations are presented to support the viability of the proposed verification system.
NASA Astrophysics Data System (ADS)
Rafiq Abuturab, Muhammad
2016-06-01
A new multiple color-image authentication system based on HSI (Hue-Saturation-Intensity) color space and QR decomposition in gyrator domains is proposed. In this scheme, original color images are converted from RGB (Red-Green-Blue) color spaces to HSI color spaces, divided into their H, S, and I components, and then obtained corresponding phase-encoded components. All the phase-encoded H, S, and I components are individually multiplied, and then modulated by random phase functions. The modulated H, S, and I components are convoluted into a single gray image with asymmetric cryptosystem. The resulting image is segregated into Q and R parts by QR decomposition. Finally, they are independently gyrator transformed to get their encoded parts. The encoded Q and R parts should be gathered without missing anyone for decryption. The angles of gyrator transform afford sensitive keys. The protocol based on QR decomposition of encoded matrix and getting back decoded matrix after multiplying matrices Q and R, enhances the security level. The random phase keys, individual phase keys, and asymmetric phase keys provide high robustness to the cryptosystem. Numerical simulation results demonstrate that this scheme is the superior than the existing techniques.
NASA Technical Reports Server (NTRS)
Nguyen, D. T.; Watson, Willie R. (Technical Monitor)
2005-01-01
The overall objectives of this research work are to formulate and validate efficient parallel algorithms, and to efficiently design/implement computer software for solving large-scale acoustic problems, arised from the unified frameworks of the finite element procedures. The adopted parallel Finite Element (FE) Domain Decomposition (DD) procedures should fully take advantages of multiple processing capabilities offered by most modern high performance computing platforms for efficient parallel computation. To achieve this objective. the formulation needs to integrate efficient sparse (and dense) assembly techniques, hybrid (or mixed) direct and iterative equation solvers, proper pre-conditioned strategies, unrolling strategies, and effective processors' communicating schemes. Finally, the numerical performance of the developed parallel finite element procedures will be evaluated by solving series of structural, and acoustic (symmetrical and un-symmetrical) problems (in different computing platforms). Comparisons with existing "commercialized" and/or "public domain" software are also included, whenever possible.
Spatiotemporal Domain Decomposition for Massive Parallel Computation of Space-Time Kernel Density
NASA Astrophysics Data System (ADS)
Hohl, A.; Delmelle, E. M.; Tang, W.
2015-07-01
Accelerated processing capabilities are deemed critical when conducting analysis on spatiotemporal datasets of increasing size, diversity and availability. High-performance parallel computing offers the capacity to solve computationally demanding problems in a limited timeframe, but likewise poses the challenge of preventing processing inefficiency due to workload imbalance between computing resources. Therefore, when designing new algorithms capable of implementing parallel strategies, careful spatiotemporal domain decomposition is necessary to account for heterogeneity in the data. In this study, we perform octtree-based adaptive decomposition of the spatiotemporal domain for parallel computation of space-time kernel density. In order to avoid edge effects near subdomain boundaries, we establish spatiotemporal buffers to include adjacent data-points that are within the spatial and temporal kernel bandwidths. Then, we quantify computational intensity of each subdomain to balance workloads among processors. We illustrate the benefits of our methodology using a space-time epidemiological dataset of Dengue fever, an infectious vector-borne disease that poses a severe threat to communities in tropical climates. Our parallel implementation of kernel density reaches substantial speedup compared to sequential processing, and achieves high levels of workload balance among processors due to great accuracy in quantifying computational intensity. Our approach is portable of other space-time analytical tests.
Mechanical and Assembly Units of Viral Capsids Identified via Quasi-Rigid Domain Decomposition
NASA Astrophysics Data System (ADS)
Polles, Guido; Indelicato, Giuliana; Potestio, Raffaello; Cermelli, Paolo; Twarock, Reidun; Micheletti, Cristian
2014-03-01
Key steps in a viral life-cycle, such as self-assembly of a protective protein container or in some cases also subsequent maturation events, are governed by the interplay of physico-chemical mechanisms involving various spatial and temporal scales. These salient aspects of a viral life cycle are hence well described and rationalised from a mesoscopic perspective. Accordingly, various experimental and computational efforts have been directed towards identifying the fundamental building blocks that are instrumental for the mechanical response, or constitute the assembly units, of a few specific viral shells. Motivated by these earlier studies we introduce and apply a general and efficient computational scheme for identifying the stable domains of a given viral capsid. The method is based on elastic network models and quasi-rigid domain decomposition. It is first applied to a heterogeneous set of well-characterized viruses (CCMV, MS2, STNV, STMV) for which the known mechanical or assembly domains are correctly identified. The validated method is next applied to other viral particles such as L-A, Pariacoto and polyoma viruses, whose fundamental functional domains are still unknown or debated and for which we formulate verifiable predictions.
NASA Astrophysics Data System (ADS)
Grinberg, L.; Karniadakis, G. E.
2010-08-01
We address the failure in scalability of large-scale parallel simulations that are based on (semi-)implicit time-stepping and hence on the solution of linear systems on thousands of processors. We develop a general algorithmic framework based on domain decomposition that removes the scalability limitations and leads to optimal allocation of available computational resources. It is a non-intrusive approach as it does not require modification of existing codes. Specifically, we present here a two-stage domain decomposition method for the Navier-Stokes equations that combines features of discontinuous and continuous Galerkin formulations. At the first stage the domain is subdivided into overlapping patches and within each patch a C0 spectral element discretization (second stage) is employed. Solution within each patch is obtained separately by applying an efficient parallel solver. Proper inter-patch boundary conditions are developed to provide solution continuity, while a Multilevel Communicating Interface (MCI) is developed to provide efficient communication between the non-overlapping groups of processors of each patch. The overall strong scaling of the method depends on the number of patches and on the scalability of the standard solver within each patch. This dual path to scalability provides great flexibility in balancing accuracy with parallel efficiency. The accuracy of the method has been evaluated in solutions of steady and unsteady 3D flow problems including blood flow in the human intracranial arterial tree. Benchmarks on BlueGene/P, CRAY XT5 and Sun Constellation Linux Cluster have demonstrated good performance on up to 96,000 cores, solving up to 8.21B degrees of freedom in unsteady flow problem. The proposed method is general and can be potentially used with other discretization methods or in other applications.
Parallel computation for reservoir thermal simulation: An overlapping domain decomposition approach
NASA Astrophysics Data System (ADS)
Wang, Zhongxiao
2005-11-01
In this dissertation, we are involved in parallel computing for the thermal simulation of multicomponent, multiphase fluid flow in petroleum reservoirs. We report the development and applications of such a simulator. Unlike many efforts made to parallelize locally the solver of a linear equations system which affects the performance the most, this research takes a global parallelization strategy by decomposing the computational domain into smaller subdomains. This dissertation addresses the domain decomposition techniques and, based on the comparison, adopts an overlapping domain decomposition method. This global parallelization method hands over each subdomain to a single processor of the parallel computer to process. Communication is required when handling overlapping regions between subdomains. For this purpose, MPI (message passing interface) is used for data communication and communication control. A physical and mathematical model is introduced for the reservoir thermal simulation. Numerical tests on two sets of industrial data of practical oilfields indicate that this model and the parallel implementation match the history data accurately. Therefore, we expect to use both the model and the parallel code to predict oil production and guide the design, implementation and real-time fine tuning of new well operating schemes. A new adaptive mechanism to synchronize processes on different processors has been introduced, which not only ensures the computational accuracy but also improves the time performance. To accelerate the convergence rate of iterative solution of the large linear equations systems derived from the discretization of governing equations of our physical and mathematical model in space and time, we adopt the ORTHOMIN method in conjunction with an incomplete LU factorization preconditioning technique. Important improvements have been made in both ORTHOMIN method and incomplete LU factorization in order to enhance time performance without affecting
Single-channel color information security system using LU decomposition in gyrator transform domains
NASA Astrophysics Data System (ADS)
Abuturab, Muhammad Rafiq
2014-07-01
A novel single-channel color information security system based on LU decomposition in gyrator transform domains is proposed. The original color image to be encoded is separated into its red, green and blue channels. They are modulated by corresponding random phase functions and then independently Fourier transformed. The transformed images of red and green channels are multiplied and then inverse Fourier transformed. The resulting image is phase- and amplitude truncated to obtain an encrypted image and an asymmetric decryption key, respectively. The encrypted image is multiplied by transformed image of blue channel and then performed LU decomposition. Finally, L and U parts are individually gyrator transformed at different transformation angles, which can be assigned to two different authorized users. The proposed single-channel encryption system is more compact than conventional three-channel encryption systems. Additionally, the ciphertexts are not color images but they are gray images which have obscure properties. The presented LU form is asymmetric. The two transformation angles of GT, three decryption keys for three channels and one asymmetric decryption key significantly improve the security and robustness of the proposed method. The encryption system can be realized digitally or optically. Numerical simulations demonstrate the feasibility and effectiveness of the suggested algorithms.
A balancing domain decomposition method by constraints for advection-diffusion problems
Tu, Xuemin; Li, Jing
2008-12-10
The balancing domain decomposition methods by constraints are extended to solving nonsymmetric, positive definite linear systems resulting from the finite element discretization of advection-diffusion equations. A pre-conditioned GMRES iteration is used to solve a Schur complement system of equations for the subdomain interface variables. In the preconditioning step of each iteration, a partially sub-assembled finite element problem is solved. A convergence rate estimate for the GMRES iteration is established, under the condition that the diameters of subdomains are small enough. It is independent of the number of subdomains and grows only slowly with the subdomain problem size. Numerical experiments for several two-dimensional advection-diffusion problems illustrate the fast convergence of the proposed algorithm.
Domain Decomposition PN Solutions to the 3D Transport Benchmark over a Range in Parameter Space
NASA Astrophysics Data System (ADS)
Van Criekingen, S.
2014-06-01
The objectives of this contribution are twofold. First, the Domain Decomposition (DD) method used in the parafish parallel transport solver is re-interpreted as a Generalized Schwarz Splitting as defined by Tang [SIAM J Sci Stat Comput, vol.13 (2), pp. 573-595, 1992]. Second, parafish provides spherical harmonic (i.e., PN) solutions to the NEA benchmark suite for 3D transport methods and codes over a range in parameter space. To the best of the author's knowledge, these are the first spherical harmonic solutions provided for this demanding benchmark suite. They have been obtained using 512 CPU cores of the JuRoPa machine installed at the Jülich Computing Center (Germany).
NASA Astrophysics Data System (ADS)
Zhao, Tao; Hwang, Feng-Nan; Cai, Xiao-Chuan
2016-07-01
We consider a quintic polynomial eigenvalue problem arising from the finite volume discretization of a quantum dot simulation problem. The problem is solved by the Jacobi-Davidson (JD) algorithm. Our focus is on how to achieve the quadratic convergence of JD in a way that is not only efficient but also scalable when the number of processor cores is large. For this purpose, we develop a projected two-level Schwarz preconditioned JD algorithm that exploits multilevel domain decomposition techniques. The pyramidal quantum dot calculation is carefully studied to illustrate the efficiency of the proposed method. Numerical experiments confirm that the proposed method has a good scalability for problems with hundreds of millions of unknowns on a parallel computer with more than 10,000 processor cores.
Anistratov, Dmitriy Y.; Stehle, Nicholas D.
2012-10-15
A large class of radiative transfer and particle transport problems contain highly diffusive regions. It is possible to reduce computational costs by solving a diffusion problem in diffusive subdomains instead of the transport equation. This enables one to decrease the dimensionality of the transport problem. In this paper we present a methodology for decomposition of a spatial domain of a transport problem into transport and diffusion subregions. We develop methods for solving one-group problems in 1D slab geometry. To identify and locate diffusive regions, we develop metrics for measuring transport effects that are based on the quasidiffusion (Eddington) factor. We present the results of test problems that demonstrate the accuracy of the proposed methodology.
Final Report, DE-FG01-06ER25718 Domain Decomposition and Parallel Computing
Widlund, Olof B.
2015-06-09
The goal of this project is to develop and improve domain decomposition algorithms for a variety of partial differential equations such as those of linear elasticity and electro-magnetics.These iterative methods are designed for massively parallel computing systems and allow the fast solution of the very large systems of algebraic equations that arise in large scale and complicated simulations. A special emphasis is placed on problems arising from Maxwell's equation. The approximate solvers, the preconditioners, are combined with the conjugate gradient method and must always include a solver of a coarse model in order to have a performance which is independent of the number of processors used in the computer simulation. A recent development allows for an adaptive construction of this coarse component of the preconditioner.
FETI Prime Domain Decomposition base Parallel Iterative Solver Library Ver.1.0
Energy Science and Technology Software Center (ESTSC)
2003-09-15
FETI Prime is a library for the iterative solution of linear equations in solid and structural mechanics. The algorithm employs preconditioned conjugate gradients, with a domain decomposition-based preconditioner. The software is written in C++ and is designed for use with massively parallel computers, using MPI. The algorithm is based on the FETI-DP method, with additional capabilities for handling constraint equations, as well as interfacing with the Salinas structural dynamics code and the Finite Element Interfacemore » (FEI) library. Practical Application: FETI Prime is designed for use with finite element-based simulation codes for solid and structural mechanics. The solver uses element matrices, connectivity information, nodal information, and force vectors computed by the host code and provides back the solution to the linear system of equations, to the user specified level of accuracy, The library is compiled with the host code and becomes an integral part of the host code executable.« less
Domain decomposition, multilevel integration, and exponential noise reduction in lattice QCD
NASA Astrophysics Data System (ADS)
Cè, Marco; Giusti, Leonardo; Schaefer, Stefan
2016-05-01
We explore the possibility of computing fermionic correlators on the lattice by combining a domain decomposition with a multilevel integration scheme. The quark propagator is expanded in series of terms with a well-defined hierarchical structure. The higher the order of a term, the (exponentially) smaller its magnitude, the less local is its dependence on the gauge field. Once inserted in a Wick contraction, the gauge-field dependence of the terms in the resulting series can be factorized so that it is suitable for multilevel Monte Carlo integration. We test the strategy in quenched QCD by computing the disconnected correlator of two flavor-diagonal pseudoscalar densities, and a nucleon two-point function. In either case we observe a significant exponential increase of the signal-to-noise ratio.
Wavefront reconstruction for extremely large telescopes via CuRe with domain decomposition.
Rosensteiner, Matthias
2012-11-01
The Cumulative Reconstructor is an accurate, extremely fast reconstruction algorithm for Shack-Hartmann wavefront sensor data. But it has shown an unacceptable high noise propagation for large apertures. Therefore, in this paper we describe a domain decomposition approach to deal with this drawback. We show that this adaptation of the algorithm gives the same reconstruction quality as the original algorithm and leads to a significant improvement with respect to noise propagation. The method is combined with an integral control and compared to the classical matrix vector multiplication algorithm on an end-to-end simulation of a single conjugate adaptive optics system. The reconstruction time is 20n (number of subapertures), and the method is parallelizable. PMID:23201793
Fast non-overlapping Schwarz domain decomposition methods for solving the neutron diffusion equation
NASA Astrophysics Data System (ADS)
Jamelot, Erell; Ciarlet, Patrick
2013-05-01
Studying numerically the steady state of a nuclear core reactor is expensive, in terms of memory storage and computational time. In order to address both requirements, one can use a domain decomposition method, implemented on a parallel computer. We present here such a method for the mixed neutron diffusion equations, discretized with Raviart-Thomas-Nédélec finite elements. This method is based on the Schwarz iterative algorithm with Robin interface conditions to handle communications. We analyse this method from the continuous point of view to the discrete point of view, and we give some numerical results in a realistic highly heterogeneous 3D configuration. Computations are carried out with the MINOS solver of the APOLLO3® neutronics code. APOLLO3 is a registered trademark in France.
Fast non-overlapping Schwarz domain decomposition methods for solving the neutron diffusion equation
Jamelot, Erell; Ciarlet, Patrick
2013-05-15
Studying numerically the steady state of a nuclear core reactor is expensive, in terms of memory storage and computational time. In order to address both requirements, one can use a domain decomposition method, implemented on a parallel computer. We present here such a method for the mixed neutron diffusion equations, discretized with Raviart–Thomas–Nédélec finite elements. This method is based on the Schwarz iterative algorithm with Robin interface conditions to handle communications. We analyse this method from the continuous point of view to the discrete point of view, and we give some numerical results in a realistic highly heterogeneous 3D configuration. Computations are carried out with the MINOS solver of the APOLLO3® neutronics code.
NASA Astrophysics Data System (ADS)
Kuraz, Michal
2016-06-01
Modelling the transport processes in a vadose zone, e.g. modelling contaminant transport or the effect of the soil water regime on changes in soil structure and composition, plays an important role in predicting the reactions of soil biotopes to anthropogenic activity. Water flow is governed by the quasilinear Richards equation. The paper concerns the implementation of a multi-time-step approach for solving a nonlinear Richards equation. When modelling porous media flow with a Richards equation, due to a possible convection dominance and a convergence of a nonlinear solver, a stable finite element approximation requires accurate temporal and spatial integration. The method presented here enables adaptive domain decomposition algorithm together with a multi-time-step treatment of actively changing subdomains.
An iterative finite-element collocation method for parabolic problems using domain decomposition
Curran, M.C.
1992-01-01
Advection-dominated flows occur widely in the transport of groundwater contaminants, the movements of fluids in enhanced oil recovery projects, and many other contexts. In numerical models of such flows, adaptive local grid refinement is a conceptually attractive approach for resolving the sharp fronts or layers that tend to characterize the solutions. However, this approach can be difficult to implement in practice. A domain decomposition method developed by Bramble, Ewing, Pasciak, and Schatz, known as the BEPS method, overcomes many of the difficulties. We demonstrate the applicability of the iterative BEPS ideas to finite-element collocation on trial spaces of piecewise Hermite bicubics. The resulting scheme allows one to refine selected parts of a spatial grid without destroying algebraic efficiencies associated with the original coarse grid. We apply the method to two dimensional time-dependent advection-diffusion problems.
An iterative finite-element collocation method for parabolic problems using domain decomposition
Curran, M.C.
1992-11-01
Advection-dominated flows occur widely in the transport of groundwater contaminants, the movements of fluids in enhanced oil recovery projects, and many other contexts. In numerical models of such flows, adaptive local grid refinement is a conceptually attractive approach for resolving the sharp fronts or layers that tend to characterize the solutions. However, this approach can be difficult to implement in practice. A domain decomposition method developed by Bramble, Ewing, Pasciak, and Schatz, known as the BEPS method, overcomes many of the difficulties. We demonstrate the applicability of the iterative BEPS ideas to finite-element collocation on trial spaces of piecewise Hermite bicubics. The resulting scheme allows one to refine selected parts of a spatial grid without destroying algebraic efficiencies associated with the original coarse grid. We apply the method to two dimensional time-dependent advection-diffusion problems.
Maliassov, S.Y.
1996-12-31
An approach to the construction of an iterative method for solving systems of linear algebraic equations arising from nonconforming finite element discretizations with nonmatching grids for second order elliptic boundary value problems with anisotropic coefficients is considered. The technique suggested is based on decomposition of the original domain into nonoverlapping subdomains. The elliptic problem is presented in the macro-hybrid form with Lagrange multipliers at the interfaces between subdomains. A block diagonal preconditioner is proposed which is spectrally equivalent to the original saddle point matrix and has the optimal order of arithmetical complexity. The preconditioner includes blocks for preconditioning subdomain and interface problems. It is shown that constants of spectral equivalence axe independent of values of coefficients and mesh step size.
Gorshkov, Aleksei V
2012-09-30
The problem of stabilizing a solution of the 2D Navier-Stokes system defined in the exterior of a bounded domain with smooth boundary is investigated. For a given initial velocity field a control on the boundary of the domain must be constructed such that the solution stabilizes to a prescribed vortex solution or trivial solution at the rate of 1/t{sup k}. On the way, related questions are investigated, concerning the behaviour of the spectrum of an operator under a relatively compact perturbation and the existence of attracting invariant manifolds. Bibliography: 21 titles.
Wavelet modelling of the gravity field by domain decomposition methods: an example over Japan
NASA Astrophysics Data System (ADS)
Panet, Isabelle; Kuroishi, Yuki; Holschneider, Matthias
2011-01-01
With the advent of satellite gravity, large gravity data sets of unprecedented quality at low and medium resolution become available. For local, high resolution field modelling, they need to be combined with the surface gravity data. Such models are then used for various applications, from the study of the Earth interior to the determination of oceanic currents. Here we show how to realize such a combination in a flexible way using spherical wavelets and applying a domain decomposition approach. This iterative method, based on the Schwarz algorithms, allows to split a large problem into smaller ones, and avoids the calculation of the entire normal system, which may be huge if high resolution is sought over wide areas. A subdomain is defined as the harmonic space spanned by a subset of the wavelet family. Based on the localization properties of the wavelets in space and frequency, we define hierarchical subdomains of wavelets at different scales. On each scale, blocks of subdomains are defined by using a tailored spatial splitting of the area. The data weighting and regularization are iteratively adjusted for the subdomains, which allows to handle heterogeneity in the data quality or the gravity variations. Different levels of approximations of the subdomains normals are also introduced, corresponding to building local averages of the data at different resolution levels. We first provide the theoretical background on domain decomposition methods. Then, we validate the method with synthetic data, considering two kinds of noise: white noise and coloured noise. We then apply the method to data over Japan, where we combine a satellite-based geopotential model, EIGEN-GL04S, and a local gravity model from a combination of land and marine gravity data and an altimetry-derived marine gravity model. A hybrid spherical harmonics/wavelet model of the geoid is obtained at about 15 km resolution and a corrector grid for the surface model is derived.
Lahmiri, Salim
2016-03-01
Hybridisation of the bi-dimensional empirical mode decomposition (BEMD) with denoising techniques has been proposed in the literature as an effective approach for image denoising. In this Letter, the Student's probability density function is introduced in the computation of the mean envelope of the data during the BEMD sifting process to make it robust to values that are far from the mean. The resulting BEMD is denoted tBEMD. In order to show the effectiveness of the tBEMD, several image denoising techniques in tBEMD domain are employed; namely, fourth order partial differential equation (PDE), linear complex diffusion process (LCDP), non-linear complex diffusion process (NLCDP), and the discrete wavelet transform (DWT). Two biomedical images and a standard digital image were considered for experiments. The original images were corrupted with additive Gaussian noise with three different levels. Based on peak-signal-to-noise ratio, the experimental results show that PDE, LCDP, NLCDP, and DWT all perform better in the tBEMD than in the classical BEMD domain. It is also found that tBEMD is faster than classical BEMD when the noise level is low. When it is high, the computational cost in terms of processing time is similar. The effectiveness of the presented approach makes it promising for clinical applications. PMID:27222723
Global well-posedness for the 2D MHD equations without magnetic diffusion in a strip domain
NASA Astrophysics Data System (ADS)
Ren, Xiaoxia; Xiang, Zhaoyin; Zhang, Zhifei
2016-04-01
We study the initial boundary value problem of two dimensional MHD equations without magnetic diffusion in a strip domain. It was proved that the MHD equations have a unique global strong solution around the equilibrium state ≤ft(0,{{\\mathbf{e}}1}\\right) for both the non-slip boundary condition and Navier slip boundary condition on the velocity.
NASA Astrophysics Data System (ADS)
Yen, Guan-Wei
1994-01-01
A computational method is developed to solve the coupled governing equations of an unsteady flowfield and those of rigid-body dynamics in six degrees-of-freedom (6-DOF). This method is capable of simulating the unsteady flowfields around multiple component configurations with at least one of the components in relative motion with respect to the others. Two of the important phenomena that such analyses can help us to understand are the unsteady aerodynamic interference and the boundary-induced component of such a flowfield. By hybridizing two dynamic domain decomposition techniques, the grid generation task is simplified, the computer memory requirement is reduced, and the governing equations of the rigid-body dynamics are simplified with certain assumptions. Three dimensional, unsteady Navier -Stokes equations are solved on each of the subdomains by a fully-vectorized, finite-volume, upwind-biased, and approximately -factored method. These equations are solved on the composite meshes of hybrid subdomain grids that can move with respect to each other. Hence, the present method combines the advantages of an efficient, geometrically conservative, minimally and automatically dissipative algorithm with the advantages and flexibility of the domain decomposition techniques. Several measures that reduce the numerical error are studied and compared with the exact solution of a moving normal shock in a tube. This solution compares very well with the analytic solution of the isentropic equations. It is concluded, that as a minimum measure, the connectivity of nonconservative overlapped scheme needs to be second-order accurate for spatial and temporal discretizations, as well as for the moving subdomain interpolations. Furthermore, the CFL numbers should be restricted to below unity, if affordable, for flows with high flow gradients. The method is further scrutinized by simulating the flow past a sinusoidally pitching airfoil, and the flow past a sinusoidally pitching and
NASA Astrophysics Data System (ADS)
Steinke, R. C.
2015-12-01
Discretizing 1-D vadose zone simulations in the moisture content domain, such as is done in the Talbot-Ogden method, provides some advantages over discretizing in depth, such as is done in Richards' Equation. These advantages include inherent mass conservation and lower computational cost. However, doing so presents a difficulty for integration with 2-D groundwater interflow simulations. The equations of motion of the bins of discrete moisture content take the depth of the water table as an input. They do not produce it as an output. Finding the correct water table depth so that the groundwater recharge from the 1-D vadose zone simulation mass balances with the lateral flows from the 2-D groundwater interflow simulation was a previously unsolved problem. In this paper we present a net-groundwater-recharge method to solve to this problem and compare it with the source-term method used with Richards' Equation.
NASA Astrophysics Data System (ADS)
Blacic, Tanya M.; Jun, Hyunggu; Rosado, Hayley; Shin, Changsoo
2016-02-01
In seismic oceanography, processed images highlight small temperature changes, but inversion is needed to obtain absolute temperatures. Local search-based full waveform inversion has a lower computational cost than global search but requires accurate starting models. Unfortunately, most marine seismic data have little associated hydrographic data and the band-limited nature of seismic data makes extracting the long wavelength sound speed trend directly from seismic data inherently challenging. Laplace and Laplace-Fourier domain inversion (LDI) can use rudimentary starting models without prior information about the medium. Data are transformed to the Laplace domain, and a smooth sound speed model is extracted by examining the zero and low frequency components of the damped wavefield. We applied LDI to five synthetic data sets based on oceanographic features and recovered smoothed versions of our synthetic models, showing the viability of LDI for creating starting models suitable for more detailed inversions.
NASA Astrophysics Data System (ADS)
Torii, Hajime
2012-12-01
A time-domain computational method for calculating 1D and 2D spectra of resonantly-coupled vibrations in condensed-phase systems is presented. This method simultaneously takes into account the diagonal frequency modulations, the off-diagonal vibrational couplings, and the dynamics of the system, and is applicable to systems of wide interest, e.g., the O-H stretching modes of water and alcohols, and the amide I modes of proteins. The case of the amide I mode of (Ala-d)4 in D2O solution is shown as an example.
Reconstructing photorealistic 3D models from image sequence using domain decomposition method
NASA Astrophysics Data System (ADS)
Xiong, Hanwei; Pan, Ming; Zhang, Xiangwei
2009-11-01
In the fields of industrial design, artistic design and heritage conservation, physical objects are usually digitalized by reverse engineering through some 3D scanning methods. Structured light and photogrammetry are two main methods to acquire 3D information, and both are expensive. Even if these expensive instruments are used, photorealistic 3D models are seldom available. In this paper, a new method to reconstruction photorealistic 3D models using a single camera is proposed. A square plate glued with coded marks is used to place the objects, and a sequence of about 20 images is taken. From the coded marks, the images are calibrated, and a snake algorithm is used to segment object from the background. A rough 3d model is obtained using shape from silhouettes algorithm. The silhouettes are decomposed into a combination of convex curves, which are used to partition the rough 3d model into some convex mesh patches. For each patch, the multi-view photo consistency constraints and smooth regulations are expressed as a finite element formulation, which can be resolved locally, and the information can be exchanged along the patches boundaries. The rough model is deformed into a fine 3d model through such a domain decomposition finite element method. The textures are assigned to each element mesh, and a photorealistic 3D model is got finally. A toy pig is used to verify the algorithm, and the result is exciting.
NASA Astrophysics Data System (ADS)
Rivera, Christian A.; Heniche, Mourad; Glowinski, Roland; Tanguy, Philippe A.
2010-07-01
A parallel approach to solve three-dimensional viscous incompressible fluid flow problems using discontinuous pressure finite elements and a Lagrange multiplier technique is presented. The strategy is based on non-overlapping domain decomposition methods, and Lagrange multipliers are used to enforce continuity at the boundaries between subdomains. The novelty of the work is the coupled approach for solving the velocity-pressure-Lagrange multiplier algebraic system of the discrete Navier-Stokes equations by a distributed memory parallel ILU (0) preconditioned Krylov method. A penalty function on the interface constraints equations is introduced to avoid the failure of the ILU factorization algorithm. To ensure portability of the code, a message based memory distributed model with MPI is employed. The method has been tested over different benchmark cases such as the lid-driven cavity and pipe flow with unstructured tetrahedral grids. It is found that the partition algorithm and the order of the physical variables are central to parallelization performance. A speed-up in the range of 5-13 is obtained with 16 processors. Finally, the algorithm is tested over an industrial case using up to 128 processors. In considering the literature, the obtained speed-ups on distributed and shared memory computers are found very competitive.
NASA Astrophysics Data System (ADS)
Ju, Ning
2016-07-01
New results are obtained for global regularity and long-time behavior of the solutions to the 2D Boussinesq equations for the flow of an incompressible fluid with positive viscosity and zero diffusivity in a smooth bounded domain. Our first result for global boundedness of the solution {(u, θ)} in {D(A)× H^1} improves considerably the main result of the recent article (Hu et al. in J Math Phys 54(8):081507, 2013). Our second result on global boundedness of the solution {(u, θ)} in {V× H^1} for both bounded domain and the whole space {{R}2} is a new one. It has been open and also seems much more challenging than the first result. Global regularity of the solution {(u, θ)} in {D(A)× H2} is also proved.
Alles, E. J.; Zhu, Y.; van Dongen, K. W. A.; McGough, R. J.
2013-01-01
The fast nearfield method, when combined with time-space decomposition, is a rapid and accurate approach for calculating transient nearfield pressures generated by ultrasound transducers. However, the standard time-space decomposition approach is only applicable to certain analytical representations of the temporal transducer surface velocity that, when applied to the fast nearfield method, are expressed as a finite sum of products of separate temporal and spatial terms. To extend time-space decomposition such that accelerated transient field simulations are enabled in the nearfield for an arbitrary transducer surface velocity, a new transient simulation method, frequency domain time-space decomposition (FDTSD), is derived. With this method, the temporal transducer surface velocity is transformed into the frequency domain, and then each complex-valued term is processed separately. Further improvements are achieved by spectral clipping, which reduces the number of terms and the computation time. Trade-offs between speed and accuracy are established for FDTSD calculations, and pressure fields obtained with the FDTSD method for a circular transducer are compared to those obtained with Field II and the impulse response method. The FDTSD approach, when combined with the fast nearfield method and spectral clipping, consistently achieves smaller errors in less time and requires less memory than Field II or the impulse response method. PMID:23160476
NASA Astrophysics Data System (ADS)
Yen, Guan-Wei
A computational method is developed to solve the coupled governing equations of an unsteady flowfield and those of rigid-body dynamics in six degrees-of-freedom (6-DOF). This method is capable of simulating the unsteady flowfields around multiple component configurations with at least one of the components in relative motion with respect to the others. Two of the important phenomena that such analyses can help us to understand are the unsteady aerodynamic interference and the boundary-induced component of such a flowfield. By hybridizing two dynamic domain decomposition techniques, the grid generation task is simplified, the computer memory requirement is reduced, and the governing equations of the rigid-body dynamics are simplified with certain assumptions. Three dimensional, unsteady Navier-Stokes equations are solved on each of the subdomains by a fully-vectorized, finite-volume, upwind-biased, and approximately-factored method. These equations are solved on the composite meshes of hybrid subdomain grids that can move with respect to each other. Hence, the present method combines the advantages of an efficient, geometrically conservative, minimally and automatically dissipative algorithm with the advantages and flexibility of the domain decomposition techniques. Several measures that reduce the numerical error are studied and compared with the exact solution of a moving normal shock in a tube. This solution compares very well with the analytic solution of the isentropic equations. It is concluded, that as a minimum measure, the connectivity of nonconservative overlapped scheme needs to be second-order accurate for spatial and temporal discretizations, as well as for the moving subdomain interpolations. Furthermore, the CFL numbers should be restricted to below unity, if affordable, for flows with high flow gradients. The method is further scrutinized by simulating the flow past a sinusoidally pitching airfoil, and the flow past a sinusoidally pitching and
NASA Astrophysics Data System (ADS)
Wang, Jui-Kai; Sibony, Patrick A.; Kardon, Randy H.; Kupersmith, Mark J.; Garvin, Mona K.
2015-03-01
Recent studies have shown that the Bruch's membrane (BM) and retinal pigment epithelium (RPE), visualized on spectral-domain optical coherence tomography (SD-OCT), is deformed anteriorly towards the vitreous in patients with intracranial hypertension and papilledema. The BM/RPE shape has been quantified using a statistical-shape-model approach; however, to date, the approach has involved the tedious and time-consuming manual placement of landmarks and correspondingly, only the shape (and shape changes) of a limited number of patients has been studied. In this work, we first present a semi-automated approach for the extraction of 20 landmarks along the BM from an optic-nerve-head (ONH) centered OCT slice from each patient. In the approach, after the manual placement of the two Bruch's membrane opening (BMO) points, the remaining 18 landmarks are automatically determined using a graph-based segmentation approach. We apply the approach to the OCT scans of 116 patients (at baseline) enrolled in the Idiopathic Intracranial Hypertension Treatment Trial and generate a statistical shape model using principal components analysis. Using the resulting shape model, the coefficient (shape measure) corresponding to the second principal component (eigenvector) for each set of landmarks indicates the degree of the BM/RPE is oriented away from the vitreous. Using a subset of 20 patients, we compare the shape measure computed using this semi-automated approach with the resulting shape measure when (1) all landmarks are specified manually (Experiment I); and (2) a different expert specifies the two BMO points (Experiment II). In each case, a correlation coefficient >= 0.99 is obtained.
Lou, Zheng; Jin, Jian-Ming . E-mail: j-jin1@uiuc.edu
2007-03-01
A novel dual-field time-domain finite-element domain-decomposition method is presented for an efficient and broadband numerical simulation of electromagnetic properties of large finite arrays. Instead of treating the entire array as a single computation domain, the method considers each array element as a smaller subdomain and computes both the electric and magnetic fields inside each subdomain. Adjacent subdomains are related to each other by the equivalent surface currents on the subdomain interfaces in an explicit manner. Furthermore, the method exploits the identical geometry of the array elements and further reduces the memory requirement and CPU time. The proposed method is highly efficient for the simulation of large finite arrays. Numerical stability and computational performance of the method are discussed. Several radiation examples are presented to demonstrate the accuracy and efficiency of the method.
Zhang, Keni; Moridis, G.J.; Wu, Y.-S.; Pruess, K.
2008-07-01
Simulation of the system behavior of hydrate-bearing geologic media involves solving fully coupled mass- and heat-balance equations. In this study, we develop a domain decomposition approach for large-scale gas hydrate simulations with coarse-granularity parallel computation. This approach partitions a simulation domain into small subdomains. The full model domain, consisting of discrete subdomains, is still simulated simultaneously by using multiple processes/processors. Each processor is dedicated to following tasks of the partitioned subdomain: updating thermophysical properties, assembling mass- and energy-balance equations, solving linear equation systems, and performing various other local computations. The linearized equation systems are solved in parallel with a parallel linear solver, using an efficient interprocess communication scheme. This new domain decomposition approach has been implemented into the TOUGH+HYDRATE code and has demonstrated excellent speedup and good scalability. In this paper, we will demonstrate applications for the new approach in simulating field-scale models for gas production from gas-hydrate deposits.
NASA Astrophysics Data System (ADS)
Meert, Alan; Vikram, Vinu; Bernardi, Mariangela
2015-02-01
We present a catalogue of 2D, point spread function-corrected de Vacouleurs, Sérsic, de Vacouleurs+Exponential, and Sérsic+Exponential fits of ˜7 × 105 spectroscopically selected galaxies drawn from the Sloan Digital Sky Survey (SDSS) Data Release 7. Fits are performed for the SDSS r band utilizing the fitting routine GALFIT and analysis pipeline PYMORPH. We compare these fits to prior catalogues. Fits are analysed using a physically motivated flagging system. The flags suggest that more than 90 per cent of two-component fits can be used for analysis. We show that the fits follow the expected behaviour for early and late galaxy types. The catalogues provide a robust set of structural and photometric parameters for future galaxy studies. We show that some biases remain in the measurements, e.g. the presence of bars significantly affect the bulge measurements although the bulge ellipticity may be used to separate barred and non-barred galaxies, and about 15 per cent of bulges of two-component fits are also affected by resolution. The catalogues are available in electronic format. We also provide an interface for generating postage stamp images of the 2D model and residual as well as the 1D profile. These images can be generated for a user-uploaded list of galaxies on demand.
NASA Astrophysics Data System (ADS)
Imamura, N.; Schultz, A.
2015-12-01
Recently, a full waveform time domain solution has been developed for the magnetotelluric (MT) and controlled-source electromagnetic (CSEM) methods. The ultimate goal of this approach is to obtain a computationally tractable direct waveform joint inversion for source fields and earth conductivity structure in three and four dimensions. This is desirable on several grounds, including the improved spatial resolving power expected from use of a multitude of source illuminations of non-zero wavenumber, the ability to operate in areas of high levels of source signal spatial complexity and non-stationarity, etc. This goal would not be obtainable if one were to adopt the finite difference time-domain (FDTD) approach for the forward problem. This is particularly true for the case of MT surveys, since an enormous number of degrees of freedom are required to represent the observed MT waveforms across the large frequency bandwidth. It means that for FDTD simulation, the smallest time steps should be finer than that required to represent the highest frequency, while the number of time steps should also cover the lowest frequency. This leads to a linear system that is computationally burdensome to solve. We have implemented our code that addresses this situation through the use of a fictitious wave domain method and GPUs to speed up the computation time. We also substantially reduce the size of the linear systems by applying concepts from successive cascade decimation, through quasi-equivalent time domain decomposition. By combining these refinements, we have made good progress toward implementing the core of a full waveform joint source field/earth conductivity inverse modeling method. From results, we found the use of previous generation of CPU/GPU speeds computations by an order of magnitude over a parallel CPU only approach. In part, this arises from the use of the quasi-equivalent time domain decomposition, which shrinks the size of the linear system dramatically.
NASA Astrophysics Data System (ADS)
Innocenti, M.; Beck, A.; Lapenta, G.; Markidis, S.
2012-12-01
The kinetic simulation of intrinsically multi scale processes such as magnetic reconnection events with realistic mass ratios is a daunting task for explicit Particle In Cell (PIC) codes, which require to use resolutions of the order of the electron Debye length even when simulating dramatically bigger domains. As an example, a simulation of reconnection in the magnetotail, with domain sizes of the order of 20 di x 10 di (˜ 7.2 106 m x 3.6 106 m, with di being the ion skin depth) and a resolution of λD,e= 687 m, with λD,e the electron Debye length, requires the astounding number of 10500 x 5240 cells. Higher grid spacings can be used if the simulation is performed with an implicit PIC code, which substitutes a much less strict accuracy constraint to the stability constraint of explicit PIC codes. The same reconnection problem as before can be simulated, with an implicit PIC code resolving the scale of interest of de /2 instead of the electron Debye length (de is the electron skin depth), with the much more manageable number of 1920 x 958 cells. However, an even smaller number of cells can be used if, instead of using the same, high resolution on the entire domain, the domain to simulate is divided into subdomains each resolved with a grid spacing related to the physical scale of interest in the specific subdomain. In the case of reconnection, the division which immediately springs to mind is between electron diffusion region, ion diffusion region and outer region, where resolutions respectively of the order of fractions of the electron skin depth, of the ion skin depth and bigger can be used. We present here a new Multi Level Multi Domain (MLMD) Implicit Moment Method (IMM) Particle In Cell (PIC) code, Parsek2D-MLMD, able to perform simulations of magnetic reconnection where the expensive high resolutions are used only when needed, while the rest of the domain is simulated with grid spacings chosen according to the local scales of interest. The major difference
Energy Science and Technology Software Center (ESTSC)
2005-07-01
Aniso2d is a two-dimensional seismic forward modeling code. The earth is parameterized by an X-Z plane in which the seismic properties Can have monoclinic with x-z plane symmetry. The program uses a user define time-domain wavelet to produce synthetic seismograms anrwhere within the two-dimensional media.
Arsov, Zoran; Schara, Milan; Strancar, Janez
2002-07-01
Using EPR spectroscopy a typical lateral domain structure was detected in the membranes of spin-labeled bovine erythrocyte ghosts. The spectral parameters were determined by decomposing the EPR spectrum into three spectral components and tuned by a hybrid-evolutionary-optimization method. In our experiments the lateral domain structure and its properties were influenced by the variation in the temperature and by the addition of n-butanol. The specific responses of the particular domain types were detected. For the most-ordered domain type a break was seen in the temperature dependence of its order parameter, while the order parameters of the two less-ordered domain types exhibited a continuous decrease. Below the break-point temperature the alcohol-induced membrane fluidity variation is mainly a consequence of the change in the proportions of the least- and the most-ordered domain type and not the change of the domain-type ordering or dynamics (with n-butanol concentration). On the other hand, the fluidity variation above the break-point temperature arises from both types of changes. Interestingly, the proportion of the domain type that has its order parameter between that of the least- and the most-ordered domain type remains almost constant with concentration as well as with temperature, which implies its stability. Such characterization of the lateral membrane domain structure could be beneficial when considering the lipid-protein interactions, because it can be assumed that the activity of the membrane-bound enzyme depends on the properties of the particular domain type. PMID:12202132
NASA Astrophysics Data System (ADS)
Subber, Waad; Matouš, Karel
2016-02-01
Large-scale practical engineering problems featuring localized phenomena often benefit from local control of mesh and time resolutions to efficiently capture the spatial and temporal scales of interest. To this end, we propose an asynchronous space-time algorithm based on a domain decomposition method for structural dynamics problems on non-matching meshes. The three-field algorithm is based on the dual-primal like domain decomposition approach utilizing the localized Lagrange multipliers along the space and time common-refinement-based interface. The proposed algorithm is parallel in nature and well suited for a heterogeneous computing environment. Moreover, two-levels of parallelism are embedded in this novel scheme. For linear dynamical problems, the algorithm is unconditionally stable, shows an optimal order of convergence with respect to space and time discretizations as well as ensures conservation of mass, momentum and energy across the non-matching grid interfaces. The method of manufactured solutions is used to verify the implementation, and an engineering application is considered, where a sandwich plate is impacted by a projectile.
Subber, Waad Sarkar, Abhijit
2014-01-15
Recent advances in high performance computing systems and sensing technologies motivate computational simulations with extremely high resolution models with capabilities to quantify uncertainties for credible numerical predictions. A two-level domain decomposition method is reported in this investigation to devise a linear solver for the large-scale system in the Galerkin spectral stochastic finite element method (SSFEM). In particular, a two-level scalable preconditioner is introduced in order to iteratively solve the large-scale linear system in the intrusive SSFEM using an iterative substructuring based domain decomposition solver. The implementation of the algorithm involves solving a local problem on each subdomain that constructs the local part of the preconditioner and a coarse problem that propagates information globally among the subdomains. The numerical and parallel scalabilities of the two-level preconditioner are contrasted with the previously developed one-level preconditioner for two-dimensional flow through porous media and elasticity problems with spatially varying non-Gaussian material properties. A distributed implementation of the parallel algorithm is carried out using MPI and PETSc parallel libraries. The scalabilities of the algorithm are investigated in a Linux cluster.
Ozgun, Ozlem Mittra, Raj; Kuzuoglu, Mustafa
2009-04-01
In this paper, we introduce a parallelized version of a novel, non-iterative domain decomposition algorithm, called Characteristic Basis Finite Element Method (CBFEM-MPI), for efficient solution of large-scale electromagnetic scattering problems, by utilizing a set of specially defined characteristic basis functions (CBFs). This approach is based on the decomposition of the computational domain into a number of non-overlapping subdomains wherein the CBFs are generated by employing a novel procedure, which differs from all those that have been used in the past. Clearly, the CBFs are obtained by calculating the fields radiated by a finite number of dipole-type sources, which are placed hypothetically along the boundary of the conducting object. The major advantages of the proposed technique are twofold: (i) it provides a substantial reduction in the matrix size, and thus, makes use of direct solvers efficiently and (ii) it enables the utilization of parallel processing techniques that considerably decrease the overall computation time. We illustrate the application of the proposed approach via several 3D electromagnetic scattering problems.
NASA Astrophysics Data System (ADS)
Spyropoulos, John T.
This thesis extends earlier research in numerical analysis and computational fluid dynamics (CFD) to obtain a novel finite element method for the transient, 3-D, incompressible Navier-Stokes equations, along with efficient, parallelizable algorithms to carry out an implementation of the method in such a fashion as to be useful in mainstream industrial settings. This new finite element procedure employs alternating-direction operator splittings to model problems of increasing complexity in a step-by-step and natural manner. The scheme employs a characteristic-Galerkin method for the numerical treatment of the nonlinear advection operator. Non-overlapping domain decomposition schemes are employed for the solution of linear Stokes-type subproblems and for the matching of the inviscid and viscous solutions in different subdomains. These problems are solved by Bramble-Pasciak-Schatz wirebasket domain decomposition methods in a stabilized mixed finite element method formulation. The scheme is coupled to an existing grid generator code that provides globally unstructured, but locally structured grids, within each subdomain. Numerical results obtained include incompressible viscous flows over a backward facing steps at various Reynolds numbers and show very good to excellent agreement with experiments as well as other published numerical results.
Brosch, Tom; Tam, Roger
2015-01-01
Deep learning has traditionally been computationally expensive, and advances in training methods have been the prerequisite for improving its efficiency in order to expand its application to a variety of image classification problems. In this letter, we address the problem of efficient training of convolutional deep belief networks by learning the weights in the frequency domain, which eliminates the time-consuming calculation of convolutions. An essential consideration in the design of the algorithm is to minimize the number of transformations to and from frequency space. We have evaluated the running time improvements using two standard benchmark data sets, showing a speed-up of up to 8 times on 2D images and up to 200 times on 3D volumes. Our training algorithm makes training of convolutional deep belief networks on 3D medical images with a resolution of up to 128×128×128 voxels practical, which opens new directions for using deep learning for medical image analysis. PMID:25380341
NASA Astrophysics Data System (ADS)
Pauli, Jutta; van Rossum, Barth; Förster, Hans; de Groot, Huub J. M.; Oschkinat, Hartmut
2000-04-01
Future structural investigations of proteins by solid-state CPMAS NMR will rely on uniformly labeled protein samples showing spectra with an excellent resolution. NMR samples of the solid α-spectrin SH3 domain were generated in four different ways, and their 13C CPMAS spectra were compared. The spectrum of a [u-13C, 15N]-labeled sample generated by precipitation shows very narrow 13C signals and resolved scalar carbon-carbon couplings. Linewidths of 16-19 Hz were found for the three alanine Cβ signals of a selectively labeled [70% 3-13C]alanine-enriched SH3 sample. The signal pattern of the isoleucine, of all prolines, valines, alanines, and serines, and of three of the four threonines were identified in 2D 13C-13C RFDR spectra of the [u-13C,15N]-labeled SH3 sample. A comparison of the 13C chemical shifts of the found signal patterns with the 13C assignment obtained in solution shows an intriguing match.
NASA Astrophysics Data System (ADS)
Vera, N. C.; GMMC
2013-05-01
In this paper we present the results of macrohybrid mixed Darcian flow in porous media in a general three-dimensional domain. The global problem is solved as a set of local subproblems which are posed using a domain decomposition method. Unknown fields of local problems, velocity and pressure are approximated using mixed finite elements. For this application, a general three-dimensional domain is considered which is discretized using tetrahedra. The discrete domain is decomposed into subdomains and reformulated the original problem as a set of subproblems, communicated through their interfaces. To solve this set of subproblems, we use finite element mixed and parallel computing. The parallelization of a problem using this methodology can, in principle, to fully exploit a computer equipment and also provides results in less time, two very important elements in modeling. Referencias G.Alduncin and N.Vera-Guzmán Parallel proximal-point algorithms for mixed _nite element models of _ow in the subsurface, Commun. Numer. Meth. Engng 2004; 20:83-104 (DOI: 10.1002/cnm.647) Z. Chen, G.Huan and Y. Ma Computational Methods for Multiphase Flows in Porous Media, SIAM, Society for Industrial and Applied Mathematics, Philadelphia, 2006. A. Quarteroni and A. Valli, Numerical Approximation of Partial Differential Equations, Springer-Verlag, Berlin, 1994. Brezzi F, Fortin M. Mixed and Hybrid Finite Element Methods. Springer: New York, 1991.
Yang, L. H.; Brooks III, E. D.; Belak, J.
1992-01-01
A molecular dynamics algorithm for performing large-scale simulations using the Parallel C Preprocessor (PCP) programming paradigm on the BBN TC2000, a massively parallel computer, is discussed. The algorithm uses a linked-cell data structure to obtain the near neighbors of each atom as time evoles. Each processor is assigned to a geometric domain containing many subcells and the storage for that domain is private to the processor. Within this scheme, the interdomain (i.e., interprocessor) communication is minimized.
NASA Technical Reports Server (NTRS)
Bailey, R. T.; Shih, T. I.-P.; Nguyen, H. L.; Roelke, R. J.
1990-01-01
An efficient computer program, called GRID2D/3D, was developed to generate single and composite grid systems within geometrically complex two- and three-dimensional (2- and 3-D) spatial domains that can deform with time. GRID2D/3D generates single grid systems by using algebraic grid generation methods based on transfinite interpolation in which the distribution of grid points within the spatial domain is controlled by stretching functions. All single grid systems generated by GRID2D/3D can have grid lines that are continuous and differentiable everywhere up to the second-order. Also, grid lines can intersect boundaries of the spatial domain orthogonally. GRID2D/3D generates composite grid systems by patching together two or more single grid systems. The patching can be discontinuous or continuous. For continuous composite grid systems, the grid lines are continuous and differentiable everywhere up to the second-order except at interfaces where different single grid systems meet. At interfaces where different single grid systems meet, the grid lines are only differentiable up to the first-order. For 2-D spatial domains, the boundary curves are described by using either cubic or tension spline interpolation. For 3-D spatial domains, the boundary surfaces are described by using either linear Coon's interpolation, bi-hyperbolic spline interpolation, or a new technique referred to as 3-D bi-directional Hermite interpolation. Since grid systems generated by algebraic methods can have grid lines that overlap one another, GRID2D/3D contains a graphics package for evaluating the grid systems generated. With the graphics package, the user can generate grid systems in an interactive manner with the grid generation part of GRID2D/3D. GRID2D/3D is written in FORTRAN 77 and can be run on any IBM PC, XT, or AT compatible computer. In order to use GRID2D/3D on workstations or mainframe computers, some minor modifications must be made in the graphics part of the program; no
NASA Astrophysics Data System (ADS)
Sourbier, F.; Operto, S.; Virieux, J.
2006-12-01
We present a distributed-memory parallel algorithm for 2D visco-acoustic full-waveform inversion of wide-angle seismic data. Our code is written in fortran90 and use MPI for parallelism. The algorithm was applied to real wide-angle data set recorded by 100 OBSs with a 1-km spacing in the eastern-Nankai trough (Japan) to image the deep structure of the subduction zone. Full-waveform inversion is applied sequentially to discrete frequencies by proceeding from the low to the high frequencies. The inverse problem is solved with a classic gradient method. Full-waveform modeling is performed with a frequency-domain finite-difference method. In the frequency-domain, solving the wave equation requires resolution of a large unsymmetric system of linear equations. We use the massively parallel direct solver MUMPS (http://www.enseeiht.fr/irit/apo/MUMPS) for distributed-memory computer to solve this system. The MUMPS solver is based on a multifrontal method for the parallel factorization. The MUMPS algorithm is subdivided in 3 main steps: a symbolic analysis step that performs re-ordering of the matrix coefficients to minimize the fill-in of the matrix during the subsequent factorization and an estimation of the assembly tree of the matrix. Second, the factorization is performed with dynamic scheduling to accomodate numerical pivoting and provides the LU factors distributed over all the processors. Third, the resolution is performed for multiple sources. To compute the gradient of the cost function, 2 simulations per shot are required (one to compute the forward wavefield and one to back-propagate residuals). The multi-source resolutions can be performed in parallel with MUMPS. In the end, each processor stores in core a sub-domain of all the solutions. These distributed solutions can be exploited to compute in parallel the gradient of the cost function. Since the gradient of the cost function is a weighted stack of the shot and residual solutions of MUMPS, each processor
NASA Astrophysics Data System (ADS)
Dolean, Victorita; Gander, Martin J.; Lanteri, Stephane; Lee, Jin-Fa; Peng, Zhen
2015-01-01
The time-harmonic Maxwell equations describe the propagation of electromagnetic waves and are therefore fundamental for the simulation of many modern devices we have become used to in everyday life. The numerical solution of these equations is hampered by two fundamental problems: first, in the high frequency regime, very fine meshes need to be used in order to avoid the pollution effect well known for the Helmholtz equation, and second the large scale systems obtained from the vector valued equations in three spatial dimensions need to be solved by iterative methods, since direct factorizations are not feasible any more at that scale. As for the Helmholtz equation, classical iterative methods applied to discretized Maxwell equations have severe convergence problems. We explain in this paper a family of domain decomposition methods based on well chosen transmission conditions. We show that all transmission conditions proposed so far in the literature, both for the first and second order formulation of Maxwell's equations, can be written and optimized in the common framework of optimized Schwarz methods, independently of the first or second order formulation one uses, and the performance of the corresponding algorithms is identical. We use a decomposition into transverse electric and transverse magnetic fields to describe these algorithms, which greatly simplifies the convergence analysis of the methods. We illustrate the performance of our algorithms with large scale numerical simulations.
Texture analysis by fractal descriptors over the wavelet domain using a best basis decomposition
NASA Astrophysics Data System (ADS)
Florindo, J. B.; Bruno, O. M.
2016-02-01
This work proposes the development and study of a novel set of fractal descriptors for texture analysis. These descriptors are obtained by exploring the fractal-like relation among the coefficients and magnitudes of a particular type of wavelet decomposition, to know, the best basis selection. The proposed method is tested in the classification of three sets of textures from the literature: Brodatz, Vistex and USPTex. The method is also applied to a challenging real-world problem, which is the identification of species of plants from the Brazilian flora. The results are compared with other classical and state-of-the-art texture descriptors and demonstrate the efficiency of the proposed technique in this task.
Lazarov, R; Pasciak, J; Jones, J
2002-02-01
Construction, analysis and numerical testing of efficient solution techniques for solving elliptic PDEs that allow for parallel implementation have been the focus of the research. A number of discretization and solution methods for solving second order elliptic problems that include mortar and penalty approximations and domain decomposition methods for finite elements and finite volumes have been investigated and analyzed. Techniques for parallel domain decomposition algorithms in the framework of PETC and HYPRE have been studied and tested. Hierarchical parallel grid refinement and adaptive solution methods have been implemented and tested on various model problems. A parallel code implementing the mortar method with algebraically constructed multiplier spaces was developed.
Greenman, G M; O'Brien, M J; Procassini, R J; Joy, K I
2009-03-09
Two enhancements to the combinatorial geometry (CG) particle tracker in the Mercury Monte Carlo transport code are presented. The first enhancement is a hybrid particle tracker wherein a mesh region is embedded within a CG region. This method permits efficient calculations of problems with contain both large-scale heterogeneous and homogeneous regions. The second enhancement relates to the addition of parallelism within the CG tracker via spatial domain decomposition. This permits calculations of problems with a large degree of geometric complexity, which are not possible through particle parallelism alone. In this method, the cells are decomposed across processors and a particles is communicated to an adjacent processor when it tracks to an interprocessor boundary. Applications that demonstrate the efficacy of these new methods are presented.
Dimarco, Giacomo; Mieussens, Luc; Rispoli, Vittorio
2014-10-01
In this work we present an efficient strategy to deal with plasma physics simulations in which localized departures from thermodynamical equilibrium are present. The method is based on the introduction of intermediate regions which allows smooth transitions between kinetic and fluid zones. In this paper we extend Domain Decomposition techniques, obtained through dynamic coupling and buffer zones, to the study of plasmas and, moreover, we combine them with Asymptotic Preserving and Asymptotically Accurate strategies for the time integration. We use a hybrid scheme in which both kinetic and fluid descriptions are considered and coupled together while the kinetic model is solved by asymptotic preserving and accurate methods, in order to guarantee high efficiency and accuracy in all regimes. The numerical scheme is validated and its performances are analyzed by numerical simulations.
An efficient, robust, domain-decomposition algorithm for particle Monte Carlo
Brunner, Thomas A. Brantley, Patrick S.
2009-06-01
A previously described algorithm [T.A. Brunner, T.J. Urbatsch, T.M. Evans, N.A. Gentile, Comparison of four parallel algorithms for domain decomposed implicit Monte Carlo, Journal of Computational Physics 212 (2) (2006) 527-539] for doing domain decomposed particle Monte Carlo calculations in the context of thermal radiation transport has been improved. It has been extended to support cases where the number of particles in a time step are unknown at the beginning of the time step. This situation arises when various physical processes, such as neutron transport, can generate additional particles during the time step, or when particle splitting is used for variance reduction. Additionally, several race conditions that existed in the previous algorithm and could cause code hangs have been fixed. This new algorithm is believed to be robust against all race conditions. The parallel scalability of the new algorithm remains excellent.
He, Jingjing; Zhou, Yibin; Guan, Xuefei; Zhang, Wei; Zhang, Weifang; Liu, Yongming
2016-01-01
Structural health monitoring has been studied by a number of researchers as well as various industries to keep up with the increasing demand for preventive maintenance routines. This work presents a novel method for reconstruct prompt, informed strain/stress responses at the hot spots of the structures based on strain measurements at remote locations. The structural responses measured from usage monitoring system at available locations are decomposed into modal responses using empirical mode decomposition. Transformation equations based on finite element modeling are derived to extrapolate the modal responses from the measured locations to critical locations where direct sensor measurements are not available. Then, two numerical examples (a two-span beam and a 19956-degree of freedom simplified airfoil) are used to demonstrate the overall reconstruction method. Finally, the present work investigates the effectiveness and accuracy of the method through a set of experiments conducted on an aluminium alloy cantilever beam commonly used in air vehicle and spacecraft. The experiments collect the vibration strain signals of the beam via optical fiber sensors. Reconstruction results are compared with theoretical solutions and a detailed error analysis is also provided. PMID:27537889
Domain decomposition based iterative methods for nonlinear elliptic finite element problems
Cai, X.C.
1994-12-31
The class of overlapping Schwarz algorithms has been extensively studied for linear elliptic finite element problems. In this presentation, the author considers the solution of systems of nonlinear algebraic equations arising from the finite element discretization of some nonlinear elliptic equations. Several overlapping Schwarz algorithms, including the additive and multiplicative versions, with inexact Newton acceleration will be discussed. The author shows that the convergence rate of the Newton`s method is independent of the mesh size used in the finite element discretization, and also independent of the number of subdomains into which the original domain in decomposed. Numerical examples will be presented.
Domain decomposition methods for systems of conservation laws: Spectral collocation approximations
NASA Technical Reports Server (NTRS)
Quarteroni, Alfio
1989-01-01
Hyperbolic systems of conversation laws are considered which are discretized in space by spectral collocation methods and advanced in time by finite difference schemes. At any time-level a domain deposition method based on an iteration by subdomain procedure was introduced yielding at each step a sequence of independent subproblems (one for each subdomain) that can be solved simultaneously. The method is set for a general nonlinear problem in several space variables. The convergence analysis, however, is carried out only for a linear one-dimensional system with continuous solutions. A precise form of the error reduction factor at each iteration is derived. Although the method is applied here to the case of spectral collocation approximation only, the idea is fairly general and can be used in a different context as well. For instance, its application to space discretization by finite differences is straight forward.
NASA Astrophysics Data System (ADS)
Pütz, Martin; Nielaba, Peter
2015-03-01
We present a numerical method for simulations of spinodal decomposition of liquid-vapor systems. The results are in excellent agreement with theoretical predictions for all expected time regimes from the initial growth of "homophase fluctuations" up to the inertial hydrodynamics regime. The numerical approach follows a modern formulation of the smoothed particle hydrodynamics method with a van der Waals equation of state and thermal conduction. The dynamics and thermal evolution of instantaneously temperature-quenched systems are investigated. Therefore, we introduce a simple scaling thermostat that allows thermal fluctuations at a constant predicted mean temperature. We find that the initial stage spinodal decomposition is strongly affected by the temperature field. The separated phases react on density changes with a change in temperature. Although, the thermal conduction acts very slowly, thermal deviations are eventually compensated. The domain growth in the late stage of demixing is found to be rather unaffected by thermal fluctuations. We observe a transition from the Lifshitz-Slyozov growth rate with 1 /3 exponent to the inertial hydrodynamics regime with a rate of 2 /3 , only excepted from simulations near the critical point where the liquid droplets are observed to nucleate directly in a spherical shape. The transition between the growth regimes is found to occur earlier for higher initial temperatures. We explain this time dependency with the phase interfaces that become more diffuse and overlap with approaching the critical point. A prolonging behavior of the demixing process is observed and also expected to depend on temperature. It is further found that the observations can excellently explain the growth behavior for pure nonisothermal simulations that are performed without thermostat.
Pütz, Martin; Nielaba, Peter
2015-03-01
We present a numerical method for simulations of spinodal decomposition of liquid-vapor systems. The results are in excellent agreement with theoretical predictions for all expected time regimes from the initial growth of "homophase fluctuations" up to the inertial hydrodynamics regime. The numerical approach follows a modern formulation of the smoothed particle hydrodynamics method with a van der Waals equation of state and thermal conduction. The dynamics and thermal evolution of instantaneously temperature-quenched systems are investigated. Therefore, we introduce a simple scaling thermostat that allows thermal fluctuations at a constant predicted mean temperature. We find that the initial stage spinodal decomposition is strongly affected by the temperature field. The separated phases react on density changes with a change in temperature. Although, the thermal conduction acts very slowly, thermal deviations are eventually compensated. The domain growth in the late stage of demixing is found to be rather unaffected by thermal fluctuations. We observe a transition from the Lifshitz-Slyozov growth rate with 1/3 exponent to the inertial hydrodynamics regime with a rate of 2/3, only excepted from simulations near the critical point where the liquid droplets are observed to nucleate directly in a spherical shape. The transition between the growth regimes is found to occur earlier for higher initial temperatures. We explain this time dependency with the phase interfaces that become more diffuse and overlap with approaching the critical point. A prolonging behavior of the demixing process is observed and also expected to depend on temperature. It is further found that the observations can excellently explain the growth behavior for pure nonisothermal simulations that are performed without thermostat. PMID:25871106
NASA Astrophysics Data System (ADS)
Laizet, Sylvain; Lamballais, Eric; Vassilicos, J. Christos
2015-11-01
Incompact3d is a high-order flow solver dedicated to Direct and Large Eddy Simulations (DNS/LES) using High Performance Computing (HPC) systems which isdevoted to turbulent flows at the interface between academic research and upstream industrial R&D. It is originating from the University of Poitiers (France) and was developed there as well as, more recently, in the Turbulence, Mixing and Flow Control Group at Imperial College London (UK). This high-order flow solver can reconcile accuracy, efficiency, versatility and scalability using a simple Cartesian mesh and up to one million computational cores. The three key ingredients of this successful cocktail to tackle turbulence on HPC systemswill be given in this talkfollowed by various applications such as fractal-generated turbulence, gravity currents in an open basin, impinging jets on a heated plate and a micro-jet device to control a turbulent jet.
NASA Astrophysics Data System (ADS)
Mocek, Lukas; Kozubek, Tomas
2011-09-01
The paper deals with the numerical solution of elliptic boundary value problems for 2D linear elasticity using the fictitious domain method in combination with the discrete Fourier transform and the FETI domain decomposition. We briefly mention the theoretical background of these methods, introduce resulting solvers, and demonstrate their efficiency on model benchmarks.
NASA Astrophysics Data System (ADS)
Mieloszyk, M.; Opoka, S.; Ostachowicz, W.
2015-07-01
This paper presents an application of Fibre Bragg Grating (FBG) sensors for Structural Health Monitoring (SHM) of offshore wind energy support structure model. The analysed structure is a tripod equipped with 16 FBG sensors. From a wide variety of Operational Modal Analysis (OMA) methods Frequency Domain Decomposition (FDD) technique is used in this paper under assumption that the input loading is similar to a white noise excitation. The FDD method can be applied using different sets of sensors, i.e. the one which contains all FBG sensors and the other set of sensors localised only on a particular tripod's leg. The cases considered during investigation were as follows: damaged and undamaged scenarios, different support conditions. The damage was simulated as an dismantled flange on an upper brace in one of the tripod legs. First the model was fixed to an antishaker table and investigated in the air under impulse excitations. Next the tripod was submerged into water basin in order to check the quality of the measurement set-up in different environmental condition. In this case the model was excited by regular waves.
NASA Technical Reports Server (NTRS)
Ku, Hwar-Ching; Ramaswamy, Bala
1993-01-01
The new multigrid (or adaptive) pseudospectral element method was carried out for the solution of incompressible flow in terms of primitive variable formulation. The desired features of the proposed method include the following: (1) the ability to treat complex geometry; (2) high resolution adapted in the interesting areas; (3) requires minimal working space; and (4) effective in a multiprocessing environment. The approach for flow problems, complex geometry or not, is to first divide the computational domain into a number of fine-grid and coarse-grid subdomains with the inter-overlapping area. Next, it is necessary to implement the Schwarz alternating procedure (SAP) to exchange the data among subdomains, where the coarse-grid correction is used to remove the high frequency error that occurs when the data interpolation from the fine-grid subdomain to the coarse-grid subdomain is conducted. The strategy behind the coarse-grid correction is to adopt the operator of the divergence of the velocity field, which intrinsically links the pressure equation, into this process. The solution of each subdomain can be efficiently solved by the direct (or iterative) eigenfunction expansion technique with the least storage requirement, i.e. O(N(exp 3)) in 3-D and O(N(exp 2)) in 2-D. Numerical results of both driven cavity and jet flow will be presented in the paper to account for the versatility of the proposed method.
NASA Astrophysics Data System (ADS)
Zhao, Wei; Niu, Tianye; Xing, Lei; Xie, Yaoqin; Xiong, Guanglei; Elmore, Kimberly; Zhu, Jun; Wang, Luyao; Min, James K.
2016-02-01
Increased noise is a general concern for dual-energy material decomposition. Here, we develop an image-domain material decomposition algorithm for dual-energy CT (DECT) by incorporating an edge-preserving filter into the Local HighlY constrained backPRojection reconstruction (HYPR-LR) framework. With effective use of the non-local mean, the proposed algorithm, which is referred to as HYPR-NLM, reduces the noise in dual-energy decomposition while preserving the accuracy of quantitative measurement and spatial resolution of the material-specific dual-energy images. We demonstrate the noise reduction and resolution preservation of the algorithm with an iodine concentrate numerical phantom by comparing the HYPR-NLM algorithm to the direct matrix inversion, HYPR-LR and iterative image-domain material decomposition (Iter-DECT). We also show the superior performance of the HYPR-NLM over the existing methods by using two sets of cardiac perfusing imaging data. The DECT material decomposition comparison study shows that all four algorithms yield acceptable quantitative measurements of iodine concentrate. Direct matrix inversion yields the highest noise level, followed by HYPR-LR and Iter-DECT. HYPR-NLM in an iterative formulation significantly reduces image noise and the image noise is comparable to or even lower than that generated using Iter-DECT. For the HYPR-NLM method, there are marginal edge effects in the difference image, suggesting the high-frequency details are well preserved. In addition, when the search window size increases from 11× 11 to 19× 19 , there are no significant changes or marginal edge effects in the HYPR-NLM difference images. The reference drawn from the comparison study includes: (1) HYPR-NLM significantly reduces the DECT material decomposition noise while preserving quantitative measurements and high-frequency edge information, and (2) HYPR-NLM is robust with respect to parameter selection.
NASA Astrophysics Data System (ADS)
Schultz, A.
2010-12-01
describe our ongoing efforts to achieve massive parallelization on a novel hybrid GPU testbed machine currently configured with 12 Intel Westmere Xeon CPU cores (or 24 parallel computational threads) with 96 GB DDR3 system memory, 4 GPU subsystems which in aggregate contain 960 NVidia Tesla GPU cores with 16 GB dedicated DDR3 GPU memory, and a second interleved bank of 4 GPU subsystems containing in aggregate 1792 NVidia Fermi GPU cores with 12 GB dedicated DDR5 GPU memory. We are applying domain decomposition methods to a modified version of Weiss' (2001) 3D frequency domain full physics EM finite difference code, an open source GPL licensed f90 code available for download from www.OpenEM.org. This will be the core of a new hybrid 3D inversion that parallelizes frequencies across CPUs and individual forward solutions across GPUs. We describe progress made in modifying the code to use direct solvers in GPU cores dedicated to each small subdomain, iteratively improving the solution by matching adjacent subdomain boundary solutions, rather than iterative Krylov space sparse solvers as currently applied to the whole domain.
NASA Astrophysics Data System (ADS)
Liu, F.; Zhuang, P.; Turner, I.; Anh, V.; Burrage, K.
2015-07-01
A FitzHugh-Nagumo monodomain model has been used to describe the propagation of the electrical potential in heterogeneous cardiac tissue. In this paper, we consider a two-dimensional fractional FitzHugh-Nagumo monodomain model on an irregular domain. The model consists of a coupled Riesz space fractional nonlinear reaction-diffusion model and an ordinary differential equation, describing the ionic fluxes as a function of the membrane potential. Second, we use a decoupling technique and focus on solving the Riesz space fractional nonlinear reaction-diffusion model. A novel spatially second-order accurate semi-implicit alternating direction method (SIADM) for this model on an approximate irregular domain is proposed. Third, stability and convergence of the SIADM are proved. Finally, some numerical examples are given to support our theoretical analysis and these numerical techniques are employed to simulate a two-dimensional fractional FitzHugh-Nagumo model on both an approximate circular and an approximate irregular domain.
NASA Technical Reports Server (NTRS)
Farhat, Charbel; Rixen, Daniel
1996-01-01
We present an optimal preconditioning algorithm that is equally applicable to the dual (FETI) and primal (Balancing) Schur complement domain decomposition methods, and which successfully addresses the problems of subdomain heterogeneities including the effects of large jumps of coefficients. The proposed preconditioner is derived from energy principles and embeds a new coarsening operator that propagates the error globally and accelerates convergence. The resulting iterative solver is illustrated with the solution of highly heterogeneous elasticity problems.
NASA Astrophysics Data System (ADS)
Gao, Zhensen; Dai, Bo; Wang, Xu; Kataoka, Nobuyuki; Wada, Naoya
2010-12-01
We propose and experimentally demonstrate a reconfigurable two-dimensional (temporal-spectral) time domain spectral phase encoding (SPE) scheme for coherent optical code-division-multiple-access (OCDMA) application. The time-domain SPE scheme is robust to wavelength drift of the light source and is very flexible and compatible with the fiber optical system. In the proposed scheme, the ultra-short optical pulse is stretched by dispersive device and the SPE is done in time domain using high speed phase modulator. A Fiber Bragg Gratings array is used for generating the two-dimensional wavelength hopping pattern while the high speed phase modulator is used for generating the spectral phase pattern. The proposed scheme can enable simultaneous generation of the time domain spectral phase encoding and DPSK data modulation using only a single phase modulator. In the experiment, the two-dimensional SPE codes have been generated and modulated with 2.5-Gb/s DPSK data using a single phase modulator. Transmission of the 2.5-Gb/s DPSK data over 49km fiber with BER<10-9 has been demonstrated successfully. The proposed scheme exhibits the potential to simplify the architecture and improve the security of the OCDMA system.
Tensor representation of color images and fast 2D quaternion discrete Fourier transform
NASA Astrophysics Data System (ADS)
Grigoryan, Artyom M.; Agaian, Sos S.
2015-03-01
In this paper, a general, efficient, split algorithm to compute the two-dimensional quaternion discrete Fourier transform (2-D QDFT), by using the special partitioning in the frequency domain, is introduced. The partition determines an effective transformation, or color image representation in the form of 1-D quaternion signals which allow for splitting the N × M-point 2-D QDFT into a set of 1-D QDFTs. Comparative estimates revealing the efficiency of the proposed algorithms with respect to the known ones are given. In particular, a proposed method of calculating the 2r × 2r -point 2-D QDFT uses 18N2 less multiplications than the well-known column-row method and method of calculation based on the symplectic decomposition. The proposed algorithm is simple to apply and design, which makes it very practical in color image processing in the frequency domain.
NASA Astrophysics Data System (ADS)
Zheng, Xiang; Yang, Chao; Cai, Xiao-Chuan; Keyes, David
2015-03-01
We present a numerical algorithm for simulating the spinodal decomposition described by the three dimensional Cahn-Hilliard-Cook (CHC) equation, which is a fourth-order stochastic partial differential equation with a noise term. The equation is discretized in space and time based on a fully implicit, cell-centered finite difference scheme, with an adaptive time-stepping strategy designed to accelerate the progress to equilibrium. At each time step, a parallel Newton-Krylov-Schwarz algorithm is used to solve the nonlinear system. We discuss various numerical and computational challenges associated with the method. The numerical scheme is validated by a comparison with an explicit scheme of high accuracy (and unreasonably high cost). We present steady state solutions of the CHC equation in two and three dimensions. The effect of the thermal fluctuation on the spinodal decomposition process is studied. We show that the existence of the thermal fluctuation accelerates the spinodal decomposition process and that the final steady morphology is sensitive to the stochastic noise. We also show the evolution of the energies and statistical moments. In terms of the parallel performance, it is found that the implicit domain decomposition approach scales well on supercomputers with a large number of processors.
Zheng, Xiang; Yang, Chao; Cai, Xiao-Chuan; Keyes, David
2015-03-15
We present a numerical algorithm for simulating the spinodal decomposition described by the three dimensional Cahn–Hilliard–Cook (CHC) equation, which is a fourth-order stochastic partial differential equation with a noise term. The equation is discretized in space and time based on a fully implicit, cell-centered finite difference scheme, with an adaptive time-stepping strategy designed to accelerate the progress to equilibrium. At each time step, a parallel Newton–Krylov–Schwarz algorithm is used to solve the nonlinear system. We discuss various numerical and computational challenges associated with the method. The numerical scheme is validated by a comparison with an explicit scheme of high accuracy (and unreasonably high cost). We present steady state solutions of the CHC equation in two and three dimensions. The effect of the thermal fluctuation on the spinodal decomposition process is studied. We show that the existence of the thermal fluctuation accelerates the spinodal decomposition process and that the final steady morphology is sensitive to the stochastic noise. We also show the evolution of the energies and statistical moments. In terms of the parallel performance, it is found that the implicit domain decomposition approach scales well on supercomputers with a large number of processors.
Dong, Xue; Niu, Tianye; Zhu, Lei
2014-05-15
Purpose: Dual-energy CT (DECT) is being increasingly used for its capability of material decomposition and energy-selective imaging. A generic problem of DECT, however, is that the decomposition process is unstable in the sense that the relative magnitude of decomposed signals is reduced due to signal cancellation while the image noise is accumulating from the two CT images of independent scans. Direct image decomposition, therefore, leads to severe degradation of signal-to-noise ratio on the resultant images. Existing noise suppression techniques are typically implemented in DECT with the procedures of reconstruction and decomposition performed independently, which do not explore the statistical properties of decomposed images during the reconstruction for noise reduction. In this work, the authors propose an iterative approach that combines the reconstruction and the signal decomposition procedures to minimize the DECT image noise without noticeable loss of resolution. Methods: The proposed algorithm is formulated as an optimization problem, which balances the data fidelity and total variation of decomposed images in one framework, and the decomposition step is carried out iteratively together with reconstruction. The noise in the CT images from the proposed algorithm becomes well correlated even though the noise of the raw projections is independent on the two CT scans. Due to this feature, the proposed algorithm avoids noise accumulation during the decomposition process. The authors evaluate the method performance on noise suppression and spatial resolution using phantom studies and compare the algorithm with conventional denoising approaches as well as combined iterative reconstruction methods with different forms of regularization. Results: On the Catphan©600 phantom, the proposed method outperforms the existing denoising methods on preserving spatial resolution at the same level of noise suppression, i.e., a reduction of noise standard deviation by one
Sakhratov, Yu. A.; Svistov, L. E.; Kuhns, P. L.; Zhou, H. D.; Reyes, A. P.
2014-11-15
We have carried out {sup 63,65}Cu NMR spectra measurements in a magnetic field up to about 15.5 T on a single crystal of the multiferroic triangular-lattice antiferromagnet CuCrO{sub 2}. The measurements were performed for perpendicular and parallel orientations of the magnetic field with respect to the c axis of the crystal, and the detailed angle dependence of the spectra on the magnetic field direction in the ab plane was studied. The shape of the spectra can be well described in the model of spiral spin structure proposed by recent neutron diffraction experiments. When the field is rotated perpendicular to the crystal c axis, we observed, directly for the first time, a remarkable reorientation of the spin plane simultaneous with rotation of the incommensurate wavevector, by quantitatively deducing the conversion of the energetically less favorable domain to a more favorable one. At high enough fields parallel to the c axis, the data are consistent with either a field-induced commensurate spiral magnetic structure or an incommensurate spiral magnetic structure with a disorder in the c direction, suggesting that high fields may have influence on interplanar ordering.
NASA Astrophysics Data System (ADS)
Bruno, Oscar P.; Cubillos, Max
2016-02-01
This paper introduces alternating-direction implicit (ADI) solvers of higher order of time-accuracy (orders two to six) for the compressible Navier-Stokes equations in two- and three-dimensional curvilinear domains. The higher-order accuracy in time results from 1) An application of the backward differentiation formulae time-stepping algorithm (BDF) in conjunction with 2) A BDF-like extrapolation technique for certain components of the nonlinear terms (which makes use of nonlinear solves unnecessary), as well as 3) A novel application of the Douglas-Gunn splitting (which greatly facilitates handling of boundary conditions while preserving higher-order accuracy in time). As suggested by our theoretical analysis of the algorithms for a variety of special cases, an extensive set of numerical experiments clearly indicate that all of the BDF-based ADI algorithms proposed in this paper are "quasi-unconditionally stable" in the following sense: each algorithm is stable for all couples (h , Δt)of spatial and temporal mesh sizes in a problem-dependent rectangular neighborhood of the form (0 ,Mh) × (0 ,Mt). In other words, for each fixed value of Δt below a certain threshold, the Navier-Stokes solvers presented in this paper are stable for arbitrarily small spatial mesh-sizes. The second-order formulation has further been rigorously shown to be unconditionally stable for linear hyperbolic and parabolic equations in two-dimensional space. Although implicit ADI solvers for the Navier-Stokes equations with nominal second-order of temporal accuracy have been proposed in the past, the algorithms presented in this paper are the first ADI-based Navier-Stokes solvers for which second-order or better accuracy has been verified in practice under non-trivial (non-periodic) boundary conditions.
NASA Astrophysics Data System (ADS)
Wu, Yuqi; Cai, Xiao-Chuan
2014-02-01
Due to the rapid advancement of supercomputing hardware, there is a growing interest in parallel algorithms for modeling the full three-dimensional interaction between the blood flow and the arterial wall. In [4], Barker and Cai developed a parallel framework for solving fluid-structure interaction problems in two dimensions. In this paper, we extend the idea to three dimensions. We introduce and study a parallel scalable domain decomposition method for solving nonlinear monolithically coupled systems arising from the discretization of the coupled system in an arbitrary Lagrangian-Eulerian framework with a fully implicit stabilized finite element method. The investigation focuses on the robustness and parallel scalability of the Newton-Krylov algorithm preconditioned with an overlapping additive Schwarz method. We validate the proposed approach and report the parallel performance for some patient-specific pulmonary artery problems. The algorithm is shown to be scalable with a large number of processors and for problems with millions of unknowns.
NASA Astrophysics Data System (ADS)
Lorin, E.; Yang, X.; Antoine, X.
2016-06-01
The paper is devoted to develop efficient domain decomposition methods for the linear Schrödinger equation beyond the semiclassical regime, which does not carry a small enough rescaled Planck constant for asymptotic methods (e.g. geometric optics) to produce a good accuracy, but which is too computationally expensive if direct methods (e.g. finite difference) are applied. This belongs to the category of computing middle-frequency wave propagation, where neither asymptotic nor direct methods can be directly used with both efficiency and accuracy. Motivated by recent works of the authors on absorbing boundary conditions (Antoine et al. (2014) [13] and Yang and Zhang (2014) [43]), we introduce Semiclassical Schwarz Waveform Relaxation methods (SSWR), which are seamless integrations of semiclassical approximation to Schwarz Waveform Relaxation methods. Two versions are proposed respectively based on Herman-Kluk propagation and geometric optics, and we prove the convergence and provide numerical evidence of efficiency and accuracy of these methods.
Aagaard, B.T.; Knepley, M.G.; Williams, C.A.
2013-01-01
We employ a domain decomposition approach with Lagrange multipliers to implement fault slip in a finite-element code, PyLith, for use in both quasi-static and dynamic crustal deformation applications. This integrated approach to solving both quasi-static and dynamic simulations leverages common finite-element data structures and implementations of various boundary conditions, discretization schemes, and bulk and fault rheologies. We have developed a custom preconditioner for the Lagrange multiplier portion of the system of equations that provides excellent scalability with problem size compared to conventional additive Schwarz methods. We demonstrate application of this approach using benchmarks for both quasi-static viscoelastic deformation and dynamic spontaneous rupture propagation that verify the numerical implementation in PyLith.
Greg Flach, Frank Smith
2011-12-31
Mesh2d is a Fortran90 program designed to generate two-dimensional structured grids of the form [x(i),y(i,j)] where [x,y] are grid coordinates identified by indices (i,j). The x(i) coordinates alone can be used to specify a one-dimensional grid. Because the x-coordinates vary only with the i index, a two-dimensional grid is composed in part of straight vertical lines. However, the nominally horizontal y(i,j0) coordinates along index i are permitted to undulate or otherwise vary. Mesh2d also assigns an integer material type to each grid cell, mtyp(i,j), in a user-specified manner. The complete grid is specified through three separate input files defining the x(i), y(i,j), and mtyp(i,j) variations.
Energy Science and Technology Software Center (ESTSC)
2011-12-31
Mesh2d is a Fortran90 program designed to generate two-dimensional structured grids of the form [x(i),y(i,j)] where [x,y] are grid coordinates identified by indices (i,j). The x(i) coordinates alone can be used to specify a one-dimensional grid. Because the x-coordinates vary only with the i index, a two-dimensional grid is composed in part of straight vertical lines. However, the nominally horizontal y(i,j0) coordinates along index i are permitted to undulate or otherwise vary. Mesh2d also assignsmore » an integer material type to each grid cell, mtyp(i,j), in a user-specified manner. The complete grid is specified through three separate input files defining the x(i), y(i,j), and mtyp(i,j) variations.« less
NASA Astrophysics Data System (ADS)
Lotsch, Bettina V.
2015-07-01
Graphene's legacy has become an integral part of today's condensed matter science and has equipped a whole generation of scientists with an armory of concepts and techniques that open up new perspectives for the postgraphene area. In particular, the judicious combination of 2D building blocks into vertical heterostructures has recently been identified as a promising route to rationally engineer complex multilayer systems and artificial solids with intriguing properties. The present review highlights recent developments in the rapidly emerging field of 2D nanoarchitectonics from a materials chemistry perspective, with a focus on the types of heterostructures available, their assembly strategies, and their emerging properties. This overview is intended to bridge the gap between two major—yet largely disjunct—developments in 2D heterostructures, which are firmly rooted in solid-state chemistry or physics. Although the underlying types of heterostructures differ with respect to their dimensions, layer alignment, and interfacial quality, there is common ground, and future synergies between the various assembly strategies are to be expected.
NASA Astrophysics Data System (ADS)
Chang, Der-Chen; Markina, Irina; Wang, Wei
2016-09-01
The k-Cauchy-Fueter operator D0(k) on one dimensional quaternionic space H is the Euclidean version of spin k / 2 massless field operator on the Minkowski space in physics. The k-Cauchy-Fueter equation for k ≥ 2 is overdetermined and its compatibility condition is given by the k-Cauchy-Fueter complex. In quaternionic analysis, these complexes play the role of Dolbeault complex in several complex variables. We prove that a natural boundary value problem associated to this complex is regular. Then by using the theory of regular boundary value problems, we show the Hodge-type orthogonal decomposition, and the fact that the non-homogeneous k-Cauchy-Fueter equation D0(k) u = f on a smooth domain Ω in H is solvable if and only if f satisfies the compatibility condition and is orthogonal to the set ℋ(k)1 (Ω) of Hodge-type elements. This set is isomorphic to the first cohomology group of the k-Cauchy-Fueter complex over Ω, which is finite dimensional, while the second cohomology group is always trivial.
A scalable 2-D parallel sparse solver
Kothari, S.C.; Mitra, S.
1995-12-01
Scalability beyond a small number of processors, typically 32 or less, is known to be a problem for existing parallel general sparse (PGS) direct solvers. This paper presents a parallel general sparse PGS direct solver for general sparse linear systems on distributed memory machines. The algorithm is based on the well-known sequential sparse algorithm Y12M. To achieve efficient parallelization, a 2-D scattered decomposition of the sparse matrix is used. The proposed algorithm is more scalable than existing parallel sparse direct solvers. Its scalability is evaluated on a 256 processor nCUBE2s machine using Boeing/Harwell benchmark matrices.
A new inversion method for (T2, D) 2D NMR logging and fluid typing
NASA Astrophysics Data System (ADS)
Tan, Maojin; Zou, Youlong; Zhou, Cancan
2013-02-01
One-dimensional nuclear magnetic resonance (1D NMR) logging technology has some significant limitations in fluid typing. However, not only can two-dimensional nuclear magnetic resonance (2D NMR) provide some accurate porosity parameters, but it can also identify fluids more accurately than 1D NMR. In this paper, based on the relaxation mechanism of (T2, D) 2D NMR in a gradient magnetic field, a hybrid inversion method that combines least-squares-based QR decomposition (LSQR) and truncated singular value decomposition (TSVD) is examined in the 2D NMR inversion of various fluid models. The forward modeling and inversion tests are performed in detail with different acquisition parameters, such as magnetic field gradients (G) and echo spacing (TE) groups. The simulated results are discussed and described in detail, the influence of the above-mentioned observation parameters on the inversion accuracy is investigated and analyzed, and the observation parameters in multi-TE activation are optimized. Furthermore, the hybrid inversion can be applied to quantitatively determine the fluid saturation. To study the effects of noise level on the hybrid method and inversion results, the numerical simulation experiments are performed using different signal-to-noise-ratios (SNRs), and the effect of different SNRs on fluid typing using three fluid models are discussed and analyzed in detail.
Fast and Memory-Efficient Topological Denoising of 2D and 3D Scalar Fields.
Günther, David; Jacobson, Alec; Reininghaus, Jan; Seidel, Hans-Peter; Sorkine-Hornung, Olga; Weinkauf, Tino
2014-12-01
Data acquisition, numerical inaccuracies, and sampling often introduce noise in measurements and simulations. Removing this noise is often necessary for efficient analysis and visualization of this data, yet many denoising techniques change the minima and maxima of a scalar field. For example, the extrema can appear or disappear, spatially move, and change their value. This can lead to wrong interpretations of the data, e.g., when the maximum temperature over an area is falsely reported being a few degrees cooler because the denoising method is unaware of these features. Recently, a topological denoising technique based on a global energy optimization was proposed, which allows the topology-controlled denoising of 2D scalar fields. While this method preserves the minima and maxima, it is constrained by the size of the data. We extend this work to large 2D data and medium-sized 3D data by introducing a novel domain decomposition approach. It allows processing small patches of the domain independently while still avoiding the introduction of new critical points. Furthermore, we propose an iterative refinement of the solution, which decreases the optimization energy compared to the previous approach and therefore gives smoother results that are closer to the input. We illustrate our technique on synthetic and real-world 2D and 3D data sets that highlight potential applications. PMID:26356972
Layer Engineering of 2D Semiconductor Junctions.
He, Yongmin; Sobhani, Ali; Lei, Sidong; Zhang, Zhuhua; Gong, Yongji; Jin, Zehua; Zhou, Wu; Yang, Yingchao; Zhang, Yuan; Wang, Xifan; Yakobson, Boris; Vajtai, Robert; Halas, Naomi J; Li, Bo; Xie, Erqing; Ajayan, Pulickel
2016-07-01
A new concept for junction fabrication by connecting multiple regions with varying layer thicknesses, based on the thickness dependence, is demonstrated. This type of junction is only possible in super-thin-layered 2D materials, and exhibits similar characteristics as p-n junctions. Rectification and photovoltaic effects are observed in chemically homogeneous MoSe2 junctions between domains of different thicknesses. PMID:27136275
Application of 2D Non-Graphene Materials and 2D Oxide Nanostructures for Biosensing Technology
Shavanova, Kateryna; Bakakina, Yulia; Burkova, Inna; Shtepliuk, Ivan; Viter, Roman; Ubelis, Arnolds; Beni, Valerio; Starodub, Nickolaj; Yakimova, Rositsa; Khranovskyy, Volodymyr
2016-01-01
The discovery of graphene and its unique properties has inspired researchers to try to invent other two-dimensional (2D) materials. After considerable research effort, a distinct “beyond graphene” domain has been established, comprising the library of non-graphene 2D materials. It is significant that some 2D non-graphene materials possess solid advantages over their predecessor, such as having a direct band gap, and therefore are highly promising for a number of applications. These applications are not limited to nano- and opto-electronics, but have a strong potential in biosensing technologies, as one example. However, since most of the 2D non-graphene materials have been newly discovered, most of the research efforts are concentrated on material synthesis and the investigation of the properties of the material. Applications of 2D non-graphene materials are still at the embryonic stage, and the integration of 2D non-graphene materials into devices is scarcely reported. However, in recent years, numerous reports have blossomed about 2D material-based biosensors, evidencing the growing potential of 2D non-graphene materials for biosensing applications. This review highlights the recent progress in research on the potential of using 2D non-graphene materials and similar oxide nanostructures for different types of biosensors (optical and electrochemical). A wide range of biological targets, such as glucose, dopamine, cortisol, DNA, IgG, bisphenol, ascorbic acid, cytochrome and estradiol, has been reported to be successfully detected by biosensors with transducers made of 2D non-graphene materials. PMID:26861346
Application of 2D Non-Graphene Materials and 2D Oxide Nanostructures for Biosensing Technology.
Shavanova, Kateryna; Bakakina, Yulia; Burkova, Inna; Shtepliuk, Ivan; Viter, Roman; Ubelis, Arnolds; Beni, Valerio; Starodub, Nickolaj; Yakimova, Rositsa; Khranovskyy, Volodymyr
2016-01-01
The discovery of graphene and its unique properties has inspired researchers to try to invent other two-dimensional (2D) materials. After considerable research effort, a distinct "beyond graphene" domain has been established, comprising the library of non-graphene 2D materials. It is significant that some 2D non-graphene materials possess solid advantages over their predecessor, such as having a direct band gap, and therefore are highly promising for a number of applications. These applications are not limited to nano- and opto-electronics, but have a strong potential in biosensing technologies, as one example. However, since most of the 2D non-graphene materials have been newly discovered, most of the research efforts are concentrated on material synthesis and the investigation of the properties of the material. Applications of 2D non-graphene materials are still at the embryonic stage, and the integration of 2D non-graphene materials into devices is scarcely reported. However, in recent years, numerous reports have blossomed about 2D material-based biosensors, evidencing the growing potential of 2D non-graphene materials for biosensing applications. This review highlights the recent progress in research on the potential of using 2D non-graphene materials and similar oxide nanostructures for different types of biosensors (optical and electrochemical). A wide range of biological targets, such as glucose, dopamine, cortisol, DNA, IgG, bisphenol, ascorbic acid, cytochrome and estradiol, has been reported to be successfully detected by biosensors with transducers made of 2D non-graphene materials. PMID:26861346
Proper Orthogonal Decomposition of Flow-Field in Non-Stationary Geometry
NASA Astrophysics Data System (ADS)
Troshin, Victor; Seifert, Avraham; Sidilkover, David; Tadmor, Gilead
2015-11-01
This work presents a proper orthogonal decomposition (POD) methodology for a flow field in a domain with moving boundaries. A relatively simple volume preserving mapping which transforms a deforming to stationary domain is described. This mapping was created by combining a transfinite interpolation and volume adjustment algorithm. The algorithm is based on iterative solution of the Laplace equation with respect to the displacement potential of the grid points. The transformed domain is suitable for proper orthogonal decomposition procedure. The presented mapping can be applied to a wide variety of flow problems which contain single or in some cases multiple deforming boundaries. Currently, this method is presented for 2D geometries, however, it can be expanded to 3D cases. This approach can assist in creation of low order models for complex aero-elastic systems which to date could not be analysed by existing POD approaches. Finally, the method is demonstrated on CFD results of pitching and plunging ellipse in still fluid.
NASA Astrophysics Data System (ADS)
Wang, Jin; Ma, Jianyong; Zhou, Changhe
2014-11-01
A 3×3 high divergent 2D-grating with period of 3.842μm at wavelength of 850nm under normal incidence is designed and fabricated in this paper. This high divergent 2D-grating is designed by the vector theory. The Rigorous Coupled Wave Analysis (RCWA) in association with the simulated annealing (SA) is adopted to calculate and optimize this 2D-grating.The properties of this grating are also investigated by the RCWA. The diffraction angles are more than 10 degrees in the whole wavelength band, which are bigger than the traditional 2D-grating. In addition, the small period of grating increases the difficulties of fabrication. So we fabricate the 2D-gratings by direct laser writing (DLW) instead of traditional manufacturing method. Then the method of ICP etching is used to obtain the high divergent 2D-grating.
Realistic and efficient 2D crack simulation
NASA Astrophysics Data System (ADS)
Yadegar, Jacob; Liu, Xiaoqing; Singh, Abhishek
2010-04-01
Although numerical algorithms for 2D crack simulation have been studied in Modeling and Simulation (M&S) and computer graphics for decades, realism and computational efficiency are still major challenges. In this paper, we introduce a high-fidelity, scalable, adaptive and efficient/runtime 2D crack/fracture simulation system by applying the mathematically elegant Peano-Cesaro triangular meshing/remeshing technique to model the generation of shards/fragments. The recursive fractal sweep associated with the Peano-Cesaro triangulation provides efficient local multi-resolution refinement to any level-of-detail. The generated binary decomposition tree also provides efficient neighbor retrieval mechanism used for mesh element splitting and merging with minimal memory requirements essential for realistic 2D fragment formation. Upon load impact/contact/penetration, a number of factors including impact angle, impact energy, and material properties are all taken into account to produce the criteria of crack initialization, propagation, and termination leading to realistic fractal-like rubble/fragments formation. The aforementioned parameters are used as variables of probabilistic models of cracks/shards formation, making the proposed solution highly adaptive by allowing machine learning mechanisms learn the optimal values for the variables/parameters based on prior benchmark data generated by off-line physics based simulation solutions that produce accurate fractures/shards though at highly non-real time paste. Crack/fracture simulation has been conducted on various load impacts with different initial locations at various impulse scales. The simulation results demonstrate that the proposed system has the capability to realistically and efficiently simulate 2D crack phenomena (such as window shattering and shards generation) with diverse potentials in military and civil M&S applications such as training and mission planning.
NASA Astrophysics Data System (ADS)
Martin, R.; Gonzalez Ortiz, A.
In the industry as well as in the geophysical community, multiphase flows are mod- elled using a finite volume approach and a multicorrector algorithm in time in order to determine implicitly the pressures, velocities and volume fractions for each phase. Pressures, and velocities are generally determined at mid-half mesh step from each other following the staggered grid approach. This ensures stability and prevents os- cillations in pressure. It allows to treat almost all the Reynolds number ranges for all speeds and viscosities. The disadvantages appear when we want to treat more complex geometries or if a generalized curvilinear formulation of the conservation equations is considered. Too many interpolations have to be done and accuracy is then lost. In order to overcome these problems, we use here a similar algorithm in time and a Rhie and Chow interpolation (1983) of the collocated variables and essentially the velocities at the interface. The Rhie and Chow interpolation of the velocities at the finite volume interfaces allows to have no oscillatons of the pressure without checkerboard effects and to stabilize all the algorithm. In a first predictor step, fluxes at the interfaces of the finite volumes are then computed using 2nd and 3rd order shock capturing schemes of MUSCL/TVD or Van Leer type, and the orthogonal stress components are treated implicitly while cross viscous/diffusion terms are treated explicitly. A pentadiagonal system in 2D or a septadiagonal in 3D must be solve but here we have chosen to solve 3 tridiagonal linear systems (the so called Alternate Direction Implicit algorithm), one in each spatial direction, to reduce the cost of computation. Then a multi-correction of interpolated velocities, pressures and volumic fractions of each phase are done in the cartesian frame or the deformed local curvilinear coordinate system till convergence and mass conservation. At the end the energy conservation equations are solved. In all this process the
Toward an Efficient Icing CFD Process Using an Interactive Software Toolkit: Smagglce 2D
NASA Technical Reports Server (NTRS)
Vickerman, Mary B.; Choo, Yung K.; Schilling, Herbert W.; Baez, Marivell; Braun, Donald C.; Cotton, Barbara J.
2001-01-01
Two-dimensional CID analysis for iced airfoils can be a labor-intensive task. The software toolkit SmaggIce 2D is being developed to help streamline the CID process and provide the unique features needed for icing. When complete, it will include a combination of partially automated and fully interactive tools for all aspects of the tasks leading up to the flow analysis: geometry preparation, domain decomposition. block boundary demoralization. gridding, and linking with a flow solver. It also includes tools to perform ice shape characterization, an important aid in determining the relationship between ice characteristics and their effects on aerodynamic performance. Completed tools, work-in-progress, and planned features of the software toolkit are presented here.
Stochastic Inversion of 2D Magnetotelluric Data
Energy Science and Technology Software Center (ESTSC)
2010-07-01
The algorithm is developed to invert 2D magnetotelluric (MT) data based on sharp boundary parametrization using a Bayesian framework. Within the algorithm, we consider the locations and the resistivity of regions formed by the interfaces are as unknowns. We use a parallel, adaptive finite-element algorithm to forward simulate frequency-domain MT responses of 2D conductivity structure. Those unknown parameters are spatially correlated and are described by a geostatistical model. The joint posterior probability distribution function ismore » explored by Markov Chain Monte Carlo (MCMC) sampling methods. The developed stochastic model is effective for estimating the interface locations and resistivity. Most importantly, it provides details uncertainty information on each unknown parameter. Hardware requirements: PC, Supercomputer, Multi-platform, Workstation; Software requirements C and Fortan; Operation Systems/version is Linux/Unix or Windows« less
Stochastic Inversion of 2D Magnetotelluric Data
Chen, Jinsong
2010-07-01
The algorithm is developed to invert 2D magnetotelluric (MT) data based on sharp boundary parametrization using a Bayesian framework. Within the algorithm, we consider the locations and the resistivity of regions formed by the interfaces are as unknowns. We use a parallel, adaptive finite-element algorithm to forward simulate frequency-domain MT responses of 2D conductivity structure. Those unknown parameters are spatially correlated and are described by a geostatistical model. The joint posterior probability distribution function is explored by Markov Chain Monte Carlo (MCMC) sampling methods. The developed stochastic model is effective for estimating the interface locations and resistivity. Most importantly, it provides details uncertainty information on each unknown parameter. Hardware requirements: PC, Supercomputer, Multi-platform, Workstation; Software requirements C and Fortan; Operation Systems/version is Linux/Unix or Windows
Energy Science and Technology Software Center (ESTSC)
2004-08-01
AnisWave2D is a 2D finite-difference code for a simulating seismic wave propagation in fully anisotropic materials. The code is implemented to run in parallel over multiple processors and is fully portable. A mesh refinement algorithm has been utilized to allow the grid-spacing to be tailored to the velocity model, avoiding the over-sampling of high-velocity materials that usually occurs in fixed-grid schemes.
Batakliev, Todor; Georgiev, Vladimir; Anachkov, Metody; Rakovsky, Slavcho
2014-01-01
Catalytic ozone decomposition is of great significance because ozone is a toxic substance commonly found or generated in human environments (aircraft cabins, offices with photocopiers, laser printers, sterilizers). Considerable work has been done on ozone decomposition reported in the literature. This review provides a comprehensive summary of the literature, concentrating on analysis of the physico-chemical properties, synthesis and catalytic decomposition of ozone. This is supplemented by a review on kinetics and catalyst characterization which ties together the previously reported results. Noble metals and oxides of transition metals have been found to be the most active substances for ozone decomposition. The high price of precious metals stimulated the use of metal oxide catalysts and particularly the catalysts based on manganese oxide. It has been determined that the kinetics of ozone decomposition is of first order importance. A mechanism of the reaction of catalytic ozone decomposition is discussed, based on detailed spectroscopic investigations of the catalytic surface, showing the existence of peroxide and superoxide surface intermediates. PMID:26109880
Batakliev, Todor; Georgiev, Vladimir; Anachkov, Metody; Rakovsky, Slavcho; Zaikov, Gennadi E
2014-06-01
Catalytic ozone decomposition is of great significance because ozone is a toxic substance commonly found or generated in human environments (aircraft cabins, offices with photocopiers, laser printers, sterilizers). Considerable work has been done on ozone decomposition reported in the literature. This review provides a comprehensive summary of the literature, concentrating on analysis of the physico-chemical properties, synthesis and catalytic decomposition of ozone. This is supplemented by a review on kinetics and catalyst characterization which ties together the previously reported results. Noble metals and oxides of transition metals have been found to be the most active substances for ozone decomposition. The high price of precious metals stimulated the use of metal oxide catalysts and particularly the catalysts based on manganese oxide. It has been determined that the kinetics of ozone decomposition is of first order importance. A mechanism of the reaction of catalytic ozone decomposition is discussed, based on detailed spectroscopic investigations of the catalytic surface, showing the existence of peroxide and superoxide surface intermediates. PMID:26109880
NASA Technical Reports Server (NTRS)
Steinthorsson, E.; Shih, T. I-P.; Roelke, R. J.
1991-01-01
In order to generate good quality systems for complicated three-dimensional spatial domains, the grid-generation method used must be able to exert rather precise controls over grid-point distributions. Several techniques are presented that enhance control of grid-point distribution for a class of algebraic grid-generation methods known as the two-, four-, and six-boundary methods. These techniques include variable stretching functions from bilinear interpolation, interpolating functions based on tension splines, and normalized K-factors. The techniques developed in this study were incorporated into a new version of GRID3D called GRID3D-v2. The usefulness of GRID3D-v2 was demonstrated by using it to generate a three-dimensional grid system in the coolent passage of a radial turbine blade with serpentine channels and pin fins.
Some nonlinear space decomposition algorithms
Tai, Xue-Cheng; Espedal, M.
1996-12-31
Convergence of a space decomposition method is proved for a general convex programming problem. The space decomposition refers to methods that decompose a space into sums of subspaces, which could be a domain decomposition or a multigrid method for partial differential equations. Two algorithms are proposed. Both can be used for linear as well as nonlinear elliptic problems and they reduce to the standard additive and multiplicative Schwarz methods for linear elliptic problems. Two {open_quotes}hybrid{close_quotes} algorithms are also presented. They converge faster than the additive one and have better parallelism than the multiplicative method. Numerical tests with a two level domain decomposition for linear, nonlinear and interface elliptic problems are presented for the proposed algorithms.
Chao, T.T.; Sanzolone, R.F.
1992-01-01
Sample decomposition is a fundamental and integral step in the procedure of geochemical analysis. It is often the limiting factor to sample throughput, especially with the recent application of the fast and modern multi-element measurement instrumentation. The complexity of geological materials makes it necessary to choose the sample decomposition technique that is compatible with the specific objective of the analysis. When selecting a decomposition technique, consideration should be given to the chemical and mineralogical characteristics of the sample, elements to be determined, precision and accuracy requirements, sample throughput, technical capability of personnel, and time constraints. This paper addresses these concerns and discusses the attributes and limitations of many techniques of sample decomposition along with examples of their application to geochemical analysis. The chemical properties of reagents as to their function as decomposition agents are also reviewed. The section on acid dissolution techniques addresses the various inorganic acids that are used individually or in combination in both open and closed systems. Fluxes used in sample fusion are discussed. The promising microwave-oven technology and the emerging field of automation are also examined. A section on applications highlights the use of decomposition techniques for the determination of Au, platinum group elements (PGEs), Hg, U, hydride-forming elements, rare earth elements (REEs), and multi-elements in geological materials. Partial dissolution techniques used for geochemical exploration which have been treated in detail elsewhere are not discussed here; nor are fire-assaying for noble metals and decomposition techniques for X-ray fluorescence or nuclear methods be discussed. ?? 1992.
MAGNUM-2D computer code: user's guide
England, R.L.; Kline, N.W.; Ekblad, K.J.; Baca, R.G.
1985-01-01
Information relevant to the general use of the MAGNUM-2D computer code is presented. This computer code was developed for the purpose of modeling (i.e., simulating) the thermal and hydraulic conditions in the vicinity of a waste package emplaced in a deep geologic repository. The MAGNUM-2D computer computes (1) the temperature field surrounding the waste package as a function of the heat generation rate of the nuclear waste and thermal properties of the basalt and (2) the hydraulic head distribution and associated groundwater flow fields as a function of the temperature gradients and hydraulic properties of the basalt. MAGNUM-2D is a two-dimensional numerical model for transient or steady-state analysis of coupled heat transfer and groundwater flow in a fractured porous medium. The governing equations consist of a set of coupled, quasi-linear partial differential equations that are solved using a Galerkin finite-element technique. A Newton-Raphson algorithm is embedded in the Galerkin functional to formulate the problem in terms of the incremental changes in the dependent variables. Both triangular and quadrilateral finite elements are used to represent the continuum portions of the spatial domain. Line elements may be used to represent discrete conduits. 18 refs., 4 figs., 1 tab.
NASA Astrophysics Data System (ADS)
Sui, Liansheng; Duan, Kuaikuai; Liang, Junli
2016-05-01
A secure double-image sharing scheme is proposed by using the Shamir's three-pass protocol in the discrete multiple-parameter fractional angular transform domain. First, an enlarged image is formed by assembling two plain images successively in the horizontal direction and scrambled in the chaotic permutation process, in which the sequences of chaotic pairs are generated by the two-dimensional Sine Logistic modulation map. Second, the scrambled image is divided into two components which are used to constitute a complex image. One component is normalized and regarded as the phase part of the complex image as well as other is considered as the amplitude part. Finally, the complex image is shared between the sender and the receiver by using the Shamir's three-pass protocol, in which the discrete multiple-parameter fractional angular transform is used as the encryption function due to its commutative property. The proposed double-image sharing scheme has an obvious advantage that the key management is convenient without distributing the random phase mask keys in advance. Moreover, the security of the image sharing scheme is enhanced with the help of extra parameters of the discrete multiple-parameter fractional angular transform. To the best of our knowledge, this is the first report on integrating the Shamir's three-pass protocol with double-image sharing scheme in the information security field. Simulation results and security analysis verify the feasibility and effectiveness of the proposed scheme.
2D constant-loss taper for mode conversion
NASA Astrophysics Data System (ADS)
Horth, Alexandre; Kashyap, Raman; Quitoriano, Nathaniel J.
2015-03-01
Proposed in this manuscript is a novel taper geometry, the constant-loss taper (CLT). This geometry is derived with 1D slabs of silicon embedded in silicon dioxide using coupled-mode theory (CMT). The efficiency of the CLT is compared to both linear and parabolic tapers using CMT and 2D finite-difference time-domain simulations. It is shown that over a short 2D, 4.45 μm long taper the CLT's mode conversion efficiency is ~90% which is 10% and 18% more efficient than a 2D parabolic or linear taper, respectively.
NASA Astrophysics Data System (ADS)
Mayor, Louise
2016-05-01
Graphene might be the most famous example, but there are other 2D materials and compounds too. Louise Mayor explains how these atomically thin sheets can be layered together to create flexible “van der Waals heterostructures”, which could lead to a range of novel applications.
ERIC Educational Resources Information Center
Napier, J.
1988-01-01
Outlines the role of the main organisms involved in woodland decomposition and discusses some of the variables affecting the rate of nutrient cycling. Suggests practical work that may be of value to high school students either as standard practice or long-term projects. (CW)
An analysis of scatter decomposition
NASA Technical Reports Server (NTRS)
Nicol, David M.; Saltz, Joel H.
1990-01-01
A formal analysis of a powerful mapping technique known as scatter decomposition is presented. Scatter decomposition divides an irregular computational domain into a large number of equal sized pieces, and distributes them modularly among processors. A probabilistic model of workload in one dimension is used to formally explain why, and when scatter decomposition works. The first result is that if correlation in workload is a convex function of distance, then scattering a more finely decomposed domain yields a lower average processor workload variance. The second result shows that if the workload process is stationary Gaussian and the correlation function decreases linearly in distance until becoming zero and then remains zero, scattering a more finely decomposed domain yields a lower expected maximum processor workload. Finally it is shown that if the correlation function decreases linearly across the entire domain, then among all mappings that assign an equal number of domain pieces to each processor, scatter decomposition minimizes the average processor workload variance. The dependence of these results on the assumption of decreasing correlation is illustrated with situations where a coarser granularity actually achieves better load balance.
Energy Science and Technology Software Center (ESTSC)
2001-01-31
This software reduces the data from two-dimensional kSA MOS program, k-Space Associates, Ann Arbor, MI. Initial MOS data is recorded without headers in 38 columns, with one row of data per acquisition per lase beam tracked. The final MOSS 2d data file is reduced, graphed, and saved in a tab-delimited column format with headers that can be plotted in any graphing software.
Nanoimprint lithography: 2D or not 2D? A review
NASA Astrophysics Data System (ADS)
Schift, Helmut
2015-11-01
Nanoimprint lithography (NIL) is more than a planar high-end technology for the patterning of wafer-like substrates. It is essentially a 3D process, because it replicates various stamp topographies by 3D displacement of material and takes advantage of the bending of stamps while the mold cavities are filled. But at the same time, it keeps all assets of a 2D technique being able to pattern thin masking layers like in photon- and electron-based traditional lithography. This review reports about 20 years of development of replication techniques at Paul Scherrer Institut, with a focus on 3D aspects of molding, which enable NIL to stay 2D, but at the same time enable 3D applications which are "more than Moore." As an example, the manufacturing of a demonstrator for backlighting applications based on thermally activated selective topography equilibration will be presented. This technique allows generating almost arbitrary sloped, convex and concave profiles in the same polymer film with dimensions in micro- and nanometer scale.
Dispersionless 2D Toda hierarchy, Hurwitz numbers and Riemann theorem
NASA Astrophysics Data System (ADS)
Natanzon, Sergey M.
2016-01-01
We describe all formal symmetric solutions of dispersionless 2D Toda hierarchy. This classification we use for solving of two classical problems: 1) The calculation of conformal mapping of an arbitrary simply connected domain to the standard disk; 2) Calculation of 2- Hurwitz numbers of genus 0.
Alloyed 2D Metal-Semiconductor Atomic Layer Junctions.
Kim, Ah Ra; Kim, Yonghun; Nam, Jaewook; Chung, Hee-Suk; Kim, Dong Jae; Kwon, Jung-Dae; Park, Sang Won; Park, Jucheol; Choi, Sun Young; Lee, Byoung Hun; Park, Ji Hyeon; Lee, Kyu Hwan; Kim, Dong-Ho; Choi, Sung Mook; Ajayan, Pulickel M; Hahm, Myung Gwan; Cho, Byungjin
2016-03-01
Heterostructures of compositionally and electronically variant two-dimensional (2D) atomic layers are viable building blocks for ultrathin optoelectronic devices. We show that the composition of interfacial transition region between semiconducting WSe2 atomic layer channels and metallic NbSe2 contact layers can be engineered through interfacial doping with Nb atoms. WxNb1-xSe2 interfacial regions considerably lower the potential barrier height of the junction, significantly improving the performance of the corresponding WSe2-based field-effect transistor devices. The creation of such alloyed 2D junctions between dissimilar atomic layer domains could be the most important factor in controlling the electronic properties of 2D junctions and the design and fabrication of 2D atomic layer devices. PMID:26839956
Sparse radar imaging using 2D compressed sensing
NASA Astrophysics Data System (ADS)
Hou, Qingkai; Liu, Yang; Chen, Zengping; Su, Shaoying
2014-10-01
Radar imaging is an ill-posed linear inverse problem and compressed sensing (CS) has been proved to have tremendous potential in this field. This paper surveys the theory of radar imaging and a conclusion is drawn that the processing of ISAR imaging can be denoted mathematically as a problem of 2D sparse decomposition. Based on CS, we propose a novel measuring strategy for ISAR imaging radar and utilize random sub-sampling in both range and azimuth dimensions, which will reduce the amount of sampling data tremendously. In order to handle 2D reconstructing problem, the ordinary solution is converting the 2D problem into 1D by Kronecker product, which will increase the size of dictionary and computational cost sharply. In this paper, we introduce the 2D-SL0 algorithm into the reconstruction of imaging. It is proved that 2D-SL0 can achieve equivalent result as other 1D reconstructing methods, but the computational complexity and memory usage is reduced significantly. Moreover, we will state the results of simulating experiments and prove the effectiveness and feasibility of our method.
2D Turbulence with Complicated Boundaries
NASA Astrophysics Data System (ADS)
Roullet, G.; McWilliams, J. C.
2014-12-01
We examine the consequences of lateral viscous boundary layers on the 2D turbulence that arises in domains with complicated boundaries (headlands, bays etc). The study is carried out numerically with LES. The numerics are carefully designed to ensure all global conservation laws, proper boundary conditions and a minimal range of dissipation scales. The turbulence dramatically differs from the classical bi-periodic case. Boundary layer separations lead to creation of many small vortices and act as a continuing energy source exciting the inverse cascade of energy throughout the domain. The detachments are very intermittent in time. In free decay, the final state depends on the effective numerical resolution: laminar with a single dominant vortex for low Re and turbulent with many vortices for large enough Re. After very long time, the turbulent end-state exhibits a striking tendency for the emergence of shielded vortices which then interact almost elastically. In the forced case, the boundary layers allow the turbulence to reach a statistical steady state without any artificial hypo-viscosity or other large-scale dissipation. Implications are discussed for the oceanic mesoscale and submesoscale turbulence.
Growth and Characterization of Silicon at the 2D Limit
NASA Astrophysics Data System (ADS)
Mannix, Andrew; Kiraly, Brian; Hersam, Mark; Guisinger, Nathan
2015-03-01
Because bulk silicon has dominated the development of microelectronics over the past 50 years, the recent interest in two-dimensional (2D) materials (e.g., graphene, MoS2, phosphorene, etc.) naturally raises questions regarding the growth and properties of silicon at the 2D limit. Utilizing atomic-scale, ultra-high vacuum (UHV) scanning tunneling microscopy (STM), we have investigated the 2D limits of silicon growth on Ag(111). In agreement with previous reports of sp2-bonded silicene phases, we observe the temperature-dependent evolution of ordered 2D phases. However, we attribute these to apparent Ag-Si surface alloys. At sufficiently high silicon coverage, we observe the precipitation of crystalline, sp3-bonded Si(111) domains. These domains are capped with a √3 honeycomb phase that is indistinguishable from the silver-induced √3 honeycomb-chained-trimer reconstruction on bulk Si(111). Further ex-situcharacterization with Raman spectroscopy, atomic force microscopy, cross-sectional transmission electron microscopy, Raman spectroscopy, and X-ray photoelectron spectroscopy reveals that these sheets are ultrathin sheets of bulk-like, (111) oriented, sp3 silicon. Even at the 2D limit, scanning tunneling spectroscopy shows that these silicon nanosheets exhibit semiconducting electronic characteristics.
Harmonic decomposition to describe the nonlinear evolution of stimulated Brillouin scattering
Hueller, S.; Masson-Laborde, P.E.; Pesme, D.; Casanova, M.; Detering, F.; Maximov, A.
2006-02-15
An efficient method to describe the nonlinear evolution of stimulated Brillouin scattering (SBS) in long scale-length plasmas is presented in the limit of a fluid description. The method is based on the decomposition of the various functions characterizing the plasma into their long- and short-wavelength components. It makes it possible to describe self-consistently the interplay between the plasma hydrodynamics, stimulated Brillouin scattering, and the generation of harmonics of the excited ion acoustic wave (IAW). This description is benchmarked numerically in one and two spatial dimensions [one dimensional (1D), two dimensional (2D)], by comparing the numerical results obtained along this method with those provided by a numerical code in which the decomposition into separate spatial scales is not made. The decomposition method proves to be very efficient in terms of computing time, especially in 2D, and very reliable, even in the extreme case of undamped ion acoustic waves. A novel picture of the SBS nonlinear behavior arises, in which the IAW harmonics generation gives rise to local defects appearing in the density and velocity hydrodynamics profiles. Consequently, SBS develops in various spatial domains which seem to be decorrelated one from each other, so that the backscattered Brillouin light is the sum of various backscattered waves generated in several independent spatial domains. It follows that the SBS reflectivity is chaotic in time and the resulting time-averaged value is significantly reduced as compared to the case when the IAW harmonics generation and flow modification are ignored. From the results of extensive numerical simulations carried out in 1D and 2D, we are able to infer the SBS reflectivity scaling law as a function of the plasma parameters and laser intensity, in the limit where the kinetic effects are negligible. It appears that this scaling law can be derived in the limit where the IAW harmonics generation is modeled simply by a
Metrology for graphene and 2D materials
NASA Astrophysics Data System (ADS)
Pollard, Andrew J.
2016-09-01
The application of graphene, a one atom-thick honeycomb lattice of carbon atoms with superlative properties, such as electrical conductivity, thermal conductivity and strength, has already shown that it can be used to benefit metrology itself as a new quantum standard for resistance. However, there are many application areas where graphene and other 2D materials, such as molybdenum disulphide (MoS2) and hexagonal boron nitride (h-BN), may be disruptive, areas such as flexible electronics, nanocomposites, sensing and energy storage. Applying metrology to the area of graphene is now critical to enable the new, emerging global graphene commercial world and bridge the gap between academia and industry. Measurement capabilities and expertise in a wide range of scientific areas are required to address this challenge. The combined and complementary approach of varied characterisation methods for structural, chemical, electrical and other properties, will allow the real-world issues of commercialising graphene and other 2D materials to be addressed. Here, examples of metrology challenges that have been overcome through a multi-technique or new approach are discussed. Firstly, the structural characterisation of defects in both graphene and MoS2 via Raman spectroscopy is described, and how nanoscale mapping of vacancy defects in graphene is also possible using tip-enhanced Raman spectroscopy (TERS). Furthermore, the chemical characterisation and removal of polymer residue on chemical vapour deposition (CVD) grown graphene via secondary ion mass spectrometry (SIMS) is detailed, as well as the chemical characterisation of iron films used to grow large domain single-layer h-BN through CVD growth, revealing how contamination of the substrate itself plays a role in the resulting h-BN layer. In addition, the role of international standardisation in this area is described, outlining the current work ongoing in both the International Organization of Standardization (ISO) and the
Competing coexisting phases in 2D water
Zanotti, Jean-Marc; Judeinstein, Patrick; Dalla-Bernardina, Simona; Creff, Gaëlle; Brubach, Jean-Blaise; Roy, Pascale; Bonetti, Marco; Ollivier, Jacques; Sakellariou, Dimitrios; Bellissent-Funel, Marie-Claire
2016-01-01
The properties of bulk water come from a delicate balance of interactions on length scales encompassing several orders of magnitudes: i) the Hydrogen Bond (HBond) at the molecular scale and ii) the extension of this HBond network up to the macroscopic level. Here, we address the physics of water when the three dimensional extension of the HBond network is frustrated, so that the water molecules are forced to organize in only two dimensions. We account for the large scale fluctuating HBond network by an analytical mean-field percolation model. This approach provides a coherent interpretation of the different events experimentally (calorimetry, neutron, NMR, near and far infra-red spectroscopies) detected in interfacial water at 160, 220 and 250 K. Starting from an amorphous state of water at low temperature, these transitions are respectively interpreted as the onset of creation of transient low density patches of 4-HBonded molecules at 160 K, the percolation of these domains at 220 K and finally the total invasion of the surface by them at 250 K. The source of this surprising behaviour in 2D is the frustration of the natural bulk tetrahedral local geometry and the underlying very significant increase in entropy of the interfacial water molecules. PMID:27185018
Competing coexisting phases in 2D water
NASA Astrophysics Data System (ADS)
Zanotti, Jean-Marc; Judeinstein, Patrick; Dalla-Bernardina, Simona; Creff, Gaëlle; Brubach, Jean-Blaise; Roy, Pascale; Bonetti, Marco; Ollivier, Jacques; Sakellariou, Dimitrios; Bellissent-Funel, Marie-Claire
2016-05-01
The properties of bulk water come from a delicate balance of interactions on length scales encompassing several orders of magnitudes: i) the Hydrogen Bond (HBond) at the molecular scale and ii) the extension of this HBond network up to the macroscopic level. Here, we address the physics of water when the three dimensional extension of the HBond network is frustrated, so that the water molecules are forced to organize in only two dimensions. We account for the large scale fluctuating HBond network by an analytical mean-field percolation model. This approach provides a coherent interpretation of the different events experimentally (calorimetry, neutron, NMR, near and far infra-red spectroscopies) detected in interfacial water at 160, 220 and 250 K. Starting from an amorphous state of water at low temperature, these transitions are respectively interpreted as the onset of creation of transient low density patches of 4-HBonded molecules at 160 K, the percolation of these domains at 220 K and finally the total invasion of the surface by them at 250 K. The source of this surprising behaviour in 2D is the frustration of the natural bulk tetrahedral local geometry and the underlying very significant increase in entropy of the interfacial water molecules.
Competing coexisting phases in 2D water.
Zanotti, Jean-Marc; Judeinstein, Patrick; Dalla-Bernardina, Simona; Creff, Gaëlle; Brubach, Jean-Blaise; Roy, Pascale; Bonetti, Marco; Ollivier, Jacques; Sakellariou, Dimitrios; Bellissent-Funel, Marie-Claire
2016-01-01
The properties of bulk water come from a delicate balance of interactions on length scales encompassing several orders of magnitudes: i) the Hydrogen Bond (HBond) at the molecular scale and ii) the extension of this HBond network up to the macroscopic level. Here, we address the physics of water when the three dimensional extension of the HBond network is frustrated, so that the water molecules are forced to organize in only two dimensions. We account for the large scale fluctuating HBond network by an analytical mean-field percolation model. This approach provides a coherent interpretation of the different events experimentally (calorimetry, neutron, NMR, near and far infra-red spectroscopies) detected in interfacial water at 160, 220 and 250 K. Starting from an amorphous state of water at low temperature, these transitions are respectively interpreted as the onset of creation of transient low density patches of 4-HBonded molecules at 160 K, the percolation of these domains at 220 K and finally the total invasion of the surface by them at 250 K. The source of this surprising behaviour in 2D is the frustration of the natural bulk tetrahedral local geometry and the underlying very significant increase in entropy of the interfacial water molecules. PMID:27185018
NKG2D ligands as therapeutic targets
Spear, Paul; Wu, Ming-Ru; Sentman, Marie-Louise; Sentman, Charles L.
2013-01-01
The Natural Killer Group 2D (NKG2D) receptor plays an important role in protecting the host from infections and cancer. By recognizing ligands induced on infected or tumor cells, NKG2D modulates lymphocyte activation and promotes immunity to eliminate ligand-expressing cells. Because these ligands are not widely expressed on healthy adult tissue, NKG2D ligands may present a useful target for immunotherapeutic approaches in cancer. Novel therapies targeting NKG2D ligands for the treatment of cancer have shown preclinical success and are poised to enter into clinical trials. In this review, the NKG2D receptor and its ligands are discussed in the context of cancer, infection, and autoimmunity. In addition, therapies targeting NKG2D ligands in cancer are also reviewed. PMID:23833565
Ultrafast 2D NMR: an emerging tool in analytical spectroscopy.
Giraudeau, Patrick; Frydman, Lucio
2014-01-01
Two-dimensional nuclear magnetic resonance (2D NMR) spectroscopy is widely used in chemical and biochemical analyses. Multidimensional NMR is also witnessing increased use in quantitative and metabolic screening applications. Conventional 2D NMR experiments, however, are affected by inherently long acquisition durations, arising from their need to sample the frequencies involved along their indirect domains in an incremented, scan-by-scan nature. A decade ago, a so-called ultrafast (UF) approach was proposed, capable of delivering arbitrary 2D NMR spectra involving any kind of homo- or heteronuclear correlation, in a single scan. During the intervening years, the performance of this subsecond 2D NMR methodology has been greatly improved, and UF 2D NMR is rapidly becoming a powerful analytical tool experiencing an expanded scope of applications. This review summarizes the principles and main developments that have contributed to the success of this approach and focuses on applications that have been recently demonstrated in various areas of analytical chemistry--from the real-time monitoring of chemical and biochemical processes, to extensions in hyphenated techniques and in quantitative applications. PMID:25014342
Ultrafast 2D NMR: An Emerging Tool in Analytical Spectroscopy
NASA Astrophysics Data System (ADS)
Giraudeau, Patrick; Frydman, Lucio
2014-06-01
Two-dimensional nuclear magnetic resonance (2D NMR) spectroscopy is widely used in chemical and biochemical analyses. Multidimensional NMR is also witnessing increased use in quantitative and metabolic screening applications. Conventional 2D NMR experiments, however, are affected by inherently long acquisition durations, arising from their need to sample the frequencies involved along their indirect domains in an incremented, scan-by-scan nature. A decade ago, a so-called ultrafast (UF) approach was proposed, capable of delivering arbitrary 2D NMR spectra involving any kind of homo- or heteronuclear correlation, in a single scan. During the intervening years, the performance of this subsecond 2D NMR methodology has been greatly improved, and UF 2D NMR is rapidly becoming a powerful analytical tool experiencing an expanded scope of applications. This review summarizes the principles and main developments that have contributed to the success of this approach and focuses on applications that have been recently demonstrated in various areas of analytical chemistry—from the real-time monitoring of chemical and biochemical processes, to extensions in hyphenated techniques and in quantitative applications.
Mode decomposition evolution equations
Wang, Yang; Wei, Guo-Wei; Yang, Siyang
2011-01-01
Partial differential equation (PDE) based methods have become some of the most powerful tools for exploring the fundamental problems in signal processing, image processing, computer vision, machine vision and artificial intelligence in the past two decades. The advantages of PDE based approaches are that they can be made fully automatic, robust for the analysis of images, videos and high dimensional data. A fundamental question is whether one can use PDEs to perform all the basic tasks in the image processing. If one can devise PDEs to perform full-scale mode decomposition for signals and images, the modes thus generated would be very useful for secondary processing to meet the needs in various types of signal and image processing. Despite of great progress in PDE based image analysis in the past two decades, the basic roles of PDEs in image/signal analysis are only limited to PDE based low-pass filters, and their applications to noise removal, edge detection, segmentation, etc. At present, it is not clear how to construct PDE based methods for full-scale mode decomposition. The above-mentioned limitation of most current PDE based image/signal processing methods is addressed in the proposed work, in which we introduce a family of mode decomposition evolution equations (MoDEEs) for a vast variety of applications. The MoDEEs are constructed as an extension of a PDE based high-pass filter (Europhys. Lett., 59(6): 814, 2002) by using arbitrarily high order PDE based low-pass filters introduced by Wei (IEEE Signal Process. Lett., 6(7): 165, 1999). The use of arbitrarily high order PDEs is essential to the frequency localization in the mode decomposition. Similar to the wavelet transform, the present MoDEEs have a controllable time-frequency localization and allow a perfect reconstruction of the original function. Therefore, the MoDEE operation is also called a PDE transform. However, modes generated from the present approach are in the spatial or time domain and can be
Perspectives for spintronics in 2D materials
NASA Astrophysics Data System (ADS)
Han, Wei
2016-03-01
The past decade has been especially creative for spintronics since the (re)discovery of various two dimensional (2D) materials. Due to the unusual physical characteristics, 2D materials have provided new platforms to probe the spin interaction with other degrees of freedom for electrons, as well as to be used for novel spintronics applications. This review briefly presents the most important recent and ongoing research for spintronics in 2D materials.
Fast 2D FWI on a multi and many-cores workstation.
NASA Astrophysics Data System (ADS)
Thierry, Philippe; Donno, Daniela; Noble, Mark
2014-05-01
Following the introduction of x86 co-processors (Xeon Phi) and the performance increase of standard 2-socket workstations using the latest 12 cores E5-v2 x86-64 CPU, we present here a MPI + OpenMP implementation of an acoustic 2D FWI (full waveform inversion) code which simultaneously runs on the CPUs and on the co-processors installed in a workstation. The main advantage of running a 2D FWI on a workstation is to be able to quickly evaluate new features such as more complicated wave equations, new cost functions, finite-difference stencils or boundary conditions. Since the co-processor is made of 61 in-order x86 cores, each of them having up to 4 threads, this many-core can be seen as a shared memory SMP (symmetric multiprocessing) machine with its own IP address. Depending on the vendor, a single workstation can handle several co-processors making the workstation as a personal cluster under the desk. The original Fortran 90 CPU version of the 2D FWI code is just recompiled to get a Xeon Phi x86 binary. This multi and many-core configuration uses standard compilers and associated MPI as well as math libraries under Linux; therefore, the cost of code development remains constant, while improving computation time. We choose to implement the code with the so-called symmetric mode to fully use the capacity of the workstation, but we also evaluate the scalability of the code in native mode (i.e running only on the co-processor) thanks to the Linux ssh and NFS capabilities. Usual care of optimization and SIMD vectorization is used to ensure optimal performances, and to analyze the application performances and bottlenecks on both platforms. The 2D FWI implementation uses finite-difference time-domain forward modeling and a quasi-Newton (with L-BFGS algorithm) optimization scheme for the model parameters update. Parallelization is achieved through standard MPI shot gathers distribution and OpenMP for domain decomposition within the co-processor. Taking advantage of the 16
Annotated Bibliography of EDGE2D Use
J.D. Strachan and G. Corrigan
2005-06-24
This annotated bibliography is intended to help EDGE2D users, and particularly new users, find existing published literature that has used EDGE2D. Our idea is that a person can find existing studies which may relate to his intended use, as well as gain ideas about other possible applications by scanning the attached tables.
Staring 2-D hadamard transform spectral imager
Gentry, Stephen M.; Wehlburg, Christine M.; Wehlburg, Joseph C.; Smith, Mark W.; Smith, Jody L.
2006-02-07
A staring imaging system inputs a 2D spatial image containing multi-frequency spectral information. This image is encoded in one dimension of the image with a cyclic Hadamarid S-matrix. The resulting image is detecting with a spatial 2D detector; and a computer applies a Hadamard transform to recover the encoded image.
Multiple-Parameter Estimation Method Based on Spatio-Temporal 2-D Processing for Bistatic MIMO Radar
Yang, Shouguo; Li, Yong; Zhang, Kunhui; Tang, Weiping
2015-01-01
A novel spatio-temporal 2-dimensional (2-D) processing method that can jointly estimate the transmitting-receiving azimuth and Doppler frequency for bistatic multiple-input multiple-output (MIMO) radar in the presence of spatial colored noise and an unknown number of targets is proposed. In the temporal domain, the cross-correlation of the matched filters’ outputs for different time-delay sampling is used to eliminate the spatial colored noise. In the spatial domain, the proposed method uses a diagonal loading method and subspace theory to estimate the direction of departure (DOD) and direction of arrival (DOA), and the Doppler frequency can then be accurately estimated through the estimation of the DOD and DOA. By skipping target number estimation and the eigenvalue decomposition (EVD) of the data covariance matrix estimation and only requiring a one-dimensional search, the proposed method achieves low computational complexity. Furthermore, the proposed method is suitable for bistatic MIMO radar with an arbitrary transmitted and received geometrical configuration. The correction and efficiency of the proposed method are verified by computer simulation results. PMID:26694385
Yang, Shouguo; Li, Yong; Zhang, Kunhui; Tang, Weiping
2015-01-01
A novel spatio-temporal 2-dimensional (2-D) processing method that can jointly estimate the transmitting-receiving azimuth and Doppler frequency for bistatic multiple-input multiple-output (MIMO) radar in the presence of spatial colored noise and an unknown number of targets is proposed. In the temporal domain, the cross-correlation of the matched filters' outputs for different time-delay sampling is used to eliminate the spatial colored noise. In the spatial domain, the proposed method uses a diagonal loading method and subspace theory to estimate the direction of departure (DOD) and direction of arrival (DOA), and the Doppler frequency can then be accurately estimated through the estimation of the DOD and DOA. By skipping target number estimation and the eigenvalue decomposition (EVD) of the data covariance matrix estimation and only requiring a one-dimensional search, the proposed method achieves low computational complexity. Furthermore, the proposed method is suitable for bistatic MIMO radar with an arbitrary transmitted and received geometrical configuration. The correction and efficiency of the proposed method are verified by computer simulation results. PMID:26694385
Light field morphing using 2D features.
Wang, Lifeng; Lin, Stephen; Lee, Seungyong; Guo, Baining; Shum, Heung-Yeung
2005-01-01
We present a 2D feature-based technique for morphing 3D objects represented by light fields. Existing light field morphing methods require the user to specify corresponding 3D feature elements to guide morph computation. Since slight errors in 3D specification can lead to significant morphing artifacts, we propose a scheme based on 2D feature elements that is less sensitive to imprecise marking of features. First, 2D features are specified by the user in a number of key views in the source and target light fields. Then the two light fields are warped view by view as guided by the corresponding 2D features. Finally, the two warped light fields are blended together to yield the desired light field morph. Two key issues in light field morphing are feature specification and warping of light field rays. For feature specification, we introduce a user interface for delineating 2D features in key views of a light field, which are automatically interpolated to other views. For ray warping, we describe a 2D technique that accounts for visibility changes and present a comparison to the ideal morphing of light fields. Light field morphing based on 2D features makes it simple to incorporate previous image morphing techniques such as nonuniform blending, as well as to morph between an image and a light field. PMID:15631126