The CP-PACS Project and Lattice QCD Results
NASA Astrophysics Data System (ADS)
Iwasaki, Y.
The aim of the CP-PACS project was to develop a massively parallel computer for performing numerical research in computational physics with primary emphasis on lattice QCD. The CP-PACS computer with a peak speed of 614 GFLOPS with 2048 processors was completed in September 1996, and has been in full operation since October 1996. We present an overview of the CP-PACS project and describe characteristics of the CP-PACS computer. The CP-PACS has been mainly used for hadron spectroscopy studies in lattice QCD. Main results in lattice QCD simulations are given.
NASA Astrophysics Data System (ADS)
Ukawa, Akira
1998-05-01
The CP-PACS computer is a massively parallel computer consisting of 2048 processing units and having a peak speed of 614 GFLOPS and 128 GByte of main memory. It was developed over the four years from 1992 to 1996 at the Center for Computational Physics, University of Tsukuba, for large-scale numerical simulations in computational physics, especially those of lattice QCD. The CP-PACS computer has been in full operation for physics computations since October 1996. In this article we describe the chronology of the development, the hardware and software characteristics of the computer, and its performance for lattice QCD simulations.
The International Conference on Vector and Parallel Computing (2nd)
1989-01-17
Computation of the SVD of Bidiagonal Matrices" ...................................... 11 " Lattice QCD -As a Large Scale Scientific Computation...vectorizcd for the IBM 3090 Vector Facility. In addition, elapsed times " Lattice QCD -As a Large Scale Scientific have been reduced by using 3090...benchmarked Lattice QCD on a large number ofcompu- come from the wavefront solver routine. This was exten- ters: CrayX-MP and Cray 2 (vector
A scalable PC-based parallel computer for lattice QCD
NASA Astrophysics Data System (ADS)
Fodor, Z.; Katz, S. D.; Pappa, G.
2003-05-01
A PC-based parallel computer for medium/large scale lattice QCD simulations is suggested. The Eo¨tvo¨s Univ., Inst. Theor. Phys. cluster consists of 137 Intel P4-1.7GHz nodes. Gigabit Ethernet cards are used for nearest neighbor communication in a two-dimensional mesh. The sustained performance for dynamical staggered (wilson) quarks on large lattices is around 70(110) GFlops. The exceptional price/performance ratio is below $1/Mflop.
Better than $l/Mflops sustained: a scalable PC-based parallel computer for lattice QCD
NASA Astrophysics Data System (ADS)
Fodor, Zoltán; Katz, Sándor D.; Papp, Gábor
2003-05-01
We study the feasibility of a PC-based parallel computer for medium to large scale lattice QCD simulations. The Eötvös Univ., Inst. Theor. Phys. cluster consists of 137 Intel P4-1.7GHz nodes with 512 MB RDRAM. The 32-bit, single precision sustained performance for dynamical QCD without communication is 1510 Mflops/node with Wilson and 970 Mflops/node with staggered fermions. This gives a total performance of 208 Gflops for Wilson and 133 Gflops for staggered QCD, respectively (for 64-bit applications the performance is approximately halved). The novel feature of our system is its communication architecture. In order to have a scalable, cost-effective machine we use Gigabit Ethernet cards for nearest-neighbor communications in a two-dimensional mesh. This type of communication is cost effective (only 30% of the hardware costs is spent on the communication). According to our benchmark measurements this type of communication results in around 40% communication time fraction for lattices upto 48 3·96 in full QCD simulations. The price/sustained-performance ratio for full QCD is better than l/Mflops for Wilson (and around 1.5/Mflops for staggered) quarks for practically any lattice size, which can fit in our parallel computer. The communication software is freely available upon request for non-profit organizations.
Kenneth Wilson and Lattice QCD
NASA Astrophysics Data System (ADS)
Ukawa, Akira
2015-09-01
We discuss the physics and computation of lattice QCD, a space-time lattice formulation of quantum chromodynamics, and Kenneth Wilson's seminal role in its development. We start with the fundamental issue of confinement of quarks in the theory of the strong interactions, and discuss how lattice QCD provides a framework for understanding this phenomenon. A conceptual issue with lattice QCD is a conflict of space-time lattice with chiral symmetry of quarks. We discuss how this problem is resolved. Since lattice QCD is a non-linear quantum dynamical system with infinite degrees of freedom, quantities which are analytically calculable are limited. On the other hand, it provides an ideal case of massively parallel numerical computations. We review the long and distinguished history of parallel-architecture supercomputers designed and built for lattice QCD. We discuss algorithmic developments, in particular the difficulties posed by the fermionic nature of quarks, and their resolution. The triad of efforts toward better understanding of physics, better algorithms, and more powerful supercomputers have produced major breakthroughs in our understanding of the strong interactions. We review the salient results of this effort in understanding the hadron spectrum, the Cabibbo-Kobayashi-Maskawa matrix elements and CP violation, and quark-gluon plasma at high temperatures. We conclude with a brief summary and a future perspective.
Applications of Massive Mathematical Computations
1990-04-01
particles from the first principles of QCD . This problem is under intensive numerical study 11-6 using special purpose parallel supercomputers in...several places around the world. The method used here is the Monte Carlo integration for a fixed 3-D plus time lattices . Reliable results are still years...mathematical and theoretical physics, but its most promising applications are in the numerical realization of QCD computations. Our programs for the solution
NASA Astrophysics Data System (ADS)
Iwasaki, Y.; CP-PACS Collaboration
1998-01-01
The CP-PACS project is a five year plan, which formally started in April 1992 and has been completed in March 1997, to develop a massively parallel computer for carrying out research in computational physics with primary emphasis on lattice QCD. The initial version of the CP-PACS computer with a theoretical peak speed of 307 GFLOPS with 1024 processors was completed in March 1996. The final version with a peak speed of 614 GFLOPS with 2048 processors was completed in September 1996, and has been in full operation since October 1996. We describe the architecture, the final specification, the hardware implementation, and the software of the CP-PACS computer. The CP-PACS has been used for hadron spectroscopy production runs since July 1996. The performance for lattice QCD applications and the LINPACK benchmark are given.
Lattice QCD calculation using VPP500
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, Seyong; Ohta, Shigemi
1995-02-01
A new vector parallel supercomputer, Fujitsu VPP500, was installed at RIKEN earlier this year. It consists of 30 vector computers, each with 1.6 GFLOPS peak speed and 256 MB memory, connected by a crossbar switch with 400 MB/s peak data transfer rate each way between any pair of nodes. The authors developed a Fortran lattice QCD simulation code for it. It runs at about 1.1 GFLOPS sustained per node for Metropolis pure-gauge update, and about 0.8 GFLOPS sustained per node for conjugate gradient inversion of staggered fermion matrix.
Lattice QCD Calculations in Nuclear Physics towards the Exascale
NASA Astrophysics Data System (ADS)
Joo, Balint
2017-01-01
The combination of algorithmic advances and new highly parallel computing architectures are enabling lattice QCD calculations to tackle ever more complex problems in nuclear physics. In this talk I will review some computational challenges that are encountered in large scale cold nuclear physics campaigns such as those in hadron spectroscopy calculations. I will discuss progress in addressing these with algorithmic improvements such as multi-grid solvers and software for recent hardware architectures such as GPUs and Intel Xeon Phi, Knights Landing. Finally, I will highlight some current topics for research and development as we head towards the Exascale era This material is funded by the U.S. Department of Energy, Office Of Science, Offices of Nuclear Physics, High Energy Physics and Advanced Scientific Computing Research, as well as the Office of Nuclear Physics under contract DE-AC05-06OR23177.
The Secret Life of Quarks, Final Report for the University of North Carolina at Chapel Hill
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fowler, Robert J.
This final report summarizes activities and results at the University of North Carolina as part of the the SciDAC-2 Project The Secret Life of Quarks: National Computational Infrastructure for Lattice Quantum Chromodynamics. The overall objective of the project is to construct the software needed to study quantum chromo- dynamics (QCD), the theory of the strong interactions of subatomic physics, and similar strongly coupled gauge theories anticipated to be of importance in the LHC era. It built upon the successful efforts of the SciDAC-1 project National Computational Infrastructure for Lattice Gauge Theory, in which a QCD Applications Programming Interface (QCD API)more » was developed that enables lat- tice gauge theorists to make effective use of a wide variety of massively parallel computers. In the SciDAC-2 project, optimized versions of the QCD API were being created for the IBM Blue- Gene/L (BG/L) and BlueGene/P (BG/P), the Cray XT3/XT4 and its successors, and clusters based on multi-core processors and Infiniband communications networks. The QCD API is being used to enhance the performance of the major QCD community codes and to create new applications. Software libraries of physics tools have been expanded to contain sharable building blocks for inclusion in application codes, performance analysis and visualization tools, and software for au- tomation of physics work flow. New software tools were designed for managing the large data sets generated in lattice QCD simulations, and for sharing them through the International Lattice Data Grid consortium. As part of the overall project, researchers at UNC were funded through ASCR to work in three general areas. The main thrust has been performance instrumentation and analysis in support of the SciDAC QCD code base as it evolved and as it moved to new computation platforms. In support of the performance activities, performance data was to be collected in a database for the purpose of broader analysis. Third, the UNC work was done at RENCI (Renaissance Computing Institute), which has extensive expertise and facilities for scientific data visualization, so we acted in an ongoing consulting and support role in that area.« less
Portable multi-node LQCD Monte Carlo simulations using OpenACC
NASA Astrophysics Data System (ADS)
Bonati, Claudio; Calore, Enrico; D'Elia, Massimo; Mesiti, Michele; Negro, Francesco; Sanfilippo, Francesco; Schifano, Sebastiano Fabio; Silvi, Giorgio; Tripiccione, Raffaele
This paper describes a state-of-the-art parallel Lattice QCD Monte Carlo code for staggered fermions, purposely designed to be portable across different computer architectures, including GPUs and commodity CPUs. Portability is achieved using the OpenACC parallel programming model, used to develop a code that can be compiled for several processor architectures. The paper focuses on parallelization on multiple computing nodes using OpenACC to manage parallelism within the node, and OpenMPI to manage parallelism among the nodes. We first discuss the available strategies to be adopted to maximize performances, we then describe selected relevant details of the code, and finally measure the level of performance and scaling-performance that we are able to achieve. The work focuses mainly on GPUs, which offer a significantly high level of performances for this application, but also compares with results measured on other processors.
High-Performance I/O: HDF5 for Lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kurth, Thorsten; Pochinsky, Andrew; Sarje, Abhinav
2015-01-01
Practitioners of lattice QCD/QFT have been some of the primary pioneer users of the state-of-the-art high-performance-computing systems, and contribute towards the stress tests of such new machines as soon as they become available. As with all aspects of high-performance-computing, I/O is becoming an increasingly specialized component of these systems. In order to take advantage of the latest available high-performance I/O infrastructure, to ensure reliability and backwards compatibility of data files, and to help unify the data structures used in lattice codes, we have incorporated parallel HDF5 I/O into the SciDAC supported USQCD software stack. Here we present the design andmore » implementation of this I/O framework. Our HDF5 implementation outperforms optimized QIO at the 10-20% level and leaves room for further improvement by utilizing appropriate dataset chunking.« less
High-Performance I/O: HDF5 for Lattice QCD
Kurth, Thorsten; Pochinsky, Andrew; Sarje, Abhinav; ...
2017-05-09
Practitioners of lattice QCD/QFT have been some of the primary pioneer users of the state-of-the-art high-performance-computing systems, and contribute towards the stress tests of such new machines as soon as they become available. As with all aspects of high-performance-computing, I/O is becoming an increasingly specialized component of these systems. In order to take advantage of the latest available high-performance I/O infrastructure, to ensure reliability and backwards compatibility of data files, and to help unify the data structures used in lattice codes, we have incorporated parallel HDF5 I/O into the SciDAC supported USQCD software stack. Here we present the design andmore » implementation of this I/O framework. Our HDF5 implementation outperforms optimized QIO at the 10-20% level and leaves room for further improvement by utilizing appropriate dataset chunking.« less
Final Report: Subcontract B623868 Algebraic Multigrid solvers for coupled PDE systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brannick, J.
The Pennsylvania State University (“Subcontractor”) continued to work on the design of algebraic multigrid solvers for coupled systems of partial differential equations (PDEs) arising in numerical modeling of various applications, with a main focus on solving the Dirac equation arising in Quantum Chromodynamics (QCD). The goal of the proposed work was to develop combined geometric and algebraic multilevel solvers that are robust and lend themselves to efficient implementation on massively parallel heterogeneous computers for these QCD systems. The research in these areas built on previous works, focusing on the following three topics: (1) the development of parallel full-multigrid (PFMG) andmore » non-Galerkin coarsening techniques in this frame work for solving the Wilson Dirac system; (2) the use of these same Wilson MG solvers for preconditioning the Overlap and Domain Wall formulations of the Dirac equation; and (3) the design and analysis of algebraic coarsening algorithms for coupled PDE systems including Stokes equation, Maxwell equation and linear elasticity.« less
Northeast Parallel Architectures Center (NPAC) at Syracuse University
1990-12-01
lattice models. On the CM-2 we will fun a lattice gauge theory simulation of quantum chromodynamics ( QCD ), and on the CM-1 we will investigate the...into a three-dimensional grid with the stipulation that adjacent processors in the lattice correspond to proximate regions of space. Light paths will...be constrained to follow lattice links and the sum over all paths from light sources to each lattice site will be computed inductively by all
Solving lattice QCD systems of equations using mixed precision solvers on GPUs
NASA Astrophysics Data System (ADS)
Clark, M. A.; Babich, R.; Barros, K.; Brower, R. C.; Rebbi, C.
2010-09-01
Modern graphics hardware is designed for highly parallel numerical tasks and promises significant cost and performance benefits for many scientific applications. One such application is lattice quantum chromodynamics (lattice QCD), where the main computational challenge is to efficiently solve the discretized Dirac equation in the presence of an SU(3) gauge field. Using NVIDIA's CUDA platform we have implemented a Wilson-Dirac sparse matrix-vector product that performs at up to 40, 135 and 212 Gflops for double, single and half precision respectively on NVIDIA's GeForce GTX 280 GPU. We have developed a new mixed precision approach for Krylov solvers using reliable updates which allows for full double precision accuracy while using only single or half precision arithmetic for the bulk of the computation. The resulting BiCGstab and CG solvers run in excess of 100 Gflops and, in terms of iterations until convergence, perform better than the usual defect-correction approach for mixed precision.
Selected inversion as key to a stable Langevin evolution across the QCD phase boundary
NASA Astrophysics Data System (ADS)
Bloch, Jacques; Schenk, Olaf
2018-03-01
We present new results of full QCD at nonzero chemical potential. In PRD 92, 094516 (2015) the complex Langevin method was shown to break down when the inverse coupling decreases and enters the transition region from the deconfined to the confined phase. We found that the stochastic technique used to estimate the drift term can be very unstable for indefinite matrices. This may be avoided by using the full inverse of the Dirac operator, which is, however, too costly for four-dimensional lattices. The major breakthrough in this work was achieved by realizing that the inverse elements necessary for the drift term can be computed efficiently using the selected inversion technique provided by the parallel sparse direct solver package PARDISO. In our new study we show that no breakdown of the complex Langevin method is encountered and that simulations can be performed across the phase boundary.
Towards understanding Regge trajectories in holographic QCD
NASA Astrophysics Data System (ADS)
Catà, Oscar
2007-05-01
We reassess a work done by Migdal on the spectrum of low-energy vector mesons in QCD in the light of the anti-de Sitter (AdS)-QCD correspondence. Recently, a tantalizing parallelism was suggested between Migdal’s work and a family of holographic duals of QCD. Despite the intriguing similarities, both approaches face a major drawback: the spectrum is in conflict with well-tested Regge scaling. However, it has recently been shown that holographic duals can be modified to accommodate Regge behavior. Therefore, it is interesting to understand whether Regge behavior can also be achieved in Migdal’s approach. In this paper we investigate this issue. We find that Migdal’s approach, which is based on a modified Padé approximant, is closely related to the issue of quark-hadron duality breakdown in QCD.
Reliable semiclassical computations in QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dine, Michael; Department of Physics, Stanford University Stanford, California 94305-4060; Festuccia, Guido
We revisit the question of whether or not one can perform reliable semiclassical QCD computations at zero temperature. We study correlation functions with no perturbative contributions, and organize the problem by means of the operator product expansion, establishing a precise criterion for the validity of a semiclassical calculation. For N{sub f}>N, a systematic computation is possible; for N{sub f}
Hadron mass and decays constant predictions of the valence approximation to lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Weingarten, D.
1993-05-01
A key goal of the lattice formulation of QCD is to reproduce the masses and decay constants of the low-lying baryons and mesons. Lattice QCD mass and decay constant predictions for the real world are supposed to be obtained from masses and decay constants calculated with finite lattice spacing and finite lattice volume by taking the limits of zero spacing and infinite volume. In addition, since the algorithms used for hadron mass and decay constant calculations become progressively slower for small quark masses, results are presently found with quark masses much larger than the expected values of the up andmore » down quark masses. Predictions for the properties of hadrons containing up and down quarks then require a further extrapolation to small quark masses. The author reports here mass and decay constant predictions combining all three extrapolations for Wilson quarks in the valence (quenched) approximation. This approximation may be viewed as replacing the momentum and frequency dependent color dielectric constant arising from quark-antiquark vacuum polarization with its zero-momentum, zero-frequency limit. These calculations used approximately one year of machine time on the GF11 parallel computer running at a sustained rate of between 5 and 7 Gflops.« less
QCDOC: A 10-teraflops scale computer for lattice QCD
NASA Astrophysics Data System (ADS)
Chen, D.; Christ, N. H.; Cristian, C.; Dong, Z.; Gara, A.; Garg, K.; Joo, B.; Kim, C.; Levkova, L.; Liao, X.; Mawhinney, R. D.; Ohta, S.; Wettig, T.
2001-03-01
The architecture of a new class of computers, optimized for lattice QCD calculations, is described. An individual node is based on a single integrated circuit containing a PowerPC 32-bit integer processor with a 1 Gflops 64-bit IEEE floating point unit, 4 Mbyte of memory, 8 Gbit/sec nearest-neighbor communications and additional control and diagnostic circuitry. The machine's name, QCDOC, derives from "QCD On a Chip".
Lattice QCD Application Development within the US DOE Exascale Computing Project
NASA Astrophysics Data System (ADS)
Brower, Richard; Christ, Norman; DeTar, Carleton; Edwards, Robert; Mackenzie, Paul
2018-03-01
In October, 2016, the US Department of Energy launched the Exascale Computing Project, which aims to deploy exascale computing resources for science and engineering in the early 2020's. The project brings together application teams, software developers, and hardware vendors in order to realize this goal. Lattice QCD is one of the applications. Members of the US lattice gauge theory community with significant collaborators abroad are developing algorithms and software for exascale lattice QCD calculations. We give a short description of the project, our activities, and our plans.
Lattice QCD Application Development within the US DOE Exascale Computing Project
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brower, Richard; Christ, Norman; DeTar, Carleton
In October, 2016, the US Department of Energy launched the Exascale Computing Project, which aims to deploy exascale computing resources for science and engineering in the early 2020's. The project brings together application teams, software developers, and hardware vendors in order to realize this goal. Lattice QCD is one of the applications. Members of the US lattice gauge theory community with significant collaborators abroad are developing algorithms and software for exascale lattice QCD calculations. We give a short description of the project, our activities, and our plans.
QCD corrections to ZZ production in gluon fusion at the LHC
Caola, Fabrizio; Melnikov, Kirill; Rontsch, Raoul; ...
2015-11-23
We compute the next-to-leading-order QCD corrections to the production of two Z-bosons in the annihilation of two gluons at the LHC. Being enhanced by a large gluon flux, these corrections provide a distinct and, potentially, the dominant part of the N 3LO QCD contributions to Z-pair production in proton collisions. The gg → ZZ annihilation is a loop-induced process that receives the dominant contribution from loops of five light quarks, that are included in our computation in the massless approximation. We find that QCD corrections increase the gg → ZZ production cross section by O(50%–100%) depending on the values ofmore » the renormalization and factorization scales used in the leading-order computation and the collider energy. Furthermore, the large corrections to the gg → ZZ channel increase the pp → ZZ cross section by about 6% to 8%, exceeding the estimated theoretical uncertainty of the recent next-to-next-to-leading-order QCD calculation.« less
Higher order corrections to mixed QCD-EW contributions to Higgs boson production in gluon fusion
NASA Astrophysics Data System (ADS)
Bonetti, Marco; Melnikov, Kirill; Tancredi, Lorenzo
2018-03-01
We present an estimate of the next-to-leading-order (NLO) QCD corrections to mixed QCD-electroweak contributions to the Higgs boson production cross section in gluon fusion, combining the recently computed three-loop virtual corrections and the approximate treatment of real emission in the soft approximation. We find that the NLO QCD corrections to the mixed QCD-electroweak contributions are nearly identical to NLO QCD corrections to QCD Higgs production. Our result confirms an earlier estimate of these O (α αs2) effects by Anastasiou et al. [J. High Energy Phys. 04 (2009) 003, 10.1088/1126-6708/2009/04/003] and provides further support for the factorization approximation of QCD and electroweak corrections.
K-->pipi amplitudes from lattice QCD with a light charm quark.
Giusti, L; Hernández, P; Laine, M; Pena, C; Wennekers, J; Wittig, H
2007-02-23
We compute the leading-order low-energy constants of the DeltaS=1 effective weak Hamiltonian in the quenched approximation of QCD with up, down, strange, and charm quarks degenerate and light. They are extracted by comparing the predictions of finite-volume chiral perturbation theory with lattice QCD computations of suitable correlation functions carried out with quark masses ranging from a few MeV up to half of the physical strange mass. We observe a DeltaI=1/2 enhancement in this corner of the parameter space of the theory. Although matching with the experimental result is not observed for the DeltaI=1/2 amplitude, our computation suggests large QCD contributions to the physical DeltaI=1/2 rule in the GIM limit, and represents the first step to quantify the role of the charm-quark mass in K-->pipi amplitudes. The use of fermions with an exact chiral symmetry is an essential ingredient in our computation.
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
1998-10-16
A workshop was held at the RIKEN-BNL Research Center on October 16, 1998, as part of the first anniversary celebration for the center. This meeting brought together the physicists from RIKEN-BNL, BNL and Columbia who are using the QCDSP (Quantum Chromodynamics on Digital Signal Processors) computer at the RIKEN-BNL Research Center for studies of QCD. Many of the talks in the workshop were devoted to domain wall fermions, a discretization of the continuum description of fermions which preserves the global symmetries of the continuum, even at finite lattice spacing. This formulation has been the subject of analytic investigation for somemore » time and has reached the stage where large-scale simulations in QCD seem very promising. With the computational power available from the QCDSP computers, scientists are looking forward to an exciting time for numerical simulations of QCD.« less
Parallel Processing and Scientific Applications
1992-11-30
Lattice QCD Calculations on the Connection Machine), SIAM News 24, 1 (May 1991) 5. C. F. Baillie and D. A. Johnston, Crumpling Dynamically Triangulated...hypercubic lattice ; in the second, the surface is randomly triangulated once at the beginning of the simulation; and in the third the random...Sharpe, QCD with Dynamical Wilson Fermions 1I, Phys. Rev. D44, 3272 (1991), 8. R. Gupta and C. F. Baillie, Critical Behavior of the 2D XY Model, Phys
NASA Astrophysics Data System (ADS)
Bornyakov, V. G.; Boyda, D. L.; Goy, V. A.; Molochkov, A. V.; Nakamura, Atsushi; Nikolaev, A. A.; Zakharov, V. I.
2017-05-01
We propose and test a new approach to computation of canonical partition functions in lattice QCD at finite density. We suggest a few steps procedure. We first compute numerically the quark number density for imaginary chemical potential i μq I . Then we restore the grand canonical partition function for imaginary chemical potential using the fitting procedure for the quark number density. Finally we compute the canonical partition functions using high precision numerical Fourier transformation. Additionally we compute the canonical partition functions using the known method of the hopping parameter expansion and compare results obtained by two methods in the deconfining as well as in the confining phases. The agreement between two methods indicates the validity of the new method. Our numerical results are obtained in two flavor lattice QCD with clover improved Wilson fermions.
Lepton-rich cold QCD matter in protoneutron stars
NASA Astrophysics Data System (ADS)
Jiménez, J. C.; Fraga, E. S.
2018-05-01
We investigate protoneutron star matter using the state-of-the-art perturbative equation of state for cold and dense QCD in the presence of a fixed lepton fraction in which both electrons and neutrinos are included. Besides computing the modifications in the equation of state due to the presence of trapped neutrinos, we show that stable strange quark matter has a more restricted parameter space. We also study the possibility of nucleation of unpaired quark matter in the core of protoneutron stars by matching the lepton-rich QCD pressure onto a hadronic equation of state, namely TM1 with trapped neutrinos. Using the inherent dependence of perturbative QCD on the renormalization scale parameter, we provide a measure of the uncertainty in the observables we compute.
a Linux PC Cluster for Lattice QCD with Exact Chiral Symmetry
NASA Astrophysics Data System (ADS)
Chiu, Ting-Wai; Hsieh, Tung-Han; Huang, Chao-Hsi; Huang, Tsung-Ren
A computational system for lattice QCD with overlap Dirac quarks is described. The platform is a home-made Linux PC cluster, built with off-the-shelf components. At present the system constitutes of 64 nodes, with each node consisting of one Pentium 4 processor (1.6/2.0/2.5 GHz), one Gbyte of PC800/1066 RDRAM, one 40/80/120 Gbyte hard disk, and a network card. The computationally intensive parts of our program are written in SSE2 codes. The speed of our system is estimated to be 70 Gflops, and its price/performance ratio is better than $1.0/Mflops for 64-bit (double precision) computations in quenched QCD. We discuss how to optimize its hardware and software for computing propagators of overlap Dirac quarks.
NASA Astrophysics Data System (ADS)
Gupta, Sourendu; Majumdar, Pushan
2018-07-01
We present the results of an effort to accelerate a Rational Hybrid Monte Carlo (RHMC) program for lattice quantum chromodynamics (QCD) simulation for 2 flavors of staggered fermions on multiple Kepler K20X GPUs distributed on different nodes of a Cray XC30. We do not use CUDA but adopt a higher level directive based programming approach using the OpenACC platform. The lattice QCD algorithm is known to be bandwidth bound; our timing results illustrate this clearly, and we discuss how this limits the parallelization gains. We achieve more than a factor three speed-up compared to the CPU only MPI program.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kurth, Thorsten; Pochinsky, Andrew; Sarje, Abhinav
Practitioners of lattice QCD/QFT have been some of the primary pioneer users of the state-of-the-art high-performance-computing systems, and contribute towards the stress tests of such new machines as soon as they become available. As with all aspects of high-performance-computing, I/O is becoming an increasingly specialized component of these systems. In order to take advantage of the latest available high-performance I/O infrastructure, to ensure reliability and backwards compatibility of data files, and to help unify the data structures used in lattice codes, we have incorporated parallel HDF5 I/O into the SciDAC supported USQCD software stack. Here we present the design andmore » implementation of this I/O framework. Our HDF5 implementation outperforms optimized QIO at the 10-20% level and leaves room for further improvement by utilizing appropriate dataset chunking.« less
The Conformal Template and New Perspectives for Quantum Chromodynamics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, Stanley J.; /SLAC
2007-03-06
Conformal symmetry provides a systematic approximation to QCD in both its perturbative and nonperturbative domains. One can use the AdS/CFT correspondence between Anti-de Sitter space and conformal gauge theories to obtain an analytically tractable approximation to QCD in the regime where the QCD coupling is large and constant. For example, there is an exact correspondence between the fifth-dimensional coordinate of AdS space and a specific impact variable which measures the separation of the quark constituents within the hadron in ordinary space-time. This connection allows one to compute the analytic form of the frame-independent light-front wavefunctions of mesons and baryons, themore » fundamental entities which encode hadron properties and allow the computation of exclusive scattering amplitudes. One can also use conformal symmetry as a template for perturbative QCD predictions where the effects of the nonzero beta function can be systematically included in the scale of the QCD coupling. This leads to fixing of the renormalization scale and commensurate scale relations which relate observables without scale or scheme ambiguity. The results are consistent with the renormalization group and the analytic connection of QCD to Abelian theory at N{sub C} {yields} 0. I also discuss a number of novel phenomenological features of QCD. Initial- and .nal-state interactions from gluon-exchange, normally neglected in the parton model, have a profound effect in QCD hard-scattering reactions, leading to leading-twist single-spin asymmetries, diffractive deep inelastic scattering, di.ractive hard hadronic reactions, the breakdown of the Lam Tung relation in Drell-Yan reactions, and nuclear shadowing and non-universal antishadowing--leading-twist physics not incorporated in the light-front wavefunctions of the target computed in isolation. I also discuss tests of hidden color in nuclear wavefunctions, the use of diffraction to materialize the Fock states of a hadronic projectile and test QCD color transparency, nonperturbative antisymmetric sea quark distributions, anomalous heavy quark e.ects, and the unexpected effects of direct higher-twist processes.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Caola, Fabrizio; Melnikov, Kirill; Rontsch, Raoul
We compute the next-to-leading-order QCD corrections to the production of two Z-bosons in the annihilation of two gluons at the LHC. Being enhanced by a large gluon flux, these corrections provide a distinct and, potentially, the dominant part of the N 3LO QCD contributions to Z-pair production in proton collisions. The gg → ZZ annihilation is a loop-induced process that receives the dominant contribution from loops of five light quarks, that are included in our computation in the massless approximation. We find that QCD corrections increase the gg → ZZ production cross section by O(50%–100%) depending on the values ofmore » the renormalization and factorization scales used in the leading-order computation and the collider energy. Furthermore, the large corrections to the gg → ZZ channel increase the pp → ZZ cross section by about 6% to 8%, exceeding the estimated theoretical uncertainty of the recent next-to-next-to-leading-order QCD calculation.« less
Holographic corrections to the Veneziano amplitude
NASA Astrophysics Data System (ADS)
Armoni, Adi; Ireson, Edwin
2017-08-01
We propose a holographic computation of the 2 → 2 meson scattering in a curved string background, dual to a QCD-like theory. We recover the Veneziano amplitude and compute a perturbative correction due to the background curvature. The result implies a small deviation from a linear trajectory, which is a requirement of the UV regime of QCD.
NNLO QCD corrections to associated W H production and H →b b ¯ decay
NASA Astrophysics Data System (ADS)
Caola, Fabrizio; Luisoni, Gionata; Melnikov, Kirill; Röntsch, Raoul
2018-04-01
We present a computation of the next-to-next-to-leading-order (NNLO) QCD corrections to the production of a Higgs boson in association with a W boson at the LHC and the subsequent decay of the Higgs boson into a b b ¯ pair, treating the b quarks as massless. We consider various kinematic distributions and find significant corrections to observables that resolve the Higgs decay products. We also find that a cut on the transverse momentum of the W boson, important for experimental analyses, may have a significant impact on kinematic distributions and radiative corrections. We show that some of these effects can be adequately described by simulating QCD radiation in Higgs boson decays to b quarks using parton showers. We also describe contributions to Higgs decay to a b b ¯ pair that first appear at NNLO and that were not considered in previous fully differential computations. The calculation of NNLO QCD corrections to production and decay sub-processes is carried out within the nested soft-collinear subtraction scheme presented by some of us earlier this year. We demonstrate that this subtraction scheme performs very well, allowing a computation of the coefficient of the second-order QCD corrections at the level of a few per mill.
Dixon, Lance J.; Luo, Ming-xing; Shtabovenko, Vladyslav; ...
2018-03-09
Here, the energy-energy correlation (EEC) between two detectors in e +e – annihilation was computed analytically at leading order in QCD almost 40 years ago, and numerically at next-to-leading order (NLO) starting in the 1980s. We present the first analytical result for the EEC at NLO, which is remarkably simple, and facilitates analytical study of the perturbative structure of the EEC. We provide the expansion of the EEC in the collinear and back-to-back regions through next-to-leading power, information which should aid resummation in these regions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dixon, Lance J.; Luo, Ming-xing; Shtabovenko, Vladyslav
Here, the energy-energy correlation (EEC) between two detectors in e +e – annihilation was computed analytically at leading order in QCD almost 40 years ago, and numerically at next-to-leading order (NLO) starting in the 1980s. We present the first analytical result for the EEC at NLO, which is remarkably simple, and facilitates analytical study of the perturbative structure of the EEC. We provide the expansion of the EEC in the collinear and back-to-back regions through next-to-leading power, information which should aid resummation in these regions.
A computational system for lattice QCD with overlap Dirac quarks
NASA Astrophysics Data System (ADS)
Chiu, Ting-Wai; Hsieh, Tung-Han; Huang, Chao-Hsi; Huang, Tsung-Ren
2003-05-01
We outline the essential features of a Linux PC cluster which is now being developed at National Taiwan University, and discuss how to optimize its hardware and software for lattice QCD with overlap Dirac quarks. At present, the cluster constitutes of 30 nodes, with each node consisting of one Pentium 4 processor (1.6/2.0 GHz), one Gbyte of PC800 RDRAM, one 40/80 Gbyte hard disk, and a network card. The speed of this system is estimated to be 30 Gflops, and its price/performance ratio is better than $1.0/Mflops for 64-bit (double precision) computations in quenched lattice QCD with overlap Dirac quarks.
NASA Astrophysics Data System (ADS)
Somogyi, Gábor
2013-04-01
We finish the definition of a subtraction scheme for computing NNLO corrections to QCD jet cross sections. In particular, we perform the integration of the soft-type contributions to the doubly unresolved counterterms via the method of Mellin-Barnes representations. With these final ingredients in place, the definition of the scheme is complete and the computation of fully differential rates for electron-positron annihilation into two and three jets at NNLO accuracy becomes feasible.
Computational Science: Ensuring America’s Competitiveness
2005-06-01
Supercharging U. S. Innovation & Competitiveness, Washington, D.C. , July 2004. Davies, C. T. H. , et al. , “High-Precision Lattice QCD Confronts Experiment...together to form a class of particles call hadrons (that include protons and neutrons) . For 30 years, researchers in lattice QCD have been trying to use...the basic QCD equations to calculate the properties of hadrons, especially their masses, using numerical lattice gauge theory calculations in order to
NASA Astrophysics Data System (ADS)
Suganuma, H.; Fukushima, M.; Toki, H.
The Table of Contents for the book is as follows: * Preface * Opening Address * Monopole Condensation and Quark Confinement * Dual QCD, Effective String Theory, and Regge Trajectories * Abelian Dominance and Monopole Condensation * Non-Abelian Stokes Theorem and Quark Confinement in QCD * Infrared Region of QCD and Confining Configurations * BRS Quartet Mechanism for Color Confinement * Color Confinement and Quartet Mechanism * Numerical Tests of the Kugo-Ojima Color Confinement Criterion * Monopoles and Confinement in Lattice QCD * SU(2) Lattice Gauge Theory at T > 0 in a Finite Box with Fixed Holonomy * Confining and Dirac Strings in Gluodynamics * Cooling, Monopoles, and Vortices in SU(2) Lattice Gauge Theory * Quark Confinement Physics from Lattice QCD * An (Almost) Perfect Lattice Action for SU(2) and SU(3) Gluodynamics * Vortices and Confinement in Lattice QCD * P-Vortices, Nexuses and Effects of Gribov Copies in the Center Gauges * Laplacian Center Vortices * Center Vortices at Strong Couplings and All Couplings * Simulations in SO(3) × Z(2) Lattice Gauge Theory * Exciting a Vortex - the Cost of Confinement * Instantons in QCD * Deformation of Instanton in External Color Fields * Field Strength Correlators in the Instanton Liquid * Instanton and Meron Physics in Lattice QCD * The Dual Ginzburg-Landau Theory for Confinement and the Role of Instantons * Lattice QCD for Quarks, Gluons and Hadrons * Hadronic Spectral Functions in QCD * Universality and Chaos in Quantum Field Theories * Lattice QCD Study of Three Quark Potential * Probing the QCD Vacuum with Flavour Singlet Objects : η' on the Lattice * Lattice Studies of Quarks and Gluons * Quarks and Hadrons in QCD * Supersymmetric Nonlinear Sigma Models * Chiral Transition and Baryon-number Susceptibility * Light Quark Masses in QCD * Chiral Symmetry of Baryons and Baryon Resonances * Confinement and Bound States in QCD * Parallel Session * Off-diagonal Gluon Mass Generation and Strong Randomness of Off-diagonal Gluon Phase in the Maximally Abelian Gauge * On the Colour Confinement and the Minimal Surface * Glueball Mass and String Tension of SU(2) Gluodynamics from Abelian Monopoles and Strings * Application of the Non-Perturbative Renormalization Group to the Nambu-Jona-Lasinio Model at Finite Temperature and Density * Confining Flux-Tube and Hadrons in QCD * Gauge Symmetry Breakdown due to Dynamical Higgs Scalar * Spatial Structure of Quark Cooper Pairs * New Approach to Axial Coupling Constants in the QCD Sum Rule and Instanton Effects * String Breaking on a Lattice * Bethe-Salpeter Approach for Mesons within the Dual Ginzburg-Landau Theory * Gauge Dependence and Matching Procedure of a Nonrelativistic QCD Boundstate Formalism * A Mathematical Approach to the SU(2)-Quark Confinement * Simulations of Odd Flavors QCD by Hybrid Monte Carlo * Non-Perturbative Renormalization Group Analysis of Dynamical Chiral Symmetry Breaking with Beyond Ladder Contributions * Charmonium Physics in Finite Temperature Lattice QCD * From Meson-Nucleon Scattering to Vector Mesons in Nuclear Matter * Symposium Program * List of Participants
Subtraction method of computing QCD jet cross sections at NNLO accuracy
NASA Astrophysics Data System (ADS)
Trócsányi, Zoltán; Somogyi, Gábor
2008-10-01
We present a general subtraction method for computing radiative corrections to QCD jet cross sections at next-to-next-to-leading order accuracy. The steps needed to set up this subtraction scheme are the same as those used in next-to-leading order computations. However, all steps need non-trivial modifications, which we implement such that that those can be defined at any order in perturbation theory. We give a status report of the implementation of the method to computing jet cross sections in electron-positron annihilation at the next-to-next-to-leading order accuracy.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Anand, Sampurn; Mohanty, Subhendra; Dey, Ujjal Kumar, E-mail: sampurn@prl.res.in, E-mail: ujjal@cts.iitkgp.ernet.in, E-mail: mohanty@prl.res.in
Cosmological phase transitions can be a source of Stochastic Gravitational Wave (SGW) background. Apart from the dynamics of the phase transition, the characteristic frequency and the fractional energy density Ω{sub gw} of the SGW depends upon the temperature of the transition. In this article, we compute the SGW spectrum in the light of QCD equation of state provided by the lattice results. We find that the inclusion of trace anomaly from lattice QCD, enhances the SGW signal generated during QCD phase transition by ∼ 50% and the peak frequency of the QCD era SGW are shifted higher by ∼ 25%more » as compared to the earlier estimates without trace anomaly. This result is extremely significant for testing the phase transition dynamics near QCD epoch.« less
Algorithms for Disconnected Diagrams in Lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gambhir, Arjun Singh; Stathopoulos, Andreas; Orginos, Konstantinos
2016-11-01
Computing disconnected diagrams in Lattice QCD (operator insertion in a quark loop) entails the computationally demanding problem of taking the trace of the all to all quark propagator. We first outline the basic algorithm used to compute a quark loop as well as improvements to this method. Then, we motivate and introduce an algorithm based on the synergy between hierarchical probing and singular value deflation. We present results for the chiral condensate using a 2+1-flavor clover ensemble and compare estimates of the nucleon charges with the basic algorithm.
Three-particle N π π state contribution to the nucleon two-point function in lattice QCD
NASA Astrophysics Data System (ADS)
Bär, Oliver
2018-05-01
The three-particle N π π state contribution to the QCD two-point function of standard nucleon interpolating fields is computed to leading order in chiral perturbation theory. Using the experimental values for two low-energy coefficients, the impact of this contribution on lattice QCD calculations of the nucleon mass is estimated. The impact is found to be at the per mille level at most and negligible in practice.
Hidden axion dark matter decaying through mixing with QCD axion and the 3.5 keV X-ray line
DOE Office of Scientific and Technical Information (OSTI.GOV)
Higaki, Tetsutaro; Kitajima, Naoya; Takahashi, Fuminobu, E-mail: thigaki@post.kek.jp, E-mail: kitajima@tuhep.phys.tohoku.ac.jp, E-mail: fumi@tuhep.phys.tohoku.ac.jp
2014-12-01
Hidden axions may be coupled to the standard model particles through a kinetic or mass mixing with QCD axion. We study a scenario in which a hidden axion constitutes a part of or the whole of dark matter and decays into photons through the mixing, explaining the 3.5 keV X-ray line signal. Interestingly, the required long lifetime of the hidden axion dark matter can be realized for the QCD axion decay constant at an intermediate scale, if the mixing is sufficiently small. In such a two component dark matter scenario, the primordial density perturbations of the hidden axion can bemore » highly non-Gaussian, leading to a possible dispersion in the X-ray line strength from various galaxy clusters and near-by galaxies. We also discuss how the parallel and orthogonal alignment of two axions affects their couplings to gauge fields. In particular, the QCD axion decay constant can be much larger than the actual Peccei-Quinn symmetry breaking.« less
Drell-Yan Lepton pair production at NNLO QCD with parton showers
Hoeche, Stefan; Li, Ye; Prestel, Stefan
2015-04-13
We present a simple approach to combine NNLO QCD calculations and parton showers, based on the UNLOPS technique. We apply the method to the computation of Drell-Yan lepton-pair production at the Large Hadron Collider. We comment on possible improvements and intrinsic uncertainties.
Parton distributions and lattice QCD calculations: A community white paper
NASA Astrophysics Data System (ADS)
Lin, Huey-Wen; Nocera, Emanuele R.; Olness, Fred; Orginos, Kostas; Rojo, Juan; Accardi, Alberto; Alexandrou, Constantia; Bacchetta, Alessandro; Bozzi, Giuseppe; Chen, Jiunn-Wei; Collins, Sara; Cooper-Sarkar, Amanda; Constantinou, Martha; Del Debbio, Luigi; Engelhardt, Michael; Green, Jeremy; Gupta, Rajan; Harland-Lang, Lucian A.; Ishikawa, Tomomi; Kusina, Aleksander; Liu, Keh-Fei; Liuti, Simonetta; Monahan, Christopher; Nadolsky, Pavel; Qiu, Jian-Wei; Schienbein, Ingo; Schierholz, Gerrit; Thorne, Robert S.; Vogelsang, Werner; Wittig, Hartmut; Yuan, C.-P.; Zanotti, James
2018-05-01
In the framework of quantum chromodynamics (QCD), parton distribution functions (PDFs) quantify how the momentum and spin of a hadron are divided among its quark and gluon constituents. Two main approaches exist to determine PDFs. The first approach, based on QCD factorization theorems, realizes a QCD analysis of a suitable set of hard-scattering measurements, often using a variety of hadronic observables. The second approach, based on first-principle operator definitions of PDFs, uses lattice QCD to compute directly some PDF-related quantities, such as their moments. Motivated by recent progress in both approaches, in this document we present an overview of lattice-QCD and global-analysis techniques used to determine unpolarized and polarized proton PDFs and their moments. We provide benchmark numbers to validate present and future lattice-QCD calculations and we illustrate how they could be used to reduce the PDF uncertainties in current unpolarized and polarized global analyses. This document represents a first step towards establishing a common language between the two communities, to foster dialogue and to further improve our knowledge of PDFs.
NASA Astrophysics Data System (ADS)
Bicudo, P.; Cardoso, M.; Oliveira, O.; Silva, P. J.
2017-10-01
We revisit the static potential for the Q Q Q ¯Q ¯ system using SU(3) lattice simulations, studying both the color singlets' ground state and first excited state. We consider geometries where the two static quarks and the two antiquarks are at the corners of rectangles of different sizes. We analyze the transition between a tetraquark system and a two-meson system with a two by two correlator matrix. We compare the potentials computed with quenched QCD and with dynamical quarks. We also compare our simulations with the results of previous studies and analyze quantitatively fits of our results with Ansätze inspired in the string flip-flop model and in its possible color excitations.
Spectroscopy of triply charmed baryons from lattice QCD
Padmanath, M.; Edwards, Robert G.; Mathur, Nilmani; ...
2014-10-14
The spectrum of excitations of triply-charmed baryons is computed using lattice QCD including dynamical light quark fields. The spectrum obtained has baryonic states with well-defined total spin up to 7/2 and the low-lying states closely resemble the expectation from models with an SU(6) x O(3) symmetry. As a result, energy splittings between extracted states, including those due to spin-orbit coupling in the heavy quark limit are computed and compared against data at other quark masses.
HELAC-Onia 2.0: An upgraded matrix-element and event generator for heavy quarkonium physics
NASA Astrophysics Data System (ADS)
Shao, Hua-Sheng
2016-01-01
We present an upgraded version (denoted as version 2.0) of the program HELAC-ONIA for the automated computation of heavy-quarkonium helicity amplitudes within non-relativistic QCD framework. The new code has been designed to include many new and useful features for practical phenomenological simulations. It is designed for job submissions under cluster environment for parallel computations via PYTHON scripts. We have interfaced HELAC-ONIA to the parton shower Monte Carlo programs PYTHIA 8 and QEDPS to take into account the parton-shower effects. Moreover, the decay module guarantees that the program can perform the spin-entangled (cascade-)decay of heavy quarkonium after its generation. We have also implemented a reweighting method to automatically estimate the uncertainties from renormalization and/or factorization scales as well as parton-distribution functions to weighted or unweighted events. A further update is the possibility to generate one-dimensional or two-dimensional plots encoded in the analysis files on the fly. Some dedicated examples are given at the end of the writeup.
Higgs bosons with large transverse momentum at the LHC
NASA Astrophysics Data System (ADS)
Kudashkin, Kirill; Lindert, Jonas M.; Melnikov, Kirill; Wever, Christopher
2018-07-01
We compute the next-to-leading order QCD corrections to the production of Higgs bosons with large transverse momentum p⊥ ≫ 2mt at the LHC. To accomplish this, we combine the two-loop amplitudes for processes gg → Hg, qg → Hq and q q bar → Hg, recently computed in the approximation of nearly massless top quarks, with the numerical calculation of the squared one-loop amplitudes for gg → Hgg, qg → Hqg and q q bar → Hgg processes. The latter computation is performed with OpenLoops. We find that the QCD corrections to the Higgs transverse momentum distribution at very high p⊥ are large but quite similar to the QCD corrections obtained for point-like Hgg coupling. Our result removes one of the largest sources of theoretical uncertainty in the description of high-p⊥ Higgs boson production and opens a way to use the high-p⊥ region to search for physics beyond the Standard Model.
Hadron electric polarizability from lattice QCD
NASA Astrophysics Data System (ADS)
Alexandru, Andrei; Lujan, Michael; Freeman, Walter; Lee, Frank
2015-04-01
Electric polarizability measures the ability of the electric field to deform a particle. Experimentally, electric and magnetic polarizabilities can be measured in Compton scattering experiments. To compute these quantities theoretically we need to understand the internal structure of the scatterer and the dynamics of its constituents. For hadrons - bound stated of quarks and gluons - this is a very difficult problem. Lattice QCD can be used to compute the polarizabilities directly in terms of quark and gluons degrees of freedom. In this talk we focus on the neutron. We present results for the electric polarizability for two different quark masses, light enough to connect to chiral perturbation theory. These are currently the lightest quark masses used in lattice QCD polarizability studies. For each pion mass we compute the polarizability at four different volumes and perform an infinite volume extrapolation. For one ensemble, we also discuss the effect of turning on the coupling between the background field and the sea quarks. We compare our results to chiral perturbation theory expectations.
Search for an Electric Dipole Moment (EDM) of 199Hg
NASA Astrophysics Data System (ADS)
Heckel, Blayne
2017-04-01
The observation of a non-zero EDM of an atom or elementary particle, at current levels of experimental sensitivity, would imply CP violation beyond the CKM matrix of the standard model of particle physics. Additional sources of CP violation have been proposed to help explain the excess of matter over anti-matter in our universe and the magnitude of ΘQCD, the strength of CP violation in the strong interaction, remains unknown. We have recently completed a set of measurements on the EDM of 199Hg, sensitive to both new sources of CP violation and ΘQCD. The experiment compares the phase accumulated by precessing Hg spins in vapor cells with electric fields parallel and anti-parallel to a common magnetic field. Our new result represents a factor of 5 improvement over previous results. A description of the EDM experiment, data, systematic error considerations will be presented. This work was supported by NSF Grant No. 1306743 and by the DOE Office of Nuclear Physics under Award No. DE-FG02-97ER41020.
C P -odd sector and θ dynamics in holographic QCD
NASA Astrophysics Data System (ADS)
Areán, Daniel; Iatrakis, Ioannis; Järvinen, Matti; Kiritsis, Elias
2017-07-01
The holographic model of V-QCD is used to analyze the physics of QCD in the Veneziano large-N limit. An unprecedented analysis of the C P -odd physics is performed going beyond the level of effective field theories. The structure of holographic saddle points at finite θ is determined, as well as its interplay with chiral symmetry breaking. Many observables (vacuum energy and higher-order susceptibilities, singlet and nonsinglet masses and mixings) are computed as functions of θ and the quark mass m . Wherever applicable the results are compared to those of chiral Lagrangians, finding agreement. In particular, we recover the Witten-Veneziano formula in the small x →0 limit, we compute the θ dependence of the pion mass, and we derive the hyperscaling relation for the topological susceptibility in the conformal window in terms of the quark mass.
Blum, Thomas; Chowdhury, Saumitra; Hayakawa, Masashi; ...
2015-01-07
The form factor that yields the light-by-light scattering contribution to the muon anomalous magnetic moment is computed in lattice QCD+QED and QED. A non-perturbative treatment of QED is used and is checked against perturbation theory. The hadronic contribution is calculated for unphysical quark and muon masses, and only the diagram with a single quark loop is computed. Statistically significant signals are obtained. Initial results appear promising, and the prospect for a complete calculation with physical masses and controlled errors is discussed.
Parton distributions and lattice QCD calculations: A community white paper
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lin, Huey-Wen; Nocera, Emanuele R.; Olness, Fred
In the framework of quantum chromodynamics (QCD), parton distribution functions (PDFs) quantify how the momentum and spin of a hadron are divided among its quark and gluon constituents. Two main approaches exist to determine PDFs. The first approach, based on QCD factorization theorems, realizes a QCD analysis of a suitable set of hard-scattering measurements, often using a variety of hadronic observables. The second approach, based on first-principle operator definitions of PDFs, uses lattice QCD to compute directly some PDF-related quantities, such as their moments. Motivated by recent progress in both approaches, in this paper we present an overview of lattice-QCDmore » and global-analysis techniques used to determine unpolarized and polarized proton PDFs and their moments. We provide benchmark numbers to validate present and future lattice-QCD calculations and we illustrate how they could be used to reduce the PDF uncertainties in current unpolarized and polarized global analyses. Finally, this document represents a first step towards establishing a common language between the two communities, to foster dialogue and to further improve our knowledge of PDFs.« less
Parton distributions and lattice QCD calculations: A community white paper
Lin, Huey-Wen; Nocera, Emanuele R.; Olness, Fred; ...
2018-01-31
In the framework of quantum chromodynamics (QCD), parton distribution functions (PDFs) quantify how the momentum and spin of a hadron are divided among its quark and gluon constituents. Two main approaches exist to determine PDFs. The first approach, based on QCD factorization theorems, realizes a QCD analysis of a suitable set of hard-scattering measurements, often using a variety of hadronic observables. The second approach, based on first-principle operator definitions of PDFs, uses lattice QCD to compute directly some PDF-related quantities, such as their moments. Motivated by recent progress in both approaches, in this paper we present an overview of lattice-QCDmore » and global-analysis techniques used to determine unpolarized and polarized proton PDFs and their moments. We provide benchmark numbers to validate present and future lattice-QCD calculations and we illustrate how they could be used to reduce the PDF uncertainties in current unpolarized and polarized global analyses. Finally, this document represents a first step towards establishing a common language between the two communities, to foster dialogue and to further improve our knowledge of PDFs.« less
Quark structure of static correlators in high temperature QCD
NASA Astrophysics Data System (ADS)
Bernard, Claude; DeGrand, Thomas A.; DeTar, Carleton; Gottlieb, Steven; Krasnitz, A.; Ogilvie, Michael C.; Sugar, R. L.; Toussaint, D.
1992-07-01
We present results of numerical simulations of quantum chromodynamics at finite temperature with two flavors of Kogut-Susskind quarks on the Intel iPSC/860 parallel processor. We investigate the properties of the objects whose exchange gives static screening lengths by reconstructing their correlated quark-antiquark structure.
Direct CP asymmetry in D → π-π+ and D → K-K+ in QCD-based approach
NASA Astrophysics Data System (ADS)
Khodjamirian, Alexander; Petrov, Alexey A.
2017-11-01
We present the first QCD-based calculation of hadronic matrix elements with penguin topology determining direct CP-violating asymmetries in D0 →π-π+ and D0 →K-K+ nonleptonic decays. The method is based on the QCD light-cone sum rules and does not rely on any model-inspired amplitude decomposition, instead leaning heavily on quark-hadron duality. We provide a Standard Model estimate of the direct CP-violating asymmetries in both pion and kaon modes and their difference and comment on further improvements of the presented computation.
Associated Higgs-W-boson production at hadron colliders: a fully exclusive QCD calculation at NNLO.
Ferrera, Giancarlo; Grazzini, Massimiliano; Tramontano, Francesco
2011-10-07
We consider QCD radiative corrections to standard model Higgs-boson production in association with a W boson in hadron collisions. We present a fully exclusive calculation up to next-to-next-to-leading order (NNLO) in QCD perturbation theory. To perform this NNLO computation, we use a recently proposed version of the subtraction formalism. Our calculation includes finite-width effects, the leptonic decay of the W boson with its spin correlations, and the decay of the Higgs boson into a bb pair. We present selected numerical results at the Tevatron and the LHC.
Tetraquark resonances computed with static lattice QCD potentials and scattering theory
NASA Astrophysics Data System (ADS)
Bicudo, Pedro; Cardoso, Marco; Peters, Antje; Pflaumer, Martin; Wagner, Marc
2018-03-01
We study tetraquark resonances with lattice QCD potentials computed for two static quarks and two dynamical quarks, the Born-Oppenheimer approximation and the emergent wave method of scattering theory. As a proof of concept we focus on systems with isospin I = 0, but consider different relative angular momenta l of the heavy b quarks. We compute the phase shifts and search for S and T matrix poles in the second Riemann sheet. We predict a new tetraquark resonance for l = 1, decaying into two B mesons, with quantum numbers I(JP) = 0(1-), mass m = 10576-4+4 MeV and decay width Γ = 112-103+90 MeV.
Hardware and software status of QCDOC
NASA Astrophysics Data System (ADS)
Boyle, P. A.; Chen, D.; Christ, N. H.; Clark, M.; Cohen, S. D.; Cristian, C.; Dong, Z.; Gara, A.; Joó, B.; Jung, C.; Kim, C.; Levkova, L.; Liao, X.; Liu, G.; Mawhinney, R. D.; Ohta, S.; Petrov, K.; Wettig, T.; Yamaguchi, A.
2004-03-01
QCDOC is a massively parallel supercomputer whose processing nodes are based on an application-specific integrated circuit (ASIC). This ASIC was custom-designed so that crucial lattice QCD kernels achieve an overall sustained performance of 50% on machines with several 10,000 nodes. This strong scalability, together with low power consumption and a price/performance ratio of $1 per sustained MFlops, enable QCDOC to attack the most demanding lattice QCD problems. The first ASICs became available in June of 2003, and the testing performed so far has shown all systems functioning according to specification. We review the hardware and software status of QCDOC and present performance figures obtained in real hardware as well as in simulation.
Dissociation of heavy quarkonium in hot QCD medium in a quasiparticle model
NASA Astrophysics Data System (ADS)
Agotiya, Vineet Kumar; Chandra, Vinod; Jamal, M. Yousuf; Nilima, Indrani
2016-11-01
Following a recent work on the effective description of the equations of state for hot QCD obtained from a hard thermal loop expression for the gluon self-energy, in terms of the quasigluons and quasiquarks and antiquarks with respective effective fugacities, the dissociation process of heavy quarkonium in hot QCD medium has been investigated. This has been done by investigating the medium modification to a heavy quark potential. The medium-modified potential has a quite different form (a long-range Coulomb tail in addition to the usual Yukawa term) in contrast to the usual picture of Debye screening. The flavor dependence binding energies of the heavy quarkonia states and the dissociation temperature have been obtained by employing the Debye mass for pure gluonic and full QCD case computed employing the quasiparticle picture. Thus, estimated dissociation patterns of the charmonium and bottomonium states, considering Debye mass from different approaches in the pure gluonic case and full QCD, have shown good agreement with the other potential model studies.
Nuclear reactions from lattice QCD
Briceño, Raúl A.; Davoudi, Zohreh; Luu, Thomas C.
2015-01-13
In this study, one of the overarching goals of nuclear physics is to rigorously compute properties of hadronic systems directly from the fundamental theory of strong interactions, Quantum Chromodynamics (QCD). In particular, the hope is to perform reliable calculations of nuclear reactions which will impact our understanding of environments that occur during big bang nucleosynthesis, the evolution of stars and supernovae, and within nuclear reactors and high energy/density facilities. Such calculations, being truly ab initio, would include all two-nucleon and three- nucleon (and higher) interactions in a consistent manner. Currently, lattice QCD provides the only reliable option for performing calculationsmore » of some of the low-energy hadronic observables. With the aim of bridging the gap between lattice QCD and nuclear many-body physics, the Institute for Nuclear Theory held a workshop on Nuclear Reactions from Lattice QCD on March 2013. In this review article, we report on the topics discussed in this workshop and the path planned to move forward in the upcoming years.« less
Analytical Computation of Energy-Energy Correlation at Next-to-Leading Order in QCD
NASA Astrophysics Data System (ADS)
Dixon, Lance J.; Luo, Ming-xing; Shtabovenko, Vladyslav; Yang, Tong-Zhi; Zhu, Hua Xing
2018-03-01
The energy-energy correlation (EEC) between two detectors in e+e- annihilation was computed analytically at leading order in QCD almost 40 years ago, and numerically at next-to-leading order (NLO) starting in the 1980s. We present the first analytical result for the EEC at NLO, which is remarkably simple, and facilitates analytical study of the perturbative structure of the EEC. We provide the expansion of the EEC in the collinear and back-to-back regions through next-to-leading power, information which should aid resummation in these regions.
Second-order QCD effects in Higgs boson production through vector boson fusion
NASA Astrophysics Data System (ADS)
Cruz-Martinez, J.; Gehrmann, T.; Glover, E. W. N.; Huss, A.
2018-06-01
We compute the factorising second-order QCD corrections to the electroweak production of a Higgs boson through vector boson fusion. Our calculation is fully differential in the kinematics of the Higgs boson and of the final state jets, and uses the antenna subtraction method to handle infrared singular configurations in the different parton-level contributions. Our results allow us to reassess the impact of the next-to-leading order (NLO) QCD corrections to electroweak Higgs-plus-three-jet production and of the next-to-next-to-leading order (NNLO) QCD corrections to electroweak Higgs-plus-two-jet production. The NNLO corrections are found to be limited in magnitude to around ± 5% and are uniform in several of the kinematical variables, displaying a kinematical dependence only in the transverse momenta and rapidity separation of the two tagging jets.
Disconnected Diagrams in Lattice QCD
NASA Astrophysics Data System (ADS)
Gambhir, Arjun Singh
In this work, we present state-of-the-art numerical methods and their applications for computing a particular class of observables using lattice quantum chromodynamics (Lattice QCD), a discretized version of the fundamental theory of quarks and gluons. These observables require calculating so called "disconnected diagrams" and are important for understanding many aspects of hadron structure, such as the strange content of the proton. We begin by introducing the reader to the key concepts of Lattice QCD and rigorously define the meaning of disconnected diagrams through an example of the Wick contractions of the nucleon. Subsequently, the calculation of observables requiring disconnected diagrams is posed as the computationally challenging problem of finding the trace of the inverse of an incredibly large, sparse matrix. This is followed by a brief primer of numerical sparse matrix techniques that overviews broadly used methods in Lattice QCD and builds the background for the novel algorithm presented in this work. We then introduce singular value deflation as a method to improve convergence of trace estimation and analyze its effects on matrices from a variety of fields, including chemical transport modeling, magnetohydrodynamics, and QCD. Finally, we apply this method to compute observables such as the strange axial charge of the proton and strange sigma terms in light nuclei. The work in this thesis is innovative for four reasons. First, we analyze the effects of deflation with a model that makes qualitative predictions about its effectiveness, taking only the singular value spectrum as input, and compare deflated variance with different types of trace estimator noise. Second, the synergy between probing methods and deflation is investigated both experimentally and theoretically. Third, we use the synergistic combination of deflation and a graph coloring algorithm known as hierarchical probing to conduct a lattice calculation of light disconnected matrix elements of the nucleon at two different values of the lattice spacing. Finally, we employ these algorithms to do a high-precision study of strange sigma terms in light nuclei; to our knowledge this is the first calculation of its kind from Lattice QCD.
Disconnected Diagrams in Lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gambhir, Arjun
In this work, we present state-of-the-art numerical methods and their applications for computing a particular class of observables using lattice quantum chromodynamics (Lattice QCD), a discretized version of the fundamental theory of quarks and gluons. These observables require calculating so called \\disconnected diagrams" and are important for understanding many aspects of hadron structure, such as the strange content of the proton. We begin by introducing the reader to the key concepts of Lattice QCD and rigorously define the meaning of disconnected diagrams through an example of the Wick contractions of the nucleon. Subsequently, the calculation of observables requiring disconnected diagramsmore » is posed as the computationally challenging problem of finding the trace of the inverse of an incredibly large, sparse matrix. This is followed by a brief primer of numerical sparse matrix techniques that overviews broadly used methods in Lattice QCD and builds the background for the novel algorithm presented in this work. We then introduce singular value deflation as a method to improve convergence of trace estimation and analyze its effects on matrices from a variety of fields, including chemical transport modeling, magnetohydrodynamics, and QCD. Finally, we apply this method to compute observables such as the strange axial charge of the proton and strange sigma terms in light nuclei. The work in this thesis is innovative for four reasons. First, we analyze the effects of deflation with a model that makes qualitative predictions about its effectiveness, taking only the singular value spectrum as input, and compare deflated variance with different types of trace estimator noise. Second, the synergy between probing methods and deflation is investigated both experimentally and theoretically. Third, we use the synergistic combination of deflation and a graph coloring algorithm known as hierarchical probing to conduct a lattice calculation of light disconnected matrix elements of the nucleon at two different values of the lattice spacing. Finally, we employ these algorithms to do a high-precision study of strange sigma terms in light nuclei; to our knowledge this is the first calculation of its kind from Lattice QCD.« less
Computing Properties of Hadrons, Nuclei and Nuclear Matter from Quantum Chromodynamics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Savage, Martin J.
This project was part of a coordinated software development effort which the nuclear physics lattice QCD community pursues in order to ensure that lattice calculations can make optimal use of present, and forthcoming leadership-class and dedicated hardware, including those of the national laboratories, and prepares for the exploitation of future computational resources in the exascale era. The UW team improved and extended software libraries used in lattice QCD calculations related to multi-nucleon systems, enhanced production running codes related to load balancing multi-nucleon production on large-scale computing platforms, and developed SQLite (addressable database) interfaces to efficiently archive and analyze multi-nucleon datamore » and developed a Mathematica interface for the SQLite databases.« less
Instanton liquid properties from lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Athenodorou, A.; Boucaud, Philippe; De Soto, F.
Here, we examined the instanton contribution to the QCD configurations generated from lattice QCD for N F = 0, N F = 2 + 1 and N F = 2 + 1 + 1 dynamical quark flavors from two different and complementary approaches. First via the use of Gradient flow, we computed instanton liquid properties using an algorithm to localize instantons in the gauge field configurations and studied their evolution with flow time. Then, the analysis of the running at low momenta of gluon Green's functions serves as an independent confirmation of the instanton density which can also be derivedmore » without the use of the Gradient flow.« less
Instanton liquid properties from lattice QCD
Athenodorou, A.; Boucaud, Philippe; De Soto, F.; ...
2018-02-22
Here, we examined the instanton contribution to the QCD configurations generated from lattice QCD for N F = 0, N F = 2 + 1 and N F = 2 + 1 + 1 dynamical quark flavors from two different and complementary approaches. First via the use of Gradient flow, we computed instanton liquid properties using an algorithm to localize instantons in the gauge field configurations and studied their evolution with flow time. Then, the analysis of the running at low momenta of gluon Green's functions serves as an independent confirmation of the instanton density which can also be derivedmore » without the use of the Gradient flow.« less
A Fast Algorithm for Lattice Hyperonic Potentials
NASA Astrophysics Data System (ADS)
Nemura, Hidekatsu; Aoki, Sinya; Doi, Takumi; Gongyo, Shinya; Hatsuda, Tetsuo; Ikeda, Yoichi; Inoue, Takashi; Iritani, Takumi; Ishii, Noriyoshi; Miyamoto, Takaya; Murano, Keiko; Sasaki, Kenji
We describe an efficient algorithm to compute a large number of baryon-baryon interactions from NN to ΞΞ by means of HAL QCD method, which lays the groundwork for the nearly physical point lattice QCD calculation with volume (96a)4 ≈ (8.2 fm)4. Preliminary results of ΛN potential calculated with quark masses corresponding to (mπ, mK) ≈ (146,525) MeV are presented.
Massive photons: An infrared regularization scheme for lattice QCD + QED
Endres, Michael G.; Shindler, Andrea; Tiburzi, Brian C.; ...
2016-08-10
The commonly adopted approach for including electromagnetic interactions in lattice QCD simulations relies on using finite volume as the infrared regularization for QED. The long-range nature of the electromagnetic interaction, however, implies that physical quantities are susceptible to power-law finite volume corrections, which must be removed by performing costly simulations at multiple lattice volumes, followed by an extrapolation to the infinite volume limit. In this work, we introduce a photon mass as an alternative means for gaining control over infrared effects associated with electromagnetic interactions. We present findings for hadron mass shifts due to electromagnetic interactions (i.e., for the proton,more » neutron, charged and neutral kaon) and corresponding mass splittings, and compare the results with those obtained from conventional QCD+QED calculations. Results are reported for numerical studies of three flavor electroquenched QCD using ensembles corresponding to 800 MeV pions, ensuring that the only appreciable volume corrections arise from QED effects. The calculations are performed with three lattice volumes with spatial extents ranging from 3.4 - 6.7 fm. As a result, we find that for equal computing time (not including the generation of the lattice configurations), the electromagnetic mass shifts can be extracted from computations on a single (our smallest) lattice volume with comparable or better precision than the conventional approach.« less
Longitudinal conductivity in strong magnetic field in perturbative QCD: Complete leading order
NASA Astrophysics Data System (ADS)
Hattori, Koichi; Li, Shiyong; Satow, Daisuke; Yee, Ho-Ung
2017-04-01
We compute the longitudinal electrical conductivity in the presence of a strong background magnetic field in complete leading order of perturbative QCD, based on the assumed hierarchy of scales αse B ≪(mq2,T2)≪e B . We formulate an effective kinetic theory of lowest Landau level quarks with the leading order QCD collision term arising from 1-to-2 processes that become possible due to 1 +1 dimensional Landau level kinematics. In the small mq/T ≪1 regime, the longitudinal conductivity behaves as σz z˜e2(e B )T /(αsmq2log (T /mq)) , where the quark mass dependence can be understood from the chiral anomaly with the axial charge relaxation provided by a finite quark mass mq. We also present parametric estimates for the longitudinal and transverse "color conductivities" in the presence of the strong magnetic field, by computing dominant damping rates for quarks and gluons that are responsible for color charge transportation. We observe that the longitudinal color conductivity is enhanced by the strong magnetic field, which implies that the sphaleron transition rate in perturbative QCD is suppressed by the strong magnetic field due to the enhanced Lenz's law in color field dynamics.
Exotic and excited-state radiative transitions in charmonium from lattice QCD
Dudek, Jozef J.; Edwards, Robert G.; Thomas, Christopher E.
2009-05-01
We compute, for the first time using lattice QCD methods, radiative transition rates involving excited charmonium states, states of high spin and exotics. Utilizing a large basis of interpolating fields we are able to project out various excited state contributions to three-point correlators computed on quenched anisotropic lattices. In the first lattice QCD calculation of the exoticmore » $$1^{-+}$$ $$\\eta_{c1}$$ radiative decay, we find a large partial width $$\\Gamma(\\eta_{c1} \\to J/\\psi \\gamma) \\sim 100 \\,\\mathrm{keV}$$. We find clear signals for electric dipole and magnetic quadrupole transition form factors in $$\\chi_{c2} \\to J/\\psi \\gamma$$, calculated for the first time in this framework, and study transitions involving excited $$\\psi$$ and $$\\chi_{c1,2}$$ states. We calculate hindered magnetic dipole transition widths without the sensitivity to assumptions made in model studies and find statistically significant signals, including a non-exotic vector hybrid candidate $Y_{\\mathrm{hyb?}} \\to \\et« less
Determination of the QCD Λ Parameter and the Accuracy of Perturbation Theory at High Energies.
Dalla Brida, Mattia; Fritzsch, Patrick; Korzec, Tomasz; Ramos, Alberto; Sint, Stefan; Sommer, Rainer
2016-10-28
We discuss the determination of the strong coupling α_{MS[over ¯]}(m_{Z}) or, equivalently, the QCD Λ parameter. Its determination requires the use of perturbation theory in α_{s}(μ) in some scheme s and at some energy scale μ. The higher the scale μ, the more accurate perturbation theory becomes, owing to asymptotic freedom. As one step in our computation of the Λ parameter in three-flavor QCD, we perform lattice computations in a scheme that allows us to nonperturbatively reach very high energies, corresponding to α_{s}=0.1 and below. We find that (continuum) perturbation theory is very accurate there, yielding a 3% error in the Λ parameter, while data around α_{s}≈0.2 are clearly insufficient to quote such a precision. It is important to realize that these findings are expected to be generic, as our scheme has advantageous properties regarding the applicability of perturbation theory.
Lattice QCD Calculation of Hadronic Light-by-Light Scattering.
Green, Jeremy; Gryniuk, Oleksii; von Hippel, Georg; Meyer, Harvey B; Pascalutsa, Vladimir
2015-11-27
We perform a lattice QCD calculation of the hadronic light-by-light scattering amplitude in a broad kinematical range. At forward kinematics, the results are compared to a phenomenological analysis based on dispersive sum rules for light-by-light scattering. The size of the pion pole contribution is investigated for momenta of typical hadronic size. The presented numerical methods can be used to compute the hadronic light-by-light contribution to the anomalous magnetic moment of the muon. Our calculations are carried out in two-flavor QCD with the pion mass in the range of 270-450 MeV and contain so far only the diagrams with fully connected quark lines.
Scientific and personal recollections of Roberto Petronzio
NASA Astrophysics Data System (ADS)
Parisi, Giorgio
2018-03-01
This paper aims to recall some of the main contributions of Roberto Petronzio to physics, with a particular regard to the period we have been working together. His seminal contributions cover an extremely wide range of topics: the foundation of the perturbative approach to QCD, various aspects of weak interaction theory, from basic questions (e.g. the mass of the Higgs) to lattice weak interaction, lattice QCD from the beginning to most recent computations.
Wilson Dslash Kernel From Lattice QCD Optimization
DOE Office of Scientific and Technical Information (OSTI.GOV)
Joo, Balint; Smelyanskiy, Mikhail; Kalamkar, Dhiraj D.
2015-07-01
Lattice Quantum Chromodynamics (LQCD) is a numerical technique used for calculations in Theoretical Nuclear and High Energy Physics. LQCD is traditionally one of the first applications ported to many new high performance computing architectures and indeed LQCD practitioners have been known to design and build custom LQCD computers. Lattice QCD kernels are frequently used as benchmarks (e.g. 168.wupwise in the SPEC suite) and are generally well understood, and as such are ideal to illustrate several optimization techniques. In this chapter we will detail our work in optimizing the Wilson-Dslash kernels for Intel Xeon Phi, however, as we will show themore » technique gives excellent performance on regular Xeon Architecture as well.« less
Baryon magnetic moments: Symmetries and relations
NASA Astrophysics Data System (ADS)
Parreño, Assumpta; Savage, Martin J.; Tiburzi, Brian C.; Wilhelm, Jonas; Chang, Emmanuel; Detmold, William; Orginos, Kostas
2018-03-01
Magnetic moments of the octet baryons are computed using lattice QCD in background magnetic fields, including the first treatment of the magnetically coupled ∑0- ⋀ system. Although the computations are performed for relatively large values of the up and down quark masses, we gain new insight into the symmetries and relations between magnetic moments by working at a three-flavor mass-symmetric point. While the spinflavor symmetry in the large Nc limit of QCD is shared by the naïve constituent quark model, we find instances where quark model predictions are considerably favored over those emerging in the large Nc limit. We suggest further calculations that would shed light on the curious patterns of baryon magnetic moments.
NASA Astrophysics Data System (ADS)
Susskind, J.; Rosenberg, R. I.
2016-12-01
The GEOS-5 Data Assimilation System (DAS) generates a global analysis every six hours by combining the previous six hour forecast for that time period with contemporaneous observations. These observations include in-situ observations as well as those taken by satellite borne instruments, such as AIRS/AMSU on EOS Aqua and CrIS/ATMS on S-NPP. Operational data assimilation methodology assimilates observed channel radiances Ri for IR sounding instruments such as AIRS and CrIS, but only for those channels i in a given scene whose radiances are thought to be unaffected by clouds. A limitation of this approach is that radiances in most tropospheric sounding channels are affected by clouds under partial cloud cover conditions, which occurs most of the time. The AIRS Science Team Version-6 retrieval algorithm generates cloud cleared radiances (CCR's) for each channel in a given scene, which represent the radiances AIRS would have observed if the scene were cloud free, and then uses them to determine quality controlled (QC'd) temperature profiles T(p) under all cloud conditions. There are potential advantages to assimilate either AIRS QC'd CCR's or QC'd T(p) instead of Ri in that the spatial coverage of observations is greater under partial cloud cover. We tested these two alternate data assimilation approaches by running three parallel data assimilation experiments over different time periods using GEOS-5. Experiment 1 assimilated all observations as done operationally, Experiment 2 assimilated QC'd values of AIRS CCRs in place of AIRS radiances, and Experiment 3 assimilated QC'd values of T(p) in place of observed radiances. Assimilation of QC'd AIRS T(p) resulted in significant improvement in seven day forecast skill compared to assimilation of CCR's or assimilation of observed radiances, especially in the Southern Hemisphere Extra-tropics.
J/ψ production and suppression in high-energy proton-nucleus collisions
Ma, Yan -Qing; Venugopalan, Raju; Zhang, Hong -Fei
2015-10-02
In this study, we apply a color glass condensate+nonrelativistic QCD (CGC+NRQCD) framework to compute J/ψ production in deuteron-nucleus collisions at RHIC and proton-nucleus collisions at the LHC. Our results match smoothly at high p⊥ to a next-to-leading order perturbative QCD+NRQCD computation. Excellent agreement is obtained for p⊥ spectra at the RHIC and LHC for central and forward rapidities, as well as for the normalized ratio R pA of these results to spectra in proton-proton collisions. In particular, we observe that the R pA data are strongly bounded by our computations of the same for each of the individual NRQCD channels;more » this result provides strong evidence that our description is robust against uncertainties in initial conditions and hadronization mechanisms.« less
Searching for the QCD Axion with Gravitational Microlensing
NASA Astrophysics Data System (ADS)
Fairbairn, Malcolm; Marsh, David J. E.; Quevillon, Jérémie
2017-07-01
The phase transition responsible for axion dark matter (DM) production can create large amplitude isocurvature perturbations, which collapse into dense objects known as axion miniclusters. We use microlensing data from the EROS survey and from recent observations with the Subaru Hyper Suprime Cam to place constraints on the minicluster scenario. We compute the microlensing event rate for miniclusters, treating them as spatially extended objects. Using the published bounds on the number of microlensing events, we bound the fraction of DM collapsed into miniclusters fMC. For an axion with temperature-dependent mass consistent with the QCD axion, we find fMC<0.083 (ma/100 μ eV )0.12 , which represents the first observational constraint on the minicluster fraction. We forecast that a high-efficiency observation of around ten nights with Subaru would be sufficient to constrain fMC≲0.004 over the entire QCD axion mass range. We make various approximations to derive these constraints, and dedicated analyses by the observing teams of EROS and Subaru are necessary to confirm our results. If accurate theoretical predictions for fMC can be made in the future, then microlensing can be used to exclude or discover the QCD axion. Further details of our computations are presented in a companion paper [M. Fairbairn, D. J. E. Marsh, J. Quevillon, and S. Rozier (to be published)].
Determination of electric dipole transitions in heavy quarkonia using potential non-relativistic QCD
NASA Astrophysics Data System (ADS)
Segovia, Jorge; Steinbeißer, Sebastian
2018-05-01
The electric dipole transitions {χ }bJ(1P)\\to γ \\Upsilon (1S) with J = 0, 1, 2 and {h}b(1P)\\to γ {η }b(1S) are computed using the weak-coupling version of a low-energy effective field theory named potential non-relativistic QCD (pNRQCD). In order to improve convergence and thus give firm predictions for the studied reactions, the full static potential is incorporated into the leading order Hamiltonian; moreover, we must handle properly renormalon effects and re-summation of large logarithms. The precision we reach is {k}γ 3/{(mv)}2× O({v}2), where kγ is the photon energy, m is the mass of the heavy quark and v its velocity. Our analysis separates those relativistic contributions that account for the electromagnetic interaction terms in the pNRQCD Lagrangian which are v 2 suppressed and those that account for wave function corrections of relative order v 2. Among the last ones, corrections from 1/m and 1/m2 potentials are computed, but not those coming from higher Fock states since they demand non-perturbative input and are {{{Λ }}}{{QCD}}2/{(mv)}2 or {{{Λ }}}{{QCD}}3/({m}3{v}4) suppressed, at least, in the strict weak coupling regime. These proceedings are based on the forthcoming publication [1].
High-precision QCD at hadron colliders:electroweak gauge boson rapidity distributions at NNLO
DOE Office of Scientific and Technical Information (OSTI.GOV)
Anastasiou, C.
2004-01-05
We compute the rapidity distributions of W and Z bosons produced at the Tevatron and the LHC through next-to-next-to leading order in QCD. Our results demonstrate remarkable stability with respect to variations of the factorization and renormalization scales for all values of rapidity accessible in current and future experiments. These processes are therefore ''gold-plated'': current theoretical knowledge yields QCD predictions accurate to better than one percent. These results strengthen the proposal to use $W$ and $Z$ production to determine parton-parton luminosities and constrain parton distribution functions at the LHC. For example, LHC data should easily be able to distinguish themore » central parton distribution fit obtained by MRST from that obtained by Alekhin.« less
NASA Astrophysics Data System (ADS)
Gehrmann-De Ridder, A.; Gehrmann, T.; Glover, E. W. N.; Huss, A.; Walker, D. M.
2018-03-01
The transverse momentum spectra of weak gauge bosons and their ratios probe the underlying dynamics and are crucial in testing our understanding of the standard model. They are an essential ingredient in precision measurements, such as the W boson mass extraction. To fully exploit the potential of the LHC data, we compute the second-order [next-to-next-to-leading-order (NNLO)] QCD corrections to the inclusive-pTW spectrum as well as to the ratios of spectra for W-/W+ and Z /W . We find that the inclusion of NNLO QCD corrections considerably improves the theoretical description of the experimental CMS data and results in a substantial reduction of the residual scale uncertainties.
Fluctuations in the quark-meson model for QCD with isospin chemical potential
NASA Astrophysics Data System (ADS)
Kamikado, Kazuhiko; Strodthoff, Nils; von Smekal, Lorenz; Wambach, Jochen
2013-01-01
We study the two-flavor quark-meson (QM) model with the functional renormalization group (FRG) to describe the effects of collective mesonic fluctuations on the phase diagram of QCD at finite baryon and isospin chemical potentials, μB and μI. With only isospin chemical potential there is a precise equivalence between the competing dynamics of chiral versus pion condensation and that of collective mesonic and baryonic fluctuations in the quark-meson-diquark model for two-color QCD at finite baryon chemical potential. Here, finite μB = 3 μ introduces an additional dimension to the phase diagram as compared to two-color QCD, however. At zero temperature, the (μI, μ) plane of this phase diagram is strongly constrained by the "Silver Blaze problem." In particular, the onset of pion condensation must occur at μI =mπ / 2, independent of μ as long as μ +μI stays below the constituent quark mass of the QM model or the liquid-gas transition line of nuclear matter in QCD. In order to maintain this relation beyond mean field it is crucial to compute the pion mass from its timelike correlator with the FRG in a consistent way.
Restoring canonical partition functions from imaginary chemical potential
NASA Astrophysics Data System (ADS)
Bornyakov, V. G.; Boyda, D.; Goy, V.; Molochkov, A.; Nakamura, A.; Nikolaev, A.; Zakharov, V. I.
2018-03-01
Using GPGPU techniques and multi-precision calculation we developed the code to study QCD phase transition line in the canonical approach. The canonical approach is a powerful tool to investigate sign problem in Lattice QCD. The central part of the canonical approach is the fugacity expansion of the grand canonical partition functions. Canonical partition functions Zn(T) are coefficients of this expansion. Using various methods we study properties of Zn(T). At the last step we perform cubic spline for temperature dependence of Zn(T) at fixed n and compute baryon number susceptibility χB/T2 as function of temperature. After that we compute numerically ∂χ/∂T and restore crossover line in QCD phase diagram. We use improved Wilson fermions and Iwasaki gauge action on the 163 × 4 lattice with mπ/mρ = 0.8 as a sandbox to check the canonical approach. In this framework we obtain coefficient in parametrization of crossover line Tc(µ2B) = Tc(C-ĸµ2B/T2c) with ĸ = -0.0453 ± 0.0099.
Shear viscosity of the quark-gluon plasma in a weak magnetic field in perturbative QCD: Leading log
NASA Astrophysics Data System (ADS)
Li, Shiyong; Yee, Ho-Ung
2018-03-01
We compute the shear viscosity of two-flavor QCD plasma in an external magnetic field in perturbative QCD at leading log order, assuming that the magnetic field is weak or soft: e B ˜g4log (1 /g )T2. We work in the assumption that the magnetic field is homogeneous and static, and the electrodynamics is nondynamical in a formal limit e →0 while e B is kept fixed. We show that the shear viscosity takes a form η =η ¯(B ¯)T3/(g4log (1 /g )) with a dimensionless function η ¯(B ¯) in terms of a dimensionless variable B ¯=(e B )/(g4log (1 /g )T2). The variable B ¯ corresponds to the relative strength of the effect of cyclotron motions compared to the QCD collisions: B ¯˜lmfp/lcyclo. We provide a full numerical result for the scaled shear viscosity η ¯(B ¯).
Freeze-out conditions in heavy ion collisions from QCD thermodynamics.
Bazavov, A; Ding, H-T; Hegde, P; Kaczmarek, O; Karsch, F; Laermann, E; Mukherjee, Swagato; Petreczky, P; Schmidt, C; Smith, D; Soeldner, W; Wagner, M
2012-11-09
We present a determination of freeze-out conditions in heavy ion collisions based on ratios of cumulants of net electric charge fluctuations. These ratios can reliably be calculated in lattice QCD for a wide range of chemical potential values by using a next-to-leading order Taylor series expansion around the limit of vanishing baryon, electric charge and strangeness chemical potentials. From a computation of up to fourth order cumulants and charge correlations we first determine the strangeness and electric charge chemical potentials that characterize freeze-out conditions in a heavy ion collision and confirm that in the temperature range 150 MeV ≤ T ≤ 170 MeV the hadron resonance gas model provides good approximations for these parameters that agree with QCD calculations on the 5%-15% level. We then show that a comparison of lattice QCD results for ratios of up to third order cumulants of electric charge fluctuations with experimental results allows us to extract the freeze-out baryon chemical potential and the freeze-out temperature.
NASA Astrophysics Data System (ADS)
Bochicchio, Marco
2017-03-01
Yang-Mills (YM) theory and QCD are known to be renormalizable, but not ultraviolet (UV) finite, order by order, in perturbation theory. It is a fundamental question whether YM theory or QCD is UV finite, or only renormalizable, order by order, in the large-N 't Hooft or Veneziano expansions. We demonstrate that the renormalization group (RG) and asymptotic freedom imply that in 't Hooft large-N expansion the S matrix in YM theory is UV finite, while in both 't Hooft and Veneziano large-N expansions, the S matrix in confining massless QCD is renormalizable but not UV finite. By the same argument, the large-N N =1 supersymmetry (SUSY) YM S matrix is UV finite as well. Besides, we demonstrate that, in both 't Hooft and Veneziano large-N expansions, the correlators of local gauge-invariant operators, as opposed to the S matrix, are renormalizable but, in general, not UV finite, either in YM theory and N =1 SUSY YM theory or a fortiori in massless QCD. Moreover, we compute explicitly the counterterms that arise from renormalizing the 't Hooft and Veneziano expansions by deriving in confining massless QCD-like theories a low-energy theorem of the Novikov-Shifman-Vainshtein-Zakharov type that relates the log derivative with respect to the gauge coupling of a k -point correlator, or the log derivative with respect to the RG-invariant scale, to a (k +1 )-point correlator with the insertion of Tr F2 at zero momentum. Finally, we argue that similar results hold in the large-N limit of a vast class of confining massive QCD-like theories, provided a renormalization scheme exists—as, for example, MS ¯ —in which the beta function is not dependent on the masses. Specifically, in both 't Hooft and Veneziano large-N expansions, the S matrix in confining massive QCD and massive N =1 SUSY QCD is renormalizable but not UV finite.
Parallel software for lattice N = 4 supersymmetric Yang-Mills theory
NASA Astrophysics Data System (ADS)
Schaich, David; DeGrand, Thomas
2015-05-01
We present new parallel software, SUSY LATTICE, for lattice studies of four-dimensional N = 4 supersymmetric Yang-Mills theory with gauge group SU(N). The lattice action is constructed to exactly preserve a single supersymmetry charge at non-zero lattice spacing, up to additional potential terms included to stabilize numerical simulations. The software evolved from the MILC code for lattice QCD, and retains a similar large-scale framework despite the different target theory. Many routines are adapted from an existing serial code (Catterall and Joseph, 2012), which SUSY LATTICE supersedes. This paper provides an overview of the new parallel software, summarizing the lattice system, describing the applications that are currently provided and explaining their basic workflow for non-experts in lattice gauge theory. We discuss the parallel performance of the code, and highlight some notable aspects of the documentation for those interested in contributing to its future development.
Baryon magnetic moments: Symmetries and relations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Parreno, Assumpta; Savage, Martin; Tiburzi, Brian
Magnetic moments of the octet baryons are computed using lattice QCD in background magnetic fields, including the first treatment of the magnetically coupled Σ0- Λ system. Although the computations are performed for relatively large values of the up and down quark masses, we gain new insight into the symmetries and relations between magnetic moments by working at a three-flavor mass-symmetric point. While the spinflavor symmetry in the large Nc limit of QCD is shared by the naïve constituent quark model, we find instances where quark model predictions are considerably favored over those emerging in the large Nc limit. We suggestmore » further calculations that would shed light on the curious patterns of baryon magnetic moments.« less
Chimera distribution amplitudes for the pion and the longitudinally polarized ρ-meson
NASA Astrophysics Data System (ADS)
Stefanis, N. G.; Pimikov, A. V.
2016-01-01
Using QCD sum rules with nonlocal condensates, we show that the distribution amplitude of the longitudinally polarized ρ-meson may have a shorttailed platykurtic profile in close analogy to our recently proposed platykurtic distribution amplitude for the pion. Such a chimera distribution de facto amalgamates the broad unimodal profile of the distribution amplitude, obtained with a Dyson-Schwinger equations-based computational scheme, with the suppressed tails characterizing the bimodal distribution amplitudes derived from QCD sum rules with nonlocal condensates. We argue that pattern formation, emerging from the collective synchronization of coupled oscillators, can provide a single theoretical scaffolding to study unimodal and bimodal distribution amplitudes of light mesons without recourse to particular computational schemes and the reasons for them.
Color fields of the static pentaquark system computed in SU(3) lattice QCD
NASA Astrophysics Data System (ADS)
Cardoso, Nuno; Bicudo, Pedro
2013-02-01
We compute the color fields of SU(3) lattice QCD created by static pentaquark systems, in a 243×48 lattice at β=6.2 corresponding to a lattice spacing a=0.07261(85)fm. We find that the pentaquark color fields are well described by a multi-Y-type shaped flux tube. The flux tube junction points are compatible with Fermat-Steiner points minimizing the total flux tube length. We also compare the pentaquark flux tube profile with the diquark-diantiquark central flux tube profile in the tetraquark and the quark-antiquark fundamental flux tube profile in the meson, and they match, thus showing that the pentaquark flux tubes are composed of fundamental flux tubes.
Non-perturbative quark mass renormalisation and running in N_{f}=3 QCD
NASA Astrophysics Data System (ADS)
Campos, I.; Fritzsch, P.; Pena, C.; Preti, D.; Ramos, A.; Vladikas, A.
2018-05-01
We determine from first principles the quark mass anomalous dimension in N_{f}=3 QCD between the electroweak and hadronic scales. This allows for a fully non-perturbative connection of the perturbative and non-perturbative regimes of the Standard Model in the hadronic sector. The computation is carried out to high accuracy, employing massless O (a)-improved Wilson quarks and finite-size scaling techniques. We also provide the matching factors required in the renormalisation of light quark masses from lattice computations with O (a)-improved Wilson fermions and a tree-level Symanzik improved gauge action. The total uncertainty due to renormalisation and running in the determination of light quark masses in the SM is thus reduced to about 1%.
Higgs boson decay into b-quarks at NNLO accuracy
NASA Astrophysics Data System (ADS)
Del Duca, Vittorio; Duhr, Claude; Somogyi, Gábor; Tramontano, Francesco; Trócsányi, Zoltán
2015-04-01
We compute the fully differential decay rate of the Standard Model Higgs boson into b-quarks at next-to-next-to-leading order (NNLO) accuracy in αs. We employ a general subtraction scheme developed for the calculation of higher order perturbative corrections to QCD jet cross sections, which is based on the universal infrared factorization properties of QCD squared matrix elements. We show that the subtractions render the various contributions to the NNLO correction finite. In particular, we demonstrate analytically that the sum of integrated subtraction terms correctly reproduces the infrared poles of the two-loop double virtual contribution to this process. We present illustrative differential distributions obtained by implementing the method in a parton level Monte Carlo program. The basic ingredients of our subtraction scheme, used here for the first time to compute a physical observable, are universal and can be employed for the computation of more involved processes.
A model for pion-pion scattering in large- N QCD
NASA Astrophysics Data System (ADS)
Veneziano, G.; Yankielowicz, S.; Onofri, E.
2017-04-01
Following up on recent work by Caron-Huot et al. we consider a generalization of the old Lovelace-Shapiro model as a toy model for ππ scattering satisfying (most of) the properties expected to hold in ('t Hooft's) large- N limit of massless QCD. In particular, the model has asymptotically linear and parallel Regge trajectories at positive t, a positive leading Regge intercept α 0 < 1, and an effective bending of the trajectories in the negative- t region producing a fixed branch point at J = 0 for t < t 0 < 0. Fixed (physical) angle scattering can be tuned to match the power-like behavior (including logarithmic corrections) predicted by perturbative QCD: A( s, t) ˜ s - β log( s)-γ F ( θ). Tree-level unitarity (i.e. positivity of residues for all values of s and J ) imposes strong constraints on the allowed region in the α0- β-γ parameter space, which nicely includes a physically interesting region around α 0 = 0 .5, β = 2 and γ = 3. The full consistency of the model would require an extension to multi-pion processes, a program we do not undertake in this paper.
Process-independent strong running coupling
Binosi, Daniele; Mezrag, Cedric; Papavassiliou, Joannis; ...
2017-09-25
Here, we unify two widely different approaches to understanding the infrared behavior of quantum chromodynamics (QCD), one essentially phenomenological, based on data, and the other computational, realized via quantum field equations in the continuum theory. Using the latter, we explain and calculate a process-independent running-coupling for QCD, a new type of effective charge that is an analogue of the Gell-Mann–Low effective coupling in quantum electrodynamics. The result is almost identical to the process-dependent effective charge defined via the Bjorken sum rule, which provides one of the most basic constraints on our knowledge of nucleon spin structure. As a result, thismore » reveals the Bjorken sum to be a near direct means by which to gain empirical insight into QCD's Gell-Mann–Low effective charge.« less
NASA Astrophysics Data System (ADS)
Jones, S. P.; Kerner, M.; Luisoni, G.
2018-04-01
We present the next-to-leading-order QCD corrections to the production of a Higgs boson in association with one jet at the LHC including the full top-quark mass dependence. The mass of the bottom quark is neglected. The two-loop integrals appearing in the virtual contribution are calculated numerically using the method of sector decomposition. We study the Higgs boson transverse momentum distribution, focusing on the high pt ,H region, where the top-quark loop is resolved. We find that the next-to-leading-order QCD corrections are large but that the ratio of the next-to-leading-order to leading-order result is similar to that obtained by computing in the limit of large top-quark mass.
Glueball spectrum and hadronic processes in low-energy QCD
NASA Astrophysics Data System (ADS)
Frasca, Marco
2010-10-01
Low-energy limit of quantum chromodynamics (QCD) is obtained using a mapping theorem recently proved. This theorem states that, classically, solutions of a massless quartic scalar field theory are approximate solutions of Yang-Mills equations in the limit of the gauge coupling going to infinity. Low-energy QCD is described by a Yukawa theory further reducible to a Nambu-Jona-Lasinio model. At the leading order one can compute glue-quark interactions and one is able to calculate the properties of the σ and η-η mesons. Finally, it is seen that all the physics of strong interactions, both in the infrared and ultraviolet limit, is described by a single constant Λ arising in the ultraviolet by dimensional transmutation and in the infrared as an integration constant.
Decoupling the NLO-coupled QED⊗QCD, DGLAP evolution equations, using Laplace transform method
NASA Astrophysics Data System (ADS)
Mottaghizadeh, Marzieh; Eslami, Parvin; Taghavi-Shahri, Fatemeh
2017-05-01
We analytically solved the QED⊗QCD-coupled DGLAP evolution equations at leading order (LO) quantum electrodynamics (QED) and next-to-leading order (NLO) quantum chromodynamics (QCD) approximations, using the Laplace transform method and then computed the proton structure function in terms of the unpolarized parton distribution functions. Our analytical solutions for parton densities are in good agreement with those from CT14QED (1.2952 < Q2 < 1010) (Ref. 6) global parametrizations and APFEL (A PDF Evolution Library) (2 < Q2 < 108) (Ref. 4). We also compared the proton structure function, F2p(x,Q2), with the experimental data released by the ZEUS and H1 collaborations at HERA. There is a nice agreement between them in the range of low and high x and Q2.
Jones, S P; Kerner, M; Luisoni, G
2018-04-20
We present the next-to-leading-order QCD corrections to the production of a Higgs boson in association with one jet at the LHC including the full top-quark mass dependence. The mass of the bottom quark is neglected. The two-loop integrals appearing in the virtual contribution are calculated numerically using the method of sector decomposition. We study the Higgs boson transverse momentum distribution, focusing on the high p_{t,H} region, where the top-quark loop is resolved. We find that the next-to-leading-order QCD corrections are large but that the ratio of the next-to-leading-order to leading-order result is similar to that obtained by computing in the limit of large top-quark mass.
Process-independent strong running coupling
DOE Office of Scientific and Technical Information (OSTI.GOV)
Binosi, Daniele; Mezrag, Cedric; Papavassiliou, Joannis
Here, we unify two widely different approaches to understanding the infrared behavior of quantum chromodynamics (QCD), one essentially phenomenological, based on data, and the other computational, realized via quantum field equations in the continuum theory. Using the latter, we explain and calculate a process-independent running-coupling for QCD, a new type of effective charge that is an analogue of the Gell-Mann–Low effective coupling in quantum electrodynamics. The result is almost identical to the process-dependent effective charge defined via the Bjorken sum rule, which provides one of the most basic constraints on our knowledge of nucleon spin structure. As a result, thismore » reveals the Bjorken sum to be a near direct means by which to gain empirical insight into QCD's Gell-Mann–Low effective charge.« less
Exploring Flavor Physics with Lattice QCD
NASA Astrophysics Data System (ADS)
Du, Daping; Fermilab/MILC Collaborations Collaboration
2016-03-01
The Standard Model has been a very good description of the subatomic particle physics. In the search for physics beyond the Standard Model in the context of flavor physics, it is important to sharpen our probes using some gold-plated processes (such as B rare decays), which requires the knowledge of the input parameters, such as the Cabibbo-Kobayashi-Maskawa (CKM) matrix elements and other nonperturbative quantities, with sufficient precision. Lattice QCD is so far the only first-principle method which could compute these quantities with competitive and systematically improvable precision using the state of the art simulation techniques. I will discuss the recent progress of lattice QCD calculations on some of these nonpurturbative quantities and their applications in flavor physics. I will also discuss the implications and future perspectives of these calculations in flavor physics.
Lattice QCD at finite temperature and density from Taylor expansion
NASA Astrophysics Data System (ADS)
Steinbrecher, Patrick
2017-01-01
In the first part, I present an overview of recent Lattice QCD simulations at finite temperature and density. In particular, we discuss fluctuations of conserved charges: baryon number, electric charge and strangeness. These can be obtained from Taylor expanding the QCD pressure as a function of corresponding chemical potentials. Our simulations were performed using quark masses corresponding to physical pion mass of about 140 MeV and allow a direct comparison to experimental data from ultra-relativistic heavy ion beams at hadron colliders such as the Relativistic Heavy Ion Collider at Brookhaven National Laboratory and the Large Hadron Collider at CERN. In the second part, we discuss computational challenges for current and future exascale Lattice simulations with a focus on new silicon developments from Intel and NVIDIA.
NASA Astrophysics Data System (ADS)
Iwasaki, Y.
1997-02-01
The CP-PACS computer with a peak speed of 300 Gflops was completed in March 1996 and has started to operate. We describe the final specification and the hardware implementation of the CP-PACS computer, and its performance for QCD codes. A plan of the grade-up of the computer scheduled for fall of 1996 is also given.
Franzosi, Diogo Buarque; Vryonidou, Eleni; Zhang, Cen
2017-10-13
Scalar and pseudo-scalar resonances decaying to top quarks are common predictions in several scenarios beyond the standard model (SM) and are extensively searched for by LHC experiments. Challenges on the experimental side require optimising the strategy based on accurate predictions. Firstly, QCD corrections are known to be large both for the SM QCD background and for the pure signal scalar production. Secondly, leading order and approximate next-to-leading order (NLO) calculations indicate that the interference between signal and background is large and drastically changes the lineshape of the signal, from a simple peak to a peak-dip structure. Therefore, a robust predictionmore » of this interference at NLO accuracy in QCD is necessary to ensure that higher-order corrections do not alter the lineshapes. We compute the exact NLO corrections, assuming a point-like coupling between the scalar and the gluons and consistently embedding the calculation in an effective field theory within an automated framework, and present results for a representative set of beyond the SM benchmarks. The results can be further matched to parton shower simulation, providing more realistic predictions. We find that NLO corrections are important and lead to a significant reduction of the uncertainties. We also discuss how our computation can be used to improve the predictions for physics scenarios where the gluon-scalar loop is resolved and the effective approach is less applicable.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Franzosi, Diogo Buarque; Vryonidou, Eleni; Zhang, Cen
Scalar and pseudo-scalar resonances decaying to top quarks are common predictions in several scenarios beyond the standard model (SM) and are extensively searched for by LHC experiments. Challenges on the experimental side require optimising the strategy based on accurate predictions. Firstly, QCD corrections are known to be large both for the SM QCD background and for the pure signal scalar production. Secondly, leading order and approximate next-to-leading order (NLO) calculations indicate that the interference between signal and background is large and drastically changes the lineshape of the signal, from a simple peak to a peak-dip structure. Therefore, a robust predictionmore » of this interference at NLO accuracy in QCD is necessary to ensure that higher-order corrections do not alter the lineshapes. We compute the exact NLO corrections, assuming a point-like coupling between the scalar and the gluons and consistently embedding the calculation in an effective field theory within an automated framework, and present results for a representative set of beyond the SM benchmarks. The results can be further matched to parton shower simulation, providing more realistic predictions. We find that NLO corrections are important and lead to a significant reduction of the uncertainties. We also discuss how our computation can be used to improve the predictions for physics scenarios where the gluon-scalar loop is resolved and the effective approach is less applicable.« less
NASA Astrophysics Data System (ADS)
Deur, Alexandre; Brodsky, Stanley J.; de Téramond, Guy F.
2016-09-01
We review the present theoretical and empirical knowledge for αs, the fundamental coupling underlying the interactions of quarks and gluons in Quantum Chromodynamics (QCD). The dependence of αs(Q2) on momentum transfer Q encodes the underlying dynamics of hadron physics-from color confinement in the infrared domain to asymptotic freedom at short distances. We review constraints on αs(Q2) at high Q2, as predicted by perturbative QCD, and its analytic behavior at small Q2, based on models of nonperturbative dynamics. In the introductory part of this review, we explain the phenomenological meaning of the coupling, the reason for its running, and the challenges facing a complete understanding of its analytic behavior in the infrared domain. In the second, more technical, part of the review, we discuss the behavior of αs(Q2) in the high momentum transfer domain of QCD. We review how αs is defined, including its renormalization scheme dependence, the definition of its renormalization scale, the utility of effective charges, as well as "Commensurate Scale Relations" which connect the various definitions of the QCD coupling without renormalization-scale ambiguity. We also report recent significant measurements and advanced theoretical analyses which have led to precise QCD predictions at high energy. As an example of an important optimization procedure, we discuss the "Principle of Maximum Conformality", which enhances QCD's predictive power by removing the dependence of the predictions for physical observables on the choice of theoretical conventions such as the renormalization scheme. In the last part of the review, we discuss the challenge of understanding the analytic behavior αs(Q2) in the low momentum transfer domain. We survey various theoretical models for the nonperturbative strongly coupled regime, such as the light-front holographic approach to QCD. This new framework predicts the form of the quark-confinement potential underlying hadron spectroscopy and dynamics, and it gives a remarkable connection between the perturbative QCD scale Λ and hadron masses. One can also identify a specific scale Q0 which demarcates the division between perturbative and nonperturbative QCD. We also review other important methods for computing the QCD coupling, including lattice QCD, the Schwinger-Dyson equations and the Gribov-Zwanziger analysis. After describing these approaches and enumerating their conflicting predictions, we discuss the origin of these discrepancies and how to remedy them. Our aim is not only to review the advances in this difficult area, but also to suggest what could be an optimal definition of αs(Q2) in order to bring better unity to the subject.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, Stanley J.; de Teramond, Guy F.; /SLAC /Southern Denmark U., CP3-Origins /Costa Rica U.
2011-01-10
AdS/QCD, the correspondence between theories in a dilaton-modified five-dimensional anti-de Sitter space and confining field theories in physical space-time, provides a remarkable semiclassical model for hadron physics. Light-front holography allows hadronic amplitudes in the AdS fifth dimension to be mapped to frame-independent light-front wavefunctions of hadrons in physical space-time. The result is a single-variable light-front Schroedinger equation which determines the eigenspectrum and the light-front wavefunctions of hadrons for general spin and orbital angular momentum. The coordinate z in AdS space is uniquely identified with a Lorentz-invariant coordinate {zeta} which measures the separation of the constituents within a hadron at equalmore » light-front time and determines the off-shell dynamics of the bound state wavefunctions as a function of the invariant mass of the constituents. The hadron eigenstates generally have components with different orbital angular momentum; e.g., the proton eigenstate in AdS/QCD with massless quarks has L = 0 and L = 1 light-front Fock components with equal probability. Higher Fock states with extra quark-anti quark pairs also arise. The soft-wall model also predicts the form of the nonperturbative effective coupling and its {beta}-function. The AdS/QCD model can be systematically improved by using its complete orthonormal solutions to diagonalize the full QCD light-front Hamiltonian or by applying the Lippmann-Schwinger method to systematically include QCD interaction terms. Some novel features of QCD are discussed, including the consequences of confinement for quark and gluon condensates. A method for computing the hadronization of quark and gluon jets at the amplitude level is outlined.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Deur, Alexandre; Brodsky, Stanley J.; de Téramond, Guy F.
Here, we review present knowledge onmore » $$\\alpha_{s}$$, the Quantum Chromodynamics (QCD) running coupling. The dependence of $$\\alpha_s(Q^2)$$ on momentum transfer $Q$ encodes the underlying dynamics of hadron physics --from color confinement in the infrared domain to asymptotic freedom at short distances. We will survey our present theoretical and empirical knowledge of $$\\alpha_s(Q^2)$$, including constraints at high $Q^2$ predicted by perturbative QCD, and constraints at small $Q^2$ based on models of nonperturbative dynamics. In the first, introductory, part of this review, we explain the phenomenological meaning of the coupling, the reason for its running, and the challenges facing a complete understanding of its analytic behavior in the infrared domain. In the second, more technical, part of the review, we discuss $$\\alpha_s(Q^2)$$ in the high momentum transfer domain of QCD. We review how $$\\alpha_s$$ is defined, including its renormalization scheme dependence, the definition of its renormalization scale, the utility of effective charges, as well as `` Commensurate Scale Relations" which connect the various definitions of the QCD coupling without renormalization scale ambiguity. We also report recent important experimental measurements and advanced theoretical analyses which have led to precise QCD predictions at high energy. As an example of an important optimization procedure, we discuss the ``Principle of Maximum Conformality" which enhances QCD's predictive power by removing the dependence of the predictions for physical observables on the choice of the gauge and renormalization scheme. In last part of the review, we discuss $$\\alpha_s(Q^2)$$ in the low momentum transfer domain, where there has been no consensus on how to define $$\\alpha_s(Q^2)$$ or its analytic behavior. We will discuss the various approaches used for low energy calculations. Among them, we will discuss the light-front holographic approach to QCD in the strongly coupled regime and its prediction for the analytic form of $$\\alpha_s(Q^2)$$. The AdS/QCD light-front holographic analysis predicts the color confinement potential underlying hadron spectroscopy and dynamics, and it gives a remarkable connection between the perturbative QCD scale $$\\Lambda$$ and hadron masses. One can also identify a specific scale $$Q_0$$ which demarcates the division between perturbative and nonperturbative QCD. We also review other important methods for computing the QCD coupling, including Lattice QCD, Schwinger-Dyson equations and the Gribov-Zwanziger analysis. After describing these approaches and enumerating conflicting results, we provide a partial discussion on the origin of these discrepancies and how to remedy them. Our aim is not only to review the advances on this difficult subject, but also to suggest what could be the best definition of $$\\alpha_s(Q^2)$$ in order to bring better unity to the subject.« less
Deur, Alexandre; Brodsky, Stanley J.; de Téramond, Guy F.
2016-05-09
Here, we review present knowledge onmore » $$\\alpha_{s}$$, the Quantum Chromodynamics (QCD) running coupling. The dependence of $$\\alpha_s(Q^2)$$ on momentum transfer $Q$ encodes the underlying dynamics of hadron physics --from color confinement in the infrared domain to asymptotic freedom at short distances. We will survey our present theoretical and empirical knowledge of $$\\alpha_s(Q^2)$$, including constraints at high $Q^2$ predicted by perturbative QCD, and constraints at small $Q^2$ based on models of nonperturbative dynamics. In the first, introductory, part of this review, we explain the phenomenological meaning of the coupling, the reason for its running, and the challenges facing a complete understanding of its analytic behavior in the infrared domain. In the second, more technical, part of the review, we discuss $$\\alpha_s(Q^2)$$ in the high momentum transfer domain of QCD. We review how $$\\alpha_s$$ is defined, including its renormalization scheme dependence, the definition of its renormalization scale, the utility of effective charges, as well as `` Commensurate Scale Relations" which connect the various definitions of the QCD coupling without renormalization scale ambiguity. We also report recent important experimental measurements and advanced theoretical analyses which have led to precise QCD predictions at high energy. As an example of an important optimization procedure, we discuss the ``Principle of Maximum Conformality" which enhances QCD's predictive power by removing the dependence of the predictions for physical observables on the choice of the gauge and renormalization scheme. In last part of the review, we discuss $$\\alpha_s(Q^2)$$ in the low momentum transfer domain, where there has been no consensus on how to define $$\\alpha_s(Q^2)$$ or its analytic behavior. We will discuss the various approaches used for low energy calculations. Among them, we will discuss the light-front holographic approach to QCD in the strongly coupled regime and its prediction for the analytic form of $$\\alpha_s(Q^2)$$. The AdS/QCD light-front holographic analysis predicts the color confinement potential underlying hadron spectroscopy and dynamics, and it gives a remarkable connection between the perturbative QCD scale $$\\Lambda$$ and hadron masses. One can also identify a specific scale $$Q_0$$ which demarcates the division between perturbative and nonperturbative QCD. We also review other important methods for computing the QCD coupling, including Lattice QCD, Schwinger-Dyson equations and the Gribov-Zwanziger analysis. After describing these approaches and enumerating conflicting results, we provide a partial discussion on the origin of these discrepancies and how to remedy them. Our aim is not only to review the advances on this difficult subject, but also to suggest what could be the best definition of $$\\alpha_s(Q^2)$$ in order to bring better unity to the subject.« less
QCD sum-rules analysis of vector (1-) heavy quarkonium meson-hybrid mixing
NASA Astrophysics Data System (ADS)
Palameta, A.; Ho, J.; Harnett, D.; Steele, T. G.
2018-02-01
We use QCD Laplace sum rules to study meson-hybrid mixing in vector (1-) heavy quarkonium. We compute the QCD cross-correlator between a heavy meson current and a heavy hybrid current within the operator product expansion. In addition to leading-order perturbation theory, we include four- and six-dimensional gluon condensate contributions as well as a six-dimensional quark condensate contribution. We construct several single and multiresonance models that take known hadron masses as inputs. We investigate which resonances couple to both currents and so exhibit meson-hybrid mixing. Compared to single resonance models that include only the ground state, we find that models that also include excited states lead to significantly improved agreement between QCD and experiment. In the charmonium sector, we find that meson-hybrid mixing is consistent with a two-resonance model consisting of the J /ψ and a 4.3 GeV resonance. In the bottomonium sector, we find evidence for meson-hybrid mixing in the ϒ (1 S ) , ϒ (2 S ), ϒ (3 S ), and ϒ (4 S ).
Blum, Thomas; Chowdhury, Saumitra; Hayakawa, Masashi; Izubuchi, Taku
2015-01-09
The most compelling possibility for a new law of nature beyond the four fundamental forces comprising the standard model of high-energy physics is the discrepancy between measurements and calculations of the muon anomalous magnetic moment. Until now a key part of the calculation, the hadronic light-by-light contribution, has only been accessible from models of QCD, the quantum description of the strong force, whose accuracy at the required level may be questioned. A first principles calculation with systematically improvable errors is needed, along with the upcoming experiments, to decisively settle the matter. For the first time, the form factor that yields the light-by-light scattering contribution to the muon anomalous magnetic moment is computed in such a framework, lattice QCD+QED and QED. A nonperturbative treatment of QED is used and checked against perturbation theory. The hadronic contribution is calculated for unphysical quark and muon masses, and only the diagram with a single quark loop is computed for which statistically significant signals are obtained. Initial results are promising, and the prospect for a complete calculation with physical masses and controlled errors is discussed.
D-Meson Mixing in 2+1-Flavor Lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chang, Chia Cheng; Bouchard, C. M.; El-Khadra, A. X.
We present results for neutral D-meson mixing in 2+1-flavor lattice QCD. We compute the matrix elements for all five operators that contribute to D mixing at short distances, including those that only arise beyond the Standard Model. Our results have an uncertainty similar to those of the ETM collaboration (with 2 and with 2+1+1 flavors). This work shares many features with a recent publication on B mixing and with ongoing work on heavy-light decay constants from the Fermilab Lattice and MILC Collaborations.
Precise QCD Predictions for the Production of a Z Boson in Association with a Hadronic Jet.
Gehrmann-De Ridder, A; Gehrmann, T; Glover, E W N; Huss, A; Morgan, T A
2016-07-08
We compute the cross section and differential distributions for the production of a Z boson in association with a hadronic jet to next-to-next-to-leading order (NNLO) in perturbative QCD, including the leptonic decay of the Z boson. We present numerical results for the transverse momentum and rapidity distributions of both the Z boson and the associated jet at the LHC. We find that the NNLO corrections increase the NLO predictions by approximately 1% and significantly reduce the scale variation uncertainty.
NNLO QCD corrections to the polarized top quark decay t (↑)→Xb+W+
NASA Astrophysics Data System (ADS)
Czarnecki, A.; Groote, S.; Körner, J. G.; Piclum, J. H.
2018-05-01
We compute the next-to-next-to-leading order (NNLO) QCD corrections to the decay t (↑)→Xb+W+ of a polarized top quark. The spin-momentum correlation in this quasi two-body decay is described by the polar angle distribution d Γ /d cos θP=Γ/2 (1 +PtαPcos θP) , where Pt is the polarization of the top quark and αP denotes the asymmetry parameter of the decay. For the latter we find αPNNLO=0.3792 ±0.0037 .
Strong running coupling at τ and Z(0) mass scales from lattice QCD.
Blossier, B; Boucaud, Ph; Brinet, M; De Soto, F; Du, X; Morenas, V; Pène, O; Petrov, K; Rodríguez-Quintero, J
2012-06-29
This Letter reports on the first computation, from data obtained in lattice QCD with u, d, s, and c quarks in the sea, of the running strong coupling via the ghost-gluon coupling renormalized in the momentum-subtraction Taylor scheme. We provide readers with estimates of α(MS[over ¯])(m(τ)(2)) and α(MS[over ¯])(m(Z)(2)) in very good agreement with experimental results. Including a dynamical c quark makes the needed running of α(MS[over ¯]) safer.
η and η' mixing from lattice QCD.
Michael, C; Ottnad, K; Urbach, C
2013-11-01
We present a lattice QCD computation of η and η' masses and mixing angles, for the first time controlling continuum and quark mass extrapolations. The results for M(η) = 551(8)(stat) (6)(yst) MeV and M(η') = 1006(54)(stat)(38)(syst)(+61)(ex) MeV are in excellent agreement with experiment. Our data show that the mixing in the quark flavor basis can be described by a single mixing angle of Ø = 46(1)(stat)(3)(syst)° indicating that the η' is mainly a flavor singlet state.
Lattice field theory applications in high energy physics
NASA Astrophysics Data System (ADS)
Gottlieb, Steven
2016-10-01
Lattice gauge theory was formulated by Kenneth Wilson in 1974. In the ensuing decades, improvements in actions, algorithms, and computers have enabled tremendous progress in QCD, to the point where lattice calculations can yield sub-percent level precision for some quantities. Beyond QCD, lattice methods are being used to explore possible beyond the standard model (BSM) theories of dynamical symmetry breaking and supersymmetry. We survey progress in extracting information about the parameters of the standard model by confronting lattice calculations with experimental results and searching for evidence of BSM effects.
Charmonium-nucleon interactions from the time-dependent HAL QCD method
NASA Astrophysics Data System (ADS)
Sugiura, Takuya; Ikeda, Yoichi; Ishii, Noriyoshi
2018-03-01
The charmonium-nucleon effective central interactions have been computed by the time-dependent HAL QCD method. This gives an updated result of a previous study based on the time-independent method, which is now known to be problematic because of the difficulty in achieving the ground-state saturation. We discuss that the result is consistent with the heavy quark symmetry. No bound state is observed from the analysis of the scattering phase shift; however, this shall lead to a future search of the hidden-charm pentaquarks by considering channel-coupling effects.
Isotensor Axial Polarizability and Lattice QCD Input for Nuclear Double-β Decay Phenomenology
NASA Astrophysics Data System (ADS)
Shanahan, Phiala E.; Tiburzi, Brian C.; Wagman, Michael L.; Winter, Frank; Chang, Emmanuel; Davoudi, Zohreh; Detmold, William; Orginos, Kostas; Savage, Martin J.; Nplqcd Collaboration
2017-08-01
The potential importance of short-distance nuclear effects in double-β decay is assessed using a lattice QCD calculation of the n n →p p transition and effective field theory methods. At the unphysical quark masses used in the numerical computation, these effects, encoded in the isotensor axial polarizability, are found to be of similar magnitude to the nuclear modification of the single axial current, which phenomenologically is the quenching of the axial charge used in nuclear many-body calculations. This finding suggests that nuclear models for neutrinoful and neutrinoless double-β decays should incorporate this previously neglected contribution if they are to provide reliable guidance for next-generation neutrinoless double-β decay searches. The prospects of constraining the isotensor axial polarizabilities of nuclei using lattice QCD input into nuclear many-body calculations are discussed.
The singular behavior of massive QCD amplitudes
NASA Astrophysics Data System (ADS)
Mitov, Alexander; Moch, Sven-Olaf
2007-05-01
We discuss the structure of infrared singularities in on-shell QCD amplitudes with massive partons and present a general factorization formula in the limit of small parton masses. The factorization formula gives rise to an all-order exponentiation of both, the soft poles in dimensional regularization and the large collinear logarithms of the parton masses. Moreover, it provides a universal relation between any on-shell amplitude with massive external partons and its corresponding massless amplitude. For the form factor of a heavy quark we present explicit results including the fixed-order expansion up to three loops in the small mass limit. For general scattering processes we show how our constructive method applies to the computation of all singularities as well as the constant (mass-independent) terms of a generic massive n-parton QCD amplitude up to the next-to-next-to-leading order corrections.
Isotensor Axial Polarizability and Lattice QCD Input for Nuclear Double-β Decay Phenomenology.
Shanahan, Phiala E; Tiburzi, Brian C; Wagman, Michael L; Winter, Frank; Chang, Emmanuel; Davoudi, Zohreh; Detmold, William; Orginos, Kostas; Savage, Martin J
2017-08-11
The potential importance of short-distance nuclear effects in double-β decay is assessed using a lattice QCD calculation of the nn→pp transition and effective field theory methods. At the unphysical quark masses used in the numerical computation, these effects, encoded in the isotensor axial polarizability, are found to be of similar magnitude to the nuclear modification of the single axial current, which phenomenologically is the quenching of the axial charge used in nuclear many-body calculations. This finding suggests that nuclear models for neutrinoful and neutrinoless double-β decays should incorporate this previously neglected contribution if they are to provide reliable guidance for next-generation neutrinoless double-β decay searches. The prospects of constraining the isotensor axial polarizabilities of nuclei using lattice QCD input into nuclear many-body calculations are discussed.
Nonperturbative quark, gluon, and meson correlators of unquenched QCD
NASA Astrophysics Data System (ADS)
Cyrol, Anton K.; Mitter, Mario; Pawlowski, Jan M.; Strodthoff, Nils
2018-03-01
We present nonperturbative first-principle results for quark, gluon, and meson 1PI correlation functions of two-flavor Landau-gauge QCD in the vacuum. These correlation functions carry the full information about the theory. They are obtained by solving their functional renormalization group equations in a systematic vertex expansion, aiming at apparent convergence. This work represents a crucial prerequisite for quantitative first-principle studies of the QCD phase diagram and the hadron spectrum within this framework. In particular, we have computed the gluon, ghost, quark, and scalar-pseudoscalar meson propagators, as well as gluon, ghost-gluon, quark-gluon, quark, quark-meson, and meson interactions. Our results stress the crucial importance of the quantitatively correct running of different vertices in the semiperturbative regime for describing the phenomena and scales of confinement and spontaneous chiral symmetry breaking without phenomenological input.
Off-shell production of top-antitop pairs in the lepton+jets channel at NLO QCD
NASA Astrophysics Data System (ADS)
Denner, Ansgar; Pellen, Mathieu
2018-02-01
The production of top-quark pairs that subsequently decay hadronically and leptonically (lepton+jets channel) is one of the key processes for the study of top-quark properties at the LHC. In this article, NLO QCD corrections of order O({α}s^3{α}^4) to the hadronic process pp\\to {μ}-{\\overline{ν}}_{μ}b\\overline{b}jj are presented. The computation includes off-shell as well as non-resonant contributions, and experimental event selections are used in order to provide realistic predictions. The results are provided in the form of cross sections and differential distributions. The QCD corrections are sizeable and different from the ones of the fully leptonic channel. This is due to the different final state where here four jets are present at leading order.
Collectively loading an application in a parallel computer
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aho, Michael E.; Attinella, John E.; Gooding, Thomas M.
Collectively loading an application in a parallel computer, the parallel computer comprising a plurality of compute nodes, including: identifying, by a parallel computer control system, a subset of compute nodes in the parallel computer to execute a job; selecting, by the parallel computer control system, one of the subset of compute nodes in the parallel computer as a job leader compute node; retrieving, by the job leader compute node from computer memory, an application for executing the job; and broadcasting, by the job leader to the subset of compute nodes in the parallel computer, the application for executing the job.
NASA Astrophysics Data System (ADS)
Zha, Wangmei
2018-02-01
The Solenoidal Tracker at RHIC (STAR) experiment takes advantage of its excellent tracking and particle identification capabilities at mid-rapidity to explore the properties of strongly interacting QCD matter created in heavy-ion collisions at RHIC. The STAR collaboration presented 7 parallel and 2 plenary talks at Strangeness in Quark Matter 2017 and covered various topics including heavy flavor measurements, bulk observables, electro-magnetic probes and the upgrade program. This paper highlights some of the selected results.
Invited Parallel Talk: Lattice results on nucleon/roper properties
NASA Astrophysics Data System (ADS)
Lin, Huey-Wen
2009-12-01
In this proceeding, I review the attempts to calculate the Nucleon resonance (including Roper as first radially excited state of nucleon and other excited states) using lattice quantum chromodynamics (QCD). The latest preliminary results from Hadron Spectrum Collaboration (HSC) with mπ thickapprox 380 MeV are reported. The Sachs electric form factor of the proton and neutron and their transition with the Roper at large Q2 are also updated in this work.
Visualization Tools for Lattice QCD - Final Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Massimo Di Pierro
2012-03-15
Our research project is about the development of visualization tools for Lattice QCD. We developed various tools by extending existing libraries, adding new algorithms, exposing new APIs, and creating web interfaces (including the new NERSC gauge connection web site). Our tools cover the full stack of operations from automating download of data, to generating VTK files (topological charge, plaquette, Polyakov lines, quark and meson propagators, currents), to turning the VTK files into images, movies, and web pages. Some of the tools have their own web interfaces. Some Lattice QCD visualization have been created in the past but, to our knowledge,more » our tools are the only ones of their kind since they are general purpose, customizable, and relatively easy to use. We believe they will be valuable to physicists working in the field. They can be used to better teach Lattice QCD concepts to new graduate students; they can be used to observe the changes in topological charge density and detect possible sources of bias in computations; they can be used to observe the convergence of the algorithms at a local level and determine possible problems; they can be used to probe heavy-light mesons with currents and determine their spatial distribution; they can be used to detect corrupted gauge configurations. There are some indirect results of this grant that will benefit a broader audience than Lattice QCD physicists.« less
Higgs boson production at hadron colliders at N3LO in QCD
NASA Astrophysics Data System (ADS)
Mistlberger, Bernhard
2018-05-01
We present the Higgs boson production cross section at Hadron colliders in the gluon fusion production mode through N3LO in perturbative QCD. Specifically, we work in an effective theory where the top quark is assumed to be infinitely heavy and all other quarks are considered to be massless. Our result is the first exact formula for a partonic hadron collider cross section at N3LO in perturbative QCD. Furthermore, our result is an analytic computation of a hadron collider cross section involving elliptic integrals. We derive numerical predictions for the Higgs boson cross section at the LHC. Previously this result was approximated by an expansion of the cross section around the production threshold of the Higgs boson and we compare our findings. Finally, we study the impact of our new result on the state of the art prediction for the Higgs boson cross section at the LHC.
Isotensor Axial Polarizability and Lattice QCD Input for Nuclear Double- β Decay Phenomenology
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shanahan, Phiala E.; Tiburzi, Brian C.; Wagman, Michael L.
The potential importance of short-distance nuclear effects in double-more » $$\\beta$$ decay is assessed using a lattice QCD calculation of the $$nn\\rightarrow pp$$ transition and effective field theory methods. At the unphysical quark masses used in the numerical computation, these effects, encoded in the isotensor axial polarisability, are found to be of similar magnitude to the nuclear modification of the single axial current, which phenomenologically is the quenching of the axial charge used in nuclear many-body calculations. This finding suggests that nuclear models for neutrinoful and neutrinoless double-$$\\beta$$ decays should incorporate this previously neglected contribution if they are to provide reliable guidance for next-generation neutrinoless double-$$\\beta$$ decay searches. The prospects of constraining the isotensor axial polarisabilities of nuclei using lattice QCD input into nuclear many-body calculations are discussed.« less
Productions of η, ρ0 and ϕ at large transverse momentum in Heavy ion Collisions
NASA Astrophysics Data System (ADS)
Dai, Wei; Zhang, Ben-Wei
2017-08-01
The suppression of the productions of the η meson in relativistic heavy-ion collisions and its ratio of η /π0 are computed theoretically in the framework of the perturbative QCD(pQCD) to confront the experimental data which matches well. We explore how the hadron production ratios as η /π0 would further disclose the informations of the production suppressions due to the energy loss of the energetic jet that propagating though the QGP medium. Also, we present our further studies on vector mesons such as ρ0 and ϕ within the same framework. The theoretical predictions based on pQCD are thus firstly given which give a decent description on the experimental measurements. It paved the way to the uniformly understanding of the strong suppression of single hadron productions at large transverse momentum which is a convincing evidence of the jet quenching effect.
Radiative Transitions in Charmonium from Lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jozef Dudek; Robert Edwards; David Richards
2006-01-17
Radiative transitions between charmonium states offer an insight into the internal structure of heavy-quark bound states within QCD. We compute, for the first time within lattice QCD, the transition form-factors of various multipolarities between the lightest few charmonium states. In addition, we compute the experimentally unobservable, but physically interesting vector form-factors of the {eta}{sub c}, J/{psi} and {chi}{sub c0}. To this end we apply an ambitious combination of lattice techniques, computing three-point functions with heavy domain wall fermions on an anisotropic lattice within the quenched approximation. With an anisotropy {xi} = 3 at a{sub s} {approx} 0.1 fm we findmore » a reasonable gross spectrum and a hyperfine splitting {approx}90 MeV, which compares favorably with other improved actions. In general, after extrapolation of lattice data at non-zero Q{sup 2} to the photopoint, our results agree within errors with all well measured experimental values. Furthermore, results are compared with the expectations of simple quark models where we find that many features are in agreement; beyond this we propose the possibility of constraining such models using our extracted values of physically unobservable quantities such as the J/{psi} quadrupole moment. We conclude that our methods are successful and propose to apply them to the problem of radiative transitions involving hybrid mesons, with the eventual goal of predicting hybrid meson photoproduction rates at the GlueX experiment.« less
Refining new-physics searches in B→Dτν with lattice QCD.
Bailey, Jon A; Bazavov, A; Bernard, C; Bouchard, C M; Detar, C; Du, Daping; El-Khadra, A X; Foley, J; Freeland, E D; Gámiz, E; Gottlieb, Steven; Heller, U M; Kim, Jongjeong; Kronfeld, A S; Laiho, J; Levkova, L; Mackenzie, P B; Meurice, Y; Neil, E T; Oktay, M B; Qiu, Si-Wei; Simone, J N; Sugar, R; Toussaint, D; Van de Water, R S; Zhou, Ran
2012-08-17
The semileptonic decay channel B→Dτν is sensitive to the presence of a scalar current, such as that mediated by a charged-Higgs boson. Recently, the BABAR experiment reported the first observation of the exclusive semileptonic decay B→Dτ(-)ν, finding an approximately 2σ disagreement with the standard-model prediction for the ratio R(D)=BR(B→Dτν)/BR(B→Dℓν), where ℓ = e,μ. We compute this ratio of branching fractions using hadronic form factors computed in unquenched lattice QCD and obtain R(D)=0.316(12)(7), where the errors are statistical and total systematic, respectively. This result is the first standard-model calculation of R(D) from ab initio full QCD. Its error is smaller than that of previous estimates, primarily due to the reduced uncertainty in the scalar form factor f(0)(q(2)). Our determination of R(D) is approximately 1σ higher than previous estimates and, thus, reduces the tension with experiment. We also compute R(D) in models with electrically charged scalar exchange, such as the type-II two-Higgs-doublet model. Once again, our result is consistent with, but approximately 1σ higher than, previous estimates for phenomenologically relevant values of the scalar coupling in the type-II model. As a by-product of our calculation, we also present the standard-model prediction for the longitudinal-polarization ratio P(L)(D)=0.325(4)(3).
Smeared quasidistributions in perturbation theory
NASA Astrophysics Data System (ADS)
Monahan, Christopher
2018-03-01
Quasi- and pseudodistributions provide a new approach to determining parton distribution functions from first principles' calculations of QCD. Here, I calculate the flavor nonsinglet unpolarized quasidistribution at one loop in perturbation theory, using the gradient flow to remove ultraviolet divergences. I demonstrate that, as expected, the gradient flow does not change the infrared structure of the quasidistribution at one loop and use the results to match the smeared matrix elements to those in the MS ¯ scheme. This matching calculation is required to relate numerical results obtained from nonperturbative lattice QCD computations to light-front parton distribution functions extracted from global analyses of experimental data.
Hammant, T C; Hart, A G; von Hippel, G M; Horgan, R R; Monahan, C J
2011-09-09
We present the first application of the background field method to nonrelativistic QCD (NRQCD) on the lattice in order to determine the one-loop radiative corrections to the coefficients of the NRQCD action in a manifestly gauge-covariant manner. The coefficients of the σ·B term in the NRQCD action and the four-fermion spin-spin interaction are computed at the one-loop level; the resulting shift of the hyperfine splitting of bottomonium is found to bring the lattice predictions in line with experiment.
Higgs boson gluon-fusion production beyond threshold in N 3LO QCD
Anastasiou, Charalampos; Duhr, Claude; Dulat, Falko; ...
2015-03-18
In this study, we compute the gluon fusion Higgs boson cross-section at N 3LO through the second term in the threshold expansion. This calculation constitutes a major milestone towards the full N 3LO cross section. Our result has the best formal accuracy in the threshold expansion currently available, and includes contributions from collinear regions besides subleading corrections from soft and hard regions, as well as certain logarithmically enhanced contributions for general kinematics. We use our results to perform a critical appraisal of the validity of the threshold approximation at N 3LO in perturbative QCD.
Aggregating job exit statuses of a plurality of compute nodes executing a parallel application
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aho, Michael E.; Attinella, John E.; Gooding, Thomas M.
Aggregating job exit statuses of a plurality of compute nodes executing a parallel application, including: identifying a subset of compute nodes in the parallel computer to execute the parallel application; selecting one compute node in the subset of compute nodes in the parallel computer as a job leader compute node; initiating execution of the parallel application on the subset of compute nodes; receiving an exit status from each compute node in the subset of compute nodes, where the exit status for each compute node includes information describing execution of some portion of the parallel application by the compute node; aggregatingmore » each exit status from each compute node in the subset of compute nodes; and sending an aggregated exit status for the subset of compute nodes in the parallel computer.« less
QCD thermodynamics with two flavors of quarks[1
NASA Astrophysics Data System (ADS)
MIMD lattice Computations (MILC) Collaboration
We present results of numerical simulations of quantum chromodynamics at finite temperature on the Intel iPSC/860 parallel processor. We performed calculations with two flavors of Kogut-Susskind quarks and of Wilson quarks on 6 × 12 3 lattices in order to study the crossover from the low temperature hadronic regime to the high temperature regime. We investigate the properties of the objects whose exchange gives static screening lengths be reconstructing their correlated quark-antiquark structure.
Computer Simulation of Electron Positron Annihilation Processes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, y
2003-10-02
With the launching of the Next Linear Collider coming closer and closer, there is a pressing need for physicists to develop a fully-integrated computer simulation of e{sup +}e{sup -} annihilation process at center-of-mass energy of 1TeV. A simulation program acts as the template for future experiments. Either new physics will be discovered, or current theoretical uncertainties will shrink due to more accurate higher-order radiative correction calculations. The existence of an efficient and accurate simulation will help us understand the new data and validate (or veto) some of the theoretical models developed to explain new physics. It should handle well interfacesmore » between different sectors of physics, e.g., interactions happening at parton levels well above the QCD scale which are described by perturbative QCD, and interactions happening at much lower energy scale, which combine partons into hadrons. Also it should achieve competitive speed in real time when the complexity of the simulation increases. This thesis contributes some tools that will be useful for the development of such simulation programs. We begin our study by the development of a new Monte Carlo algorithm intended to perform efficiently in selecting weight-1 events when multiple parameter dimensions are strongly correlated. The algorithm first seeks to model the peaks of the distribution by features, adapting these features to the function using the EM algorithm. The representation of the distribution provided by these features is then improved using the VEGAS algorithm for the Monte Carlo integration. The two strategies mesh neatly into an effective multi-channel adaptive representation. We then present a new algorithm for the simulation of parton shower processes in high energy QCD. We want to find an algorithm which is free of negative weights, produces its output as a set of exclusive events, and whose total rate exactly matches the full Feynman amplitude calculation. Our strategy is to create the whole QCD shower as a tree structure generated by a multiple Poisson process. Working with the whole shower allows us to include correlations between gluon emissions from different sources. QCD destructive interference is controlled by the implementation of ''angular-ordering,'' as in the HERWIG Monte Carlo program. We discuss methods for systematic improvement of the approach to include higher order QCD effects.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Endres, Michael G.; Shindler, Andrea; Tiburzi, Brian C.
The commonly adopted approach for including electromagnetic interactions in lattice QCD simulations relies on using finite volume as the infrared regularization for QED. The long-range nature of the electromagnetic interaction, however, implies that physical quantities are susceptible to power-law finite volume corrections, which must be removed by performing costly simulations at multiple lattice volumes, followed by an extrapolation to the infinite volume limit. In this work, we introduce a photon mass as an alternative means for gaining control over infrared effects associated with electromagnetic interactions. We present findings for hadron mass shifts due to electromagnetic interactions (i.e., for the proton,more » neutron, charged and neutral kaon) and corresponding mass splittings, and compare the results with those obtained from conventional QCD+QED calculations. Results are reported for numerical studies of three flavor electroquenched QCD using ensembles corresponding to 800 MeV pions, ensuring that the only appreciable volume corrections arise from QED effects. The calculations are performed with three lattice volumes with spatial extents ranging from 3.4 - 6.7 fm. As a result, we find that for equal computing time (not including the generation of the lattice configurations), the electromagnetic mass shifts can be extracted from computations on a single (our smallest) lattice volume with comparable or better precision than the conventional approach.« less
Aoki, Y; Endrodi, G; Fodor, Z; Katz, S D; Szabó, K K
2006-10-12
Quantum chromodynamics (QCD) is the theory of the strong interaction, explaining (for example) the binding of three almost massless quarks into a much heavier proton or neutron--and thus most of the mass of the visible Universe. The standard model of particle physics predicts a QCD-related transition that is relevant for the evolution of the early Universe. At low temperatures, the dominant degrees of freedom are colourless bound states of hadrons (such as protons and pions). However, QCD is asymptotically free, meaning that at high energies or temperatures the interaction gets weaker and weaker, causing hadrons to break up. This behaviour underlies the predicted cosmological transition between the low-temperature hadronic phase and a high-temperature quark-gluon plasma phase (for simplicity, we use the word 'phase' to characterize regions with different dominant degrees of freedom). Despite enormous theoretical effort, the nature of this finite-temperature QCD transition (that is, first-order, second-order or analytic crossover) remains ambiguous. Here we determine the nature of the QCD transition using computationally demanding lattice calculations for physical quark masses. Susceptibilities are extrapolated to vanishing lattice spacing for three physical volumes, the smallest and largest of which differ by a factor of five. This ensures that a true transition should result in a dramatic increase of the susceptibilities. No such behaviour is observed: our finite-size scaling analysis shows that the finite-temperature QCD transition in the hot early Universe was not a real phase transition, but an analytic crossover (involving a rapid change, as opposed to a jump, as the temperature varied). As such, it will be difficult to find experimental evidence of this transition from astronomical observations.
Higgs pair production at NLO QCD for CP-violating Higgs sectors
NASA Astrophysics Data System (ADS)
Gröber, R.; Mühlleitner, M.; Spira, M.
2017-12-01
Higgs pair production through gluon fusion is an important process at the LHC to test the dynamics underlying electroweak symmetry breaking. Higgs sectors beyond the Standard Model (SM) can substantially modify this cross section through novel couplings not present in the SM or the on-shell production of new heavy Higgs bosons that subsequently decay into Higgs pairs. CP violation in the Higgs sector is important for the explanation of the observed matter-antimatter asymmetry through electroweak baryogenesis. In this work we compute the next-to-leading order (NLO) QCD corrections in the heavy top quark limit, including the effects of CP violation in the Higgs sector. We choose the effective theory (EFT) approach, which provides a rather model-independent way to explore New Physics (NP) effects by adding dimension-6 operators, both CP-conserving and CP-violating ones, to the SM Lagrangian. Furthermore, we perform the computation within a specific UV-complete model and choose as benchmark model the general 2-Higgs-Doublet Model with CP violation, the C2HDM. Depending on the dimension-6 coefficients, the relative NLO QCD corrections are affected by several per cent through the new CP-violating operators. This is also the case for SM-like Higgs pair production in the C2HDM, while the relative QCD corrections in the production of heavier C2HDM Higgs boson pairs deviate more strongly from the SM case. The absolute cross sections both in the EFT and the C2HDM can be modified by more than an order of magnitude. In particular, in the C2HDM the resonant production of Higgs pairs can by far exceed the SM cross section.
Born-Oppenheimer approximation in an effective field theory language
NASA Astrophysics Data System (ADS)
Brambilla, Nora; Krein, Gastão; Tarrús Castellà, Jaume; Vairo, Antonio
2018-01-01
The Born-Oppenheimer approximation is the standard tool for the study of molecular systems. It is founded on the observation that the energy scale of the electron dynamics in a molecule is larger than that of the nuclei. A very similar physical picture can be used to describe QCD states containing heavy quarks as well as light-quarks or gluonic excitations. In this work, we derive the Born-Oppenheimer approximation for QED molecular systems in an effective field theory framework by sequentially integrating out degrees of freedom living at energies above the typical energy scale where the dynamics of the heavy degrees of freedom occurs. In particular, we compute the matching coefficients of the effective field theory for the case of the H2+ diatomic molecule that are relevant to compute its spectrum up to O (m α5). Ultrasoft photon loops contribute at this order, being ultimately responsible for the molecular Lamb shift. In the effective field theory the scaling of all the operators is homogeneous, which facilitates the determination of all the relevant contributions, an observation that may become useful for high-precision calculations. Using the above case as a guidance, we construct under some conditions an effective field theory for QCD states formed by a color-octet heavy quark-antiquark pair bound with a color-octet light-quark pair or excited gluonic state, highlighting the similarities and differences between the QED and QCD systems. Assuming that the multipole expansion is applicable, we construct the heavy-quark potential up to next-to-leading order in the multipole expansion in terms of nonperturbative matching coefficients to be obtained from lattice QCD.
NASA Astrophysics Data System (ADS)
Bicudo, Pedro; Cardoso, Marco; Peters, Antje; Pflaumer, Martin; Wagner, Marc
2017-09-01
We study tetraquark resonances with lattice QCD potentials computed for a static b ¯b ¯ pair in the presence of two lighter quarks u d , the Born-Oppenheimer approximation and the emergent wave method. As a proof of concept we focus on the system with isospin I =0 , but consider different relative angular momenta l of the heavy quarks b ¯b ¯. For l =0 a bound state has already been predicted with quantum numbers I (JP)=0 (1+). Exploring various angular momenta we now compute the phase shifts and search for S and T matrix poles in the second Riemann sheet. We predict a tetraquark resonance for l =1 , decaying into two B mesons, with quantum numbers I (JP)=0 (1-) , mass m =10 57 6-4+4 MeV and decay width Γ =11 2-103+90 MeV .
Total Top-Quark Pair-Production Cross Section at Hadron Colliders Through O(αS4)
NASA Astrophysics Data System (ADS)
Czakon, Michał; Fiedler, Paul; Mitov, Alexander
2013-06-01
We compute the next-to-next-to-leading order (NNLO) quantum chromodynamics (QCD) correction to the total cross section for the reaction gg→tt¯+X. Together with the partonic channels we computed previously, the result derived in this Letter completes the set of NNLO QCD corrections to the total top pair-production cross section at hadron colliders. Supplementing the fixed order results with soft-gluon resummation with next-to-next-to-leading logarithmic accuracy, we estimate that the theoretical uncertainty of this observable due to unknown higher order corrections is about 3% at the LHC and 2.2% at the Tevatron. We observe a good agreement between the standard model predictions and the available experimental measurements. The very high theoretical precision of this observable allows a new level of scrutiny in parton distribution functions and new physics searches.
Total top-quark pair-production cross section at hadron colliders through O(αS(4)).
Czakon, Michał; Fiedler, Paul; Mitov, Alexander
2013-06-21
We compute the next-to-next-to-leading order (NNLO) quantum chromodynamics (QCD) correction to the total cross section for the reaction gg → tt + X. Together with the partonic channels we computed previously, the result derived in this Letter completes the set of NNLO QCD corrections to the total top pair-production cross section at hadron colliders. Supplementing the fixed order results with soft-gluon resummation with next-to-next-to-leading logarithmic accuracy, we estimate that the theoretical uncertainty of this observable due to unknown higher order corrections is about 3% at the LHC and 2.2% at the Tevatron. We observe a good agreement between the standard model predictions and the available experimental measurements. The very high theoretical precision of this observable allows a new level of scrutiny in parton distribution functions and new physics searches.
Soft evolution of multi-jet final states
Gerwick, Erik; Schumann, Steffen; Höche, Stefan; ...
2015-02-16
We present a new framework for computing resummed and matched distributions in processes with many hard QCD jets. The intricate color structure of soft gluon emission at large angles renders resummed calculations highly non-trivial in this case. We automate all ingredients necessary for the color evolution of the soft function at next-to-leading-logarithmic accuracy, namely the selection of the color bases and the projections of color operators and Born amplitudes onto those bases. Explicit results for all QCD processes with up to 2 → 5 partons are given. We also devise a new tree-level matching scheme for resummed calculations which exploitsmore » a quasi-local subtraction based on the Catani-Seymour dipole formalism. We implement both resummation and matching in the Sherpa event generator. As a proof of concept, we compute the resummed and matched transverse-thrust distribution for hadronic collisions.« less
Hadron electric polarizability from lattice QCD
NASA Astrophysics Data System (ADS)
Alexandru, Andrei
2017-09-01
Electromagnetic polarizabilities are important parameters for hadron structure, describing the response of the charge and current distributions inside the hadron to an external electromagnetic field. For most hadrons these quantities are poorly constrained experimentally since they can only be measured indirectly. Lattice QCD can be used to compute these quantities directly in terms of quark and gluons degrees of freedom, using the background field method. We present results for the neutron electric polarizability for two different quark masses, light enough to connect to chiral perturbation theory. These are currently the lightest quark masses used in polarizability studies. For each pion mass we compute the polarizability at four different volumes and perform an infinite volume extrapolation. We also discuss the effect of turning on the coupling between the background field and the sea quarks. A.A. is supported in part by the National Science Foundation CAREER Grant PHY-1151648 and by U.S. DOE Grant No. DE-FG02-95ER40907.
MS overline -on-shell quark mass relation up to four loops in QCD and a general SU (N ) gauge group
NASA Astrophysics Data System (ADS)
Marquard, Peter; Smirnov, Alexander V.; Smirnov, Vladimir A.; Steinhauser, Matthias; Wellmann, David
2016-10-01
We compute the relation between heavy quark masses defined in the modified minimal subtraction and the on-shell schemes. Detailed results are presented for all coefficients of the SU (Nc) color factors. The reduction of the four-loop on-shell integrals is performed for a general QCD gauge parameter. Altogether there are about 380 master integrals. Some of them are computed analytically, others with high numerical precision using Mellin-Barnes representations, and the rest numerically with the help of FIESTA. We discuss in detail the precise numerical evaluation of the four-loop master integrals. Updated relations between various short-distance masses and the MS ¯ quark mass to next-to-next-to-next-to-leading order accuracy are provided for the charm, bottom and top quarks. We discuss the dependence on the renormalization and factorization scale.
Job Management and Task Bundling
NASA Astrophysics Data System (ADS)
Berkowitz, Evan; Jansen, Gustav R.; McElvain, Kenneth; Walker-Loud, André
2018-03-01
High Performance Computing is often performed on scarce and shared computing resources. To ensure computers are used to their full capacity, administrators often incentivize large workloads that are not possible on smaller systems. Measurements in Lattice QCD frequently do not scale to machine-size workloads. By bundling tasks together we can create large jobs suitable for gigantic partitions. We discuss METAQ and mpi_jm, software developed to dynamically group computational tasks together, that can intelligently backfill to consume idle time without substantial changes to users' current workflows or executables.
Domain walls and the C P anomaly in softly broken supersymmetric QCD
NASA Astrophysics Data System (ADS)
Draper, Patrick
2018-04-01
In ordinary QCD with light, degenerate, fundamental flavors, C P symmetry is spontaneously broken at θ =π , and domain wall solutions connecting the vacua can be constructed in chiral perturbation theory. In some cases the breaking of C P saturates a 't Hooft anomaly, and anomaly inflow requires nontrivial massless excitations on the domain walls. Analogously, C P can be spontaneously broken in supersymmetric QCD (SQCD) with light flavors and small soft breaking parameters. We study C P breaking and domain walls in softly broken SQCD with Nf
Higgs Amplitudes from N=4 Supersymmetric Yang-Mills Theory.
Brandhuber, Andreas; Kostacińska, Martyna; Penante, Brenda; Travaglini, Gabriele
2017-10-20
Higgs plus multigluon amplitudes in QCD can be computed in an effective Lagrangian description. In the infinite top-mass limit, an amplitude with a Higgs boson and n gluons is computed by the form factor of the operator TrF^{2}. Up to two loops and for three gluons, its maximally transcendental part is captured entirely by the form factor of the protected stress tensor multiplet operator T_{2} in N=4 supersymmetric Yang-Mills theory. The next order correction involves the calculation of the form factor of the higher-dimensional, trilinear operator TrF^{3}. We present explicit results at two loops for three gluons, including the subleading transcendental terms derived from a particular descendant of the Konishi operator that contains TrF^{3}. These are expressed in terms of a few universal building blocks already identified in earlier calculations. We show that the maximally transcendental part of this quantity, computed in nonsupersymmetric Yang-Mills theory, is identical to the form factor of another protected operator, T_{3}, in the maximally supersymmetric theory. Our results suggest that the maximally transcendental part of Higgs amplitudes in QCD can be entirely computed through N=4 super Yang-Mills theory.
Z -Boson Production in Association with a Jet at Next-To-Next-To-Leading Order in Perturbative QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Boughezal, Radja; Campbell, John; Ellis, R. Keith
2016-04-01
We present the first complete calculation of Z-boson production in association with a jet in hadronic collisions through next-to-next-to-leading order in perturbative QCD. Our computation uses the recently proposed N-jettiness subtraction scheme to regulate the infrared divergences that appear in the real-emission contributions. We present phenomenological results for 13 TeV proton-proton collisions with fully realistic fiducial cuts on the final-state particles. The remaining theoretical uncertainties after the inclusion of our calculations are at the percent level, making the Z + jet channel ready for precision studies at the LHC run II.
QCD Coupling from a Nonperturbative Determination of the Three-Flavor Λ Parameter
Bruno, Mattia; Brida, Mattia Dalla; Fritzsch, Patrick; ...
2017-09-08
We present a lattice determination of the Λ parameter in three-flavor QCD and the strong coupling at the Z pole mass. Computing the nonperturbative running of the coupling in the range from 0.2 to 70 GeV, and using experimental input values for the masses and decay constants of the pion and the kaon, we obtain Λ(3)MS=341(12) MeV. The nonperturbative running up to very high energies guarantees that systematic effects associated with perturbation theory are well under control. Using the four-loop prediction for Λ(5)MS/Λ(3)MS yields α(5)MS(mZ)=0.11852(84).
The Chiral Separation Effect in quenched finite-density QCD
NASA Astrophysics Data System (ADS)
Puhr, Matthias; Buividovich, Pavel
2018-03-01
We present results of a study of the Chiral Separation Effect (CSE) in quenched finite-density QCD. Using a recently developed numerical method we calculate the conserved axial current for exactly chiral overlap fermions at finite density for the first time. We compute the anomalous transport coeffcient for the CSE in the confining and deconfining phase and investigate possible deviations from the universal value. In both phases we find that non-perturbative corrections to the CSE are absent and we reproduce the universal value for the transport coeffcient within small statistical errors. Our results suggest that the CSE can be used to determine the renormalisation factor of the axial current.
A four-dimensional model with the fermionic determinant exactly evaluated
NASA Astrophysics Data System (ADS)
Mignaco, J. A.; Rego Monteiro, M. A.
1986-07-01
A method is presented to compute the fermion determinant of some class of field theories. By this method the following results of the fermion determinant in two dimensions are easily recovered: (i) Schwinger model without reference to a particular gauge. (ii) QCD in the light-cone gauge. (iii) Gauge invariant result of QCD. The method is finally applied to give an analytical solution of the fermion determinant of a four-dimensional, non-abelian, Dirac-like theory with massless fermions interacting with an external vector field through a pseudo-vectorial coupling. Fellow of the Conselho Nacional de Desenvolvimento Cientifico e Tecnologico (CNPq), Brazil.
Configurational entropy and ρ and ϕ mesons production in QCD
NASA Astrophysics Data System (ADS)
Karapetyan, G.
2018-06-01
In the present work the electroproduction for diffractive ρ and ϕ mesons by considering AdS/QCD correspondence and Color Glass Condensate (CGC) approximation are studied with respect to the associated dipole cross section, whose parameters are studied and analysed in the framework of the configurational entropy. Our results suggest different quantum states of the nuclear matter, showing that the extremal points of the nuclear configurational entropy is able to reflect a true description of the ρ and ϕ mesons production, using current data concerning light quark masses. During the computations parameters, obtained in fitting procedure, coincide to the experimental within ∼ 0.1%.
Zimmermann's forest formula, infrared divergences and the QCD beta function
NASA Astrophysics Data System (ADS)
Herzog, Franz
2018-01-01
We review Zimmermann's forest formula, which solves Bogoliubov's recursive R-operation for the subtraction of ultraviolet divergences in perturbative Quantum Field Theory. We further discuss a generalisation of the R-operation which subtracts besides ultraviolet also Euclidean infrared divergences. This generalisation, which goes under the name of the R*-operation, can be used efficiently to compute renormalisation constants. We will discuss several results obtained by this method with focus on the QCD beta function at five loops as well as the application to hadronic Higgs boson decay rates at N4LO. This article summarizes a talk given at the Wolfhart Zimmermann Memorial Symposium.
Multipion systems in lattice QCD and the three-pion interaction.
Beane, Silas R; Detmold, William; Luu, Thomas C; Orginos, Kostas; Savage, Martin J; Torok, Aaron
2008-02-29
The ground-state energies of 2, 3, 4, and 5 pi(+)'s in a spatial volume V approximately (2.5 fm)(3) are computed with lattice QCD. By eliminating the leading contribution from three-pi(+) interactions, particular combinations of these n-pi(+) ground-state energies provide precise extractions of the pi(+)pi(+) scattering length in agreement with that obtained from calculations involving only two pi(+)'s. The three-pi(+) interaction can be isolated by forming other combinations of the n-pi(+) ground-state energies. We find a result that is consistent with a repulsive three-pi(+) interaction for m_(pi) less, similar352 MeV.
Counting the number of Feynman graphs in QCD
NASA Astrophysics Data System (ADS)
Kaneko, T.
2018-05-01
Information about the number of Feynman graphs for a given physical process in a given field theory is especially useful for confirming the result of a Feynman graph generator used in an automatic system of perturbative calculations. A method of counting the number of Feynman graphs with weight of symmetry factor was established based on zero-dimensional field theory, and was used in scalar theories and QED. In this article this method is generalized to more complicated models by direct calculation of generating functions on a computer algebra system. This method is applied to QCD with and without counter terms, where many higher order are being calculated automatically.
Suppression of Baryon Diffusion and Transport in a Baryon Rich Strongly Coupled Quark-Gluon Plasma
NASA Astrophysics Data System (ADS)
Rougemont, Romulo; Noronha, Jorge; Noronha-Hostler, Jacquelyn
2015-11-01
Five dimensional black hole solutions that describe the QCD crossover transition seen in (2 +1 ) -flavor lattice QCD calculations at zero and nonzero baryon densities are used to obtain predictions for the baryon susceptibility, baryon conductivity, baryon diffusion constant, and thermal conductivity of the strongly coupled quark-gluon plasma in the range of temperatures 130 MeV ≤T ≤300 MeV and baryon chemical potentials 0 ≤μB≤400 MeV . Diffusive transport is predicted to be suppressed in this region of the QCD phase diagram, which is consistent with the existence of a critical end point at larger baryon densities. We also calculate the fourth-order baryon susceptibility at zero baryon chemical potential and find quantitative agreement with recent lattice results. The baryon transport coefficients computed in this Letter can be readily implemented in state-of-the-art hydrodynamic codes used to investigate the dense QGP currently produced at RHIC's low energy beam scan.
Bevilacqua, G; Hartanto, H B; Kraus, M; Worek, M
2016-02-05
We present a complete description of top quark pair production in association with a jet in the dilepton channel. Our calculation is accurate to next-to-leading order (NLO) in QCD and includes all nonresonant diagrams, interferences, and off-shell effects of the top quark. Moreover, nonresonant and off-shell effects due to the finite W gauge boson width are taken into account. This calculation constitutes the first fully realistic NLO computation for top quark pair production with a final state jet in hadronic collisions. Numerical results for differential distributions as well as total cross sections are presented for the Large Hadron Collider at 8 TeV. With our inclusive cuts, NLO predictions reduce the unphysical scale dependence by more than a factor of 3 and lower the total rate by about 13% compared to leading-order QCD predictions. In addition, the size of the top quark off-shell effects is estimated to be below 2%.
Relativistic corrections to exclusive χc J+γ production from e+e- annihilation
NASA Astrophysics Data System (ADS)
Brambilla, Nora; Chen, Wen; Jia, Yu; Shtabovenko, Vladyslav; Vairo, Antonio
2018-05-01
We calculate in the nonrelativistic QCD (NRQCD) factorization framework all leading relativistic corrections to the exclusive production of χc J+γ in e+e- annihilation. In particular, we compute for the first time contributions induced by octet operators with a chromoelectric field. The matching coefficients multiplying production long distance matrix elements (LDMEs) are determined through perturbative matching between QCD and NRQCD at the amplitude level. Technical challenges encountered in the nonrelativistic expansion of the QCD amplitudes are discussed in detail. The main source of uncertainty comes from the not so well known LDMEs. Accounting for it, we provide the following estimates for the production cross sections at √{s }=10.6 GeV : σ (e+e-→χc 0+γ )=(1.4 ±0.3 ) fb , σ (e+e-→χc 1+γ )=(15.0 ±3.3 ) fb , and σ (e+e-→χc 2+γ )=(4.5 ±1.4 ) fb .
Next-to-Next-to-Leading-Order QCD Corrections to the Hadronic Width of Pseudoscalar Quarkonium
NASA Astrophysics Data System (ADS)
Feng, Feng; Jia, Yu; Sang, Wen-Long
2017-12-01
We compute the next-to-next-to-leading-order QCD corrections to the hadronic decay rates of the pseudoscalar quarkonia, at the lowest order in velocity expansion. The validity of nonrelativistic QCD (NRQCD) factorization for inclusive quarkonium decay process, for the first time, is verified to relative order αs2. As a by-product, the renormalization group equation of the leading NRQCD four-fermion operator O1(1S0 ) is also deduced to this perturbative order. By incorporating this new piece of correction together with available relativistic corrections, we find that there exists severe tension between the state-of-the-art NRQCD predictions and the measured ηc hadronic width and, in particular, the branching fraction of ηc→γ γ . NRQCD appears to be capable of accounting for ηb hadronic decay to a satisfactory degree, and our most refined prediction is Br(ηb→γ γ )=(4.8 ±0.7 )×10-5.
Nuclear Physics and Lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Beane, Silas
2003-11-01
Impressive progress is currently being made in computing properties and interac- tions of the low-lying hadrons using lattice QCD. However, cost limitations will, for the foreseeable future, necessitate the use of quark masses, Mq, that are signif- icantly larger than those of nature, lattice spacings, a, that are not significantly smaller than the physical scale of interest, and lattice sizes, L, that are not sig- nificantly larger than the physical scale of interest. Extrapolations in the quark masses, lattice spacing and lattice volume are therefore required. The hierarchy of mass scales is: L 1 j Mq j â ºC jmore » a 1 . The appropriate EFT for incorporating the light quark masses, the finite lattice spacing and the lattice size into hadronic observables is C-PT, which provides systematic expansions in the small parame- ters e m L, 1/ Lâ ºC, p/â ºC, Mq/â ºC and aâ ºC . The lattice introduces other unphysical scales as well. Lattice QCD quarks will increasingly be artificially separated« less
Charmed bottom baryon spectroscopy from lattice QCD
Brown, Zachary S.; Detmold, William; Meinel, Stefan; ...
2014-11-19
In this study, we calculate the masses of baryons containing one, two, or three heavy quarks using lattice QCD. We consider all possible combinations of charm and bottom quarks, and compute a total of 36 different states with J P = 1/2 + and J P = 3/2 +. We use domain-wall fermions for the up, down, and strange quarks, a relativistic heavy-quark action for the charm quarks, and nonrelativistic QCD for the bottom quarks. Our analysis includes results from two different lattice spacings and seven different pion masses. We perform extrapolations of the baryon masses to the continuum limitmore » and to the physical pion mass using SU(4|2) heavy-hadron chiral perturbation theory including 1/m Q and finite-volume effects. For the 14 singly heavy baryons that have already been observed, our results agree with the experimental values within the uncertainties. We compare our predictions for the hitherto unobserved states with other lattice calculations and quark-model studies.« less
$$B\\to Kl^+l^-$$ decay form factors from three-flavor lattice QCD
Bailey, Jon A.
2016-01-27
We compute the form factors for the B → Kl +l - semileptonic decay process in lattice QCD using gauge-field ensembles with 2+1 flavors of sea quark, generated by the MILC Collaboration. The ensembles span lattice spacings from 0.12 to 0.045 fm and have multiple sea-quark masses to help control the chiral extrapolation. The asqtad improved staggered action is used for the light valence and sea quarks, and the clover action with the Fermilab interpretation is used for the heavy b quark. We present results for the form factors f+(q 2), f 0(q 2), and f T(q 2), where q 2more » is the momentum transfer, together with a comprehensive examination of systematic errors. Lattice QCD determines the form factors for a limited range of q 2, and we use the model-independent z expansion to cover the whole kinematically allowed range. We present our final form-factor results as coefficients of the z expansion and the correlations between them, where the errors on the coefficients include statistical and all systematic uncertainties. Lastly, we use this complete description of the form factors to test QCD predictions of the form factors at high and low q 2.« less
NASA Technical Reports Server (NTRS)
Reinsch, K. G. (Editor); Schmidt, W. (Editor); Ecer, A. (Editor); Haeuser, Jochem (Editor); Periaux, J. (Editor)
1992-01-01
A conference was held on parallel computational fluid dynamics and produced related papers. Topics discussed in these papers include: parallel implicit and explicit solvers for compressible flow, parallel computational techniques for Euler and Navier-Stokes equations, grid generation techniques for parallel computers, and aerodynamic simulation om massively parallel systems.
Structure constant of twist-2 light-ray operators in the Regge limit
Balitsky, Ian; Kazakov, Vladimir; Sobko, Evgeny
2016-03-11
We compute the correlation function of three twist-2 operators in N = 4 SYM in the leading BFKL approximation at any N c. In this limit, the result is applicable to other gauge theories, including QCD.
T -matrix approach to quark-gluon plasma
NASA Astrophysics Data System (ADS)
Liu, Shuai Y. F.; Rapp, Ralf
2018-03-01
A self-consistent thermodynamic T -matrix approach is deployed to study the microscopic properties of the quark-gluon plasma (QGP), encompassing both light- and heavy-parton degrees of freedom in a unified framework. The starting point is a relativistic effective Hamiltonian with a universal color force. The input in-medium potential is quantitatively constrained by computing the heavy-quark (HQ) free energy from the static T -matrix and fitting it to pertinent lattice-QCD (lQCD) data. The corresponding T -matrix is then applied to compute the equation of state (EoS) of the QGP in a two-particle irreducible formalism, including the full off-shell properties of the selfconsistent single-parton spectral functions and their two-body interaction. In particular, the skeleton diagram functional is fully resummed to account for emerging bound and scattering states as the critical temperature is approached from above. We find that the solution satisfying three sets of lQCD data (EoS, HQ free energy, and quarkonium correlator ratios) is not unique. As limiting cases we discuss a weakly coupled solution, which features color potentials close to the free energy, relatively sharp quasiparticle spectral functions and weak hadronic resonances near Tc, and a strongly coupled solution with a strong color potential (much larger than the free energy), resulting in broad nonquasiparticle parton spectral functions and strong hadronic resonance states which dominate the EoS when approaching Tc.
Subtraction with hadronic initial states at NLO: an NNLO-compatible scheme
NASA Astrophysics Data System (ADS)
Somogyi, Gábor
2009-05-01
We present an NNLO-compatible subtraction scheme for computing QCD jet cross sections of hadron-initiated processes at NLO accuracy. The scheme is constructed specifically with those complications in mind, that emerge when extending the subtraction algorithm to next-to-next-to-leading order. It is therefore possible to embed the present scheme in a full NNLO computation without any modifications.
The Research of the Parallel Computing Development from the Angle of Cloud Computing
NASA Astrophysics Data System (ADS)
Peng, Zhensheng; Gong, Qingge; Duan, Yanyu; Wang, Yun
2017-10-01
Cloud computing is the development of parallel computing, distributed computing and grid computing. The development of cloud computing makes parallel computing come into people’s lives. Firstly, this paper expounds the concept of cloud computing and introduces two several traditional parallel programming model. Secondly, it analyzes and studies the principles, advantages and disadvantages of OpenMP, MPI and Map Reduce respectively. Finally, it takes MPI, OpenMP models compared to Map Reduce from the angle of cloud computing. The results of this paper are intended to provide a reference for the development of parallel computing.
Spatial data analytics on heterogeneous multi- and many-core parallel architectures using python
Laura, Jason R.; Rey, Sergio J.
2017-01-01
Parallel vector spatial analysis concerns the application of parallel computational methods to facilitate vector-based spatial analysis. The history of parallel computation in spatial analysis is reviewed, and this work is placed into the broader context of high-performance computing (HPC) and parallelization research. The rise of cyber infrastructure and its manifestation in spatial analysis as CyberGIScience is seen as a main driver of renewed interest in parallel computation in the spatial sciences. Key problems in spatial analysis that have been the focus of parallel computing are covered. Chief among these are spatial optimization problems, computational geometric problems including polygonization and spatial contiguity detection, the use of Monte Carlo Markov chain simulation in spatial statistics, and parallel implementations of spatial econometric methods. Future directions for research on parallelization in computational spatial analysis are outlined.
On the interface between perturbative and nonperturbative QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Deur, Alexandre; Brodsky, Stanley J.; de Teramond, Guy F.
2016-04-04
The QCD running couplingmore » $$\\alpha_s(Q^2)$$ sets the strength of the interactions of quarks and gluons as a function of the momentum transfer $Q$. The $Q^2$ dependence of the coupling is required to describe hadronic interactions at both large and short distances. In this article we adopt the light-front holographic approach to strongly-coupled QCD, a formalism which incorporates confinement, predicts the spectroscopy of hadrons composed of light quarks, and describes the low-$Q^2$ analytic behavior of the strong coupling $$\\alpha_s(Q^2)$$. The high-$Q^2$ dependence of the coupling $$\\alpha_s(Q^2)$$ is specified by perturbative QCD and its renormalization group equation. The matching of the high and low $Q^2$ regimes of $$\\alpha_s(Q^2)$$ then determines the scale $$Q_0$$ which sets the interface between perturbative and nonperturbative hadron dynamics. The value of $$Q_0$$ can be used to set the factorization scale for DGLAP evolution of hadronic structure functions and the ERBL evolution of distribution amplitudes. We discuss the scheme-dependence of the value of $$Q_0$$ and the infrared fixed-point of the QCD coupling. Our analysis is carried out for the $$\\bar{MS}$$, $$g_1$$, $MOM$ and $V$ renormalization schemes. Our results show that the discrepancies on the value of $$\\alpha_s$$ at large distance seen in the literature can be explained by different choices of renormalization schemes. Lastly, we also provide the formulae to compute $$\\alpha_s(Q^2)$$ over the entire range of space-like momentum transfer for the different renormalization schemes discussed in this article.« less
Testing Quantum Chromodynamics with Antiprotons
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, S.
2004-10-21
The antiproton storage ring HESR to be constructed at GSI will open up a new range of perturbative and nonperturbative tests of QCD in exclusive and inclusive reactions. I discuss 21 tests of QCD using antiproton beams which can illuminate novel features of QCD. The proposed experiments include the formation of exotic hadrons, measurements of timelike generalized parton distributions, the production of charm at threshold, transversity measurements in Drell-Yan reactions, and searches for single-spin asymmetries. The interactions of antiprotons in nuclear targets will allow tests of exotic nuclear phenomena such as color transparency, hidden color, reduced nuclear amplitudes, and themore » non-universality of nuclear antishadowing. The central tool used in these lectures are light-front Fock state wavefunctions which encode the bound-state properties of hadrons in terms of their quark and gluon degrees of freedom at the amplitude level. The freedom to choose the light-like quantization four-vector provides an explicitly covariant formulation of light-front quantization and can be used to determine the analytic structure of light-front wave functions. QCD becomes scale free and conformally symmetric in the analytic limit of zero quark mass and zero {beta} function. This ''conformal correspondence principle'' determines the form of the expansion polynomials for distribution amplitudes and the behavior of non-perturbative wavefunctions which control hard exclusive processes at leading twist. The conformal template also can be used to derive commensurate scale relations which connect observables in QCD without scale or scheme ambiguity. The AdS/CFT correspondence of large N{sub C} supergravity theory in higher-dimensional anti-de Sitter space with supersymmetric QCD in 4-dimensional space-time has important implications for hadron phenomenology in the conformal limit, including the nonperturbative derivation of counting rules for exclusive processes and the behavior of structure functions at large x{sub bj}. String/gauge duality also predicts the QCD power-law fall-off of light-front Fock-state hadronic wavefunctions with arbitrary orbital angular momentum at high momentum transfer. I also review recent work which shows that the diffractive component of deep inelastic scattering, single spin asymmetries, as well as nuclear shadowing and antishadowing, cannot be computed from the LFWFs of hadrons in isolation.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Molnar, Denes
2014-04-14
The Section below summarizes research activities and achievements during the first four years of the PI’s Early Career Research Project (ECRP). Two main areas have been advanced: i) radiative 3 ↔ 2 radiative transport, via development of a new computer code MPC/Grid that solves the Boltzmann transport equation in full 6+1D (3X+3V+time) on both single-CPU and parallel computers; ii) development of a self-consistent framework to convert viscous fluids to particles, and application of this framework to relativistic heavy-ion collisions, in particular, determination of the shear viscosity. Year 5 of the ECRP is under a separate award number, and therefore itmore » has its own report document ’Final Technical Report for Year 5 of the Early Career Research Project “Viscosity and equation of state of hot and dense QCDmatter”’ (award DE-SC0008028). The PI’s group was also part of the DOE JET Topical Collaboration, a multi-institution project that overlapped in time significantly with the ECRP. Purdue achievements as part of the JET Topical Collaboration are in a separate report “Final Technical Report summarizing Purdue research activities as part of the DOE JET Topical Collaboration” (award DE-SC0004077).« less
Broadcasting collective operation contributions throughout a parallel computer
Faraj, Ahmad [Rochester, MN
2012-02-21
Methods, systems, and products are disclosed for broadcasting collective operation contributions throughout a parallel computer. The parallel computer includes a plurality of compute nodes connected together through a data communications network. Each compute node has a plurality of processors for use in collective parallel operations on the parallel computer. Broadcasting collective operation contributions throughout a parallel computer according to embodiments of the present invention includes: transmitting, by each processor on each compute node, that processor's collective operation contribution to the other processors on that compute node using intra-node communications; and transmitting on a designated network link, by each processor on each compute node according to a serial processor transmission sequence, that processor's collective operation contribution to the other processors on the other compute nodes using inter-node communications.
Application Portable Parallel Library
NASA Technical Reports Server (NTRS)
Cole, Gary L.; Blech, Richard A.; Quealy, Angela; Townsend, Scott
1995-01-01
Application Portable Parallel Library (APPL) computer program is subroutine-based message-passing software library intended to provide consistent interface to variety of multiprocessor computers on market today. Minimizes effort needed to move application program from one computer to another. User develops application program once and then easily moves application program from parallel computer on which created to another parallel computer. ("Parallel computer" also include heterogeneous collection of networked computers). Written in C language with one FORTRAN 77 subroutine for UNIX-based computers and callable from application programs written in C language or FORTRAN 77.
Z-boson production in association with a jet at next-to-next-to-leading order in perturbative QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Boughezal, Radja; Campbell, John M.; Ellis, R. Keith
2016-04-14
Here, we present the first complete calculation of Z-boson production in association with a jet in hadronic collisions through next-to-next-to-leading order in perturbative QCD. Our computation uses the recently proposed N-jettiness subtraction scheme to regulate the infrared divergences that appear in the real-emission contributions. We present phenomenological results for 13 TeV proton-proton collisions with fully realistic fiducial cuts on the final-state particles. The remaining theoretical uncertainties after the inclusion of our calculations are at the percent level, making the Z+jet channel ready for precision studies at the LHC run II.
Electroweak and QCD corrections to top-pair hadroproduction in association with heavy bosons
Frixione, Stefano; Hirschi, V.; Pagani, D.; ...
2015-06-26
Here, we compute the contribution of order α S 2α 2 to the cross section of a top-antitop pair in association with at least one heavy Standard Model boson — Z, W ±, and Higgs — by including all effects of QCD, QED, and weak origin and by working in the automated MadGraph5_aMC@NLO framework. Furthermore, this next-to-leading order contribution is then combined with that of order αS3α, and with the two dominant lowest-order ones, α S 2α and α Sα 2, to obtain phenomenological results relevant to a 8, 13, and 100 TeV pp collider.
NNLO QCD predictions for fully-differential top-quark pair production at the Tevatron
NASA Astrophysics Data System (ADS)
Czakon, Michal; Fiedler, Paul; Heymes, David; Mitov, Alexander
2016-05-01
We present a comprehensive study of differential distributions for Tevatron top-pair events at the level of stable top quarks. All calculations are performed in NNLO QCD with the help of a fully differential partonic Monte-Carlo and are exact at this order in perturbation theory. We present predictions for all kinematic distributions for which data exists. Particular attention is paid on the top-quark forward-backward asymmetry which we study in detail. We compare the NNLO results with existing approximate NNLO predictions as well as differential distributions computed with different parton distribution sets. Theory errors are significantly smaller than current experimental ones with overall agreement between theory and data.
Electric Dipole Moment of the Neutron from 2+1 Flavor Lattice QCD.
Guo, F-K; Horsley, R; Meissner, U-G; Nakamura, Y; Perlt, H; Rakow, P E L; Schierholz, G; Schiller, A; Zanotti, J M
2015-08-07
We compute the electric dipole moment d(n) of the neutron from a fully dynamical simulation of lattice QCD with 2+1 flavors of clover fermions and nonvanishing θ term. The latter is rotated into a pseudoscalar density in the fermionic action using the axial anomaly. To make the action real, the vacuum angle θ is taken to be purely imaginary. The physical value of dd(n) is obtained by analytic continuation. We find d(n)=-3.9(2)(9)×10(-16) θ e cm, which, when combined with the experimental limit on d(n), leads to the upper bound |θ|≲7.4×10(-11).
Helicity amplitudes for QCD with massive quarks
NASA Astrophysics Data System (ADS)
Ochirov, Alexander
2018-04-01
The novel massive spinor-helicity formalism of Arkani-Hamed, Huang and Huang provides an elegant way to calculate scattering amplitudes in quantum chromodynamics for arbitrary quark spin projections. In this note we compute two families of tree-level QCD amplitudes with one massive quark pair and n - 2 gluons. The two cases include all gluons with identical helicity and one opposite-helicity gluon being color-adjacent to one of the quarks. Our results naturally incorporate the previously known amplitudes for both quark spins quantized along one of the gluonic momenta. In the all-multiplicity formulae presented here the spin quantization axes can be tuned at will, which includes the case of the definite-helicity quark states.
Strangeness contribution to the proton spin from lattice QCD.
Bali, Gunnar S; Collins, Sara; Göckeler, Meinulf; Horsley, Roger; Nakamura, Yoshifumi; Nobile, Andrea; Pleiter, Dirk; Rakow, P E L; Schäfer, Andreas; Schierholz, Gerrit; Zanotti, James M
2012-06-01
We compute the strangeness and light-quark contributions Δs, Δu, and Δd to the proton spin in n(f)=2 lattice QCD at a pion mass of about 285 MeV and at a lattice spacing a≈0.073 fm, using the nonperturbatively improved Sheikholeslami-Wohlert Wilson action. We carry out the renormalization of these matrix elements, which involves mixing between contributions from different quark flavors. Our main result is the small negative value Δs(MS)(√(7.4) GeV)=-0.020(10)(4) of the strangeness contribution to the nucleon spin. The second error is an estimate of the uncertainty, due to the missing extrapolation to the physical point.
Higgs Boson Production in Association with a Jet at Next-to-Next-to-Leading Order.
Boughezal, Radja; Caola, Fabrizio; Melnikov, Kirill; Petriello, Frank; Schulze, Markus
2015-08-21
We present precise predictions for Higgs boson production in association with a jet. We work in the Higgs effective field theory framework and compute next-to-next-to-leading order QCD corrections to the gluon-gluon and quark-gluon channels, which is sufficient for reliable LHC phenomenology. We present fully differential results as well as total cross sections for the LHC. Our next-to-next-to-leading order predictions reduce the unphysical scale dependence by more than a factor of 2 and enhance the total rate by about twenty percent compared to next-to-leading order QCD predictions. Our results demonstrate for the first time satisfactory convergence of the perturbative series.
An estimate for the thermal photon rate from lattice QCD
NASA Astrophysics Data System (ADS)
Brandt, Bastian B.; Francis, Anthony; Harris, Tim; Meyer, Harvey B.; Steinberg, Aman
2018-03-01
We estimate the production rate of photons by the quark-gluon plasma in lattice QCD. We propose a new correlation function which provides better control over the systematic uncertainty in estimating the photon production rate at photon momenta in the range πT/2 to 2πT. The relevant Euclidean vector current correlation functions are computed with Nf = 2 Wilson clover fermions in the chirally-symmetric phase. In order to estimate the photon rate, an ill-posed problem for the vector-channel spectral function must be regularized. We use both a direct model for the spectral function and a modelindependent estimate from the Backus-Gilbert method to give an estimate for the photon rate.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, Stanley J.; /SLAC /Southern Denmark U., CP3-Origins
2011-08-12
I review a number of topics where conventional wisdom in hadron physics has been challenged. For example, hadrons can be produced at large transverse momentum directly within a hard higher-twist QCD subprocess, rather than from jet fragmentation. Such 'direct' processes can explain the deviations from perturbative QCD predictions in measurements of inclusive hadron cross sections at fixed x{sub T} = 2p{sub T}/{radical}s, as well as the 'baryon anomaly', the anomalously large proton-to-pion ratio seen in high centrality heavy ion collisions. Initial-state and final-state interactions of the struck quark, the soft-gluon rescattering associated with its Wilson line, lead to Bjorken-scaling single-spinmore » asymmetries, diffractive deep inelastic scattering, the breakdown of the Lam-Tung relation in Drell-Yan reactions, as well as nuclear shadowing and antishadowing. The Gribov-Glauber theory predicts that antishadowing of nuclear structure functions is not universal, but instead depends on the flavor quantum numbers of each quark and antiquark, thus explaining the anomalous nuclear dependence measured in deep-inelastic neutrino scattering. Since shadowing and antishadowing arise from the physics of leading-twist diffractive deep inelastic scattering, one cannot attribute such phenomena to the structure of the nucleus itself. It is thus important to distinguish 'static' structure functions, the probability distributions computed from the square of the target light-front wavefunctions, versus 'dynamical' structure functions which include the effects of the final-state rescattering of the struck quark. The importance of the J = 0 photon-quark QCD contact interaction in deeply virtual Compton scattering is also emphasized. The scheme-independent BLM method for setting the renormalization scale is discussed. Eliminating the renormalization scale ambiguity greatly improves the precision of QCD predictions and increases the sensitivity of searches for new physics at the LHC. Other novel features of QCD are discussed, including the consequences of confinement for quark and gluon condensates.« less
Renormalizable Quantum Field Theories in the Large -n Limit
NASA Astrophysics Data System (ADS)
Guruswamy, Sathya
1995-01-01
In this thesis, we study two examples of renormalizable quantum field theories in the large-N limit. Chapter one is a general introduction describing physical motivations for studying such theories. In chapter two, we describe the large-N method in field theory and discuss the pioneering work of 't Hooft in large-N two-dimensional Quantum Chromodynamics (QCD). In chapter three we study a spherically symmetric approximation to four-dimensional QCD ('spherical QCD'). We recast spherical QCD into a bilocal (constrained) theory of hadrons which in the large-N limit is equivalent to large-N spherical QCD for all energy scales. The linear approximation to this theory gives an eigenvalue equation which is the analogue of the well-known 't Hooft's integral equation in two dimensions. This eigenvalue equation is a scale invariant one and therefore leads to divergences in the theory. We give a non-perturbative renormalization prescription to cure this and obtain a beta function which shows that large-N spherical QCD is asymptotically free. In chapter four, we review the essentials of conformal field theories in two and higher dimensions, particularly in the context of critical phenomena. In chapter five, we study the O(N) non-linear sigma model on three-dimensional curved spaces in the large-N limit and show that there is a non-trivial ultraviolet stable critical point at which it becomes conformally invariant. We study this model at this critical point on examples of spaces of constant curvature and compute the mass gap in the theory, the free energy density (which turns out to be a universal function of the information contained in the geometry of the manifold) and the two-point correlation functions. The results we get give an indication that this model is an example of a three-dimensional analogue of a rational conformal field theory. A conclusion with a brief summary and remarks follows at the end.
Performance of parallel computation using CUDA for solving the one-dimensional elasticity equations
NASA Astrophysics Data System (ADS)
Darmawan, J. B. B.; Mungkasi, S.
2017-01-01
In this paper, we investigate the performance of parallel computation in solving the one-dimensional elasticity equations. Elasticity equations are usually implemented in engineering science. Solving these equations fast and efficiently is desired. Therefore, we propose the use of parallel computation. Our parallel computation uses CUDA of the NVIDIA. Our research results show that parallel computation using CUDA has a great advantage and is powerful when the computation is of large scale.
Increasing processor utilization during parallel computation rundown
NASA Technical Reports Server (NTRS)
Jones, W. H.
1986-01-01
Some parallel processing environments provide for asynchronous execution and completion of general purpose parallel computations from a single computational phase. When all the computations from such a phase are complete, a new parallel computational phase is begun. Depending upon the granularity of the parallel computations to be performed, there may be a shortage of available work as a particular computational phase draws to a close (computational rundown). This can result in the waste of computing resources and the delay of the overall problem. In many practical instances, strict sequential ordering of phases of parallel computation is not totally required. In such cases, the beginning of one phase can be correctly computed before the end of a previous phase is completed. This allows additional work to be generated somewhat earlier to keep computing resources busy during each computational rundown. The conditions under which this can occur are identified and the frequency of occurrence of such overlapping in an actual parallel Navier-Stokes code is reported. A language construct is suggested and possible control strategies for the management of such computational phase overlapping are discussed.
Broadcasting a message in a parallel computer
Berg, Jeremy E [Rochester, MN; Faraj, Ahmad A [Rochester, MN
2011-08-02
Methods, systems, and products are disclosed for broadcasting a message in a parallel computer. The parallel computer includes a plurality of compute nodes connected together using a data communications network. The data communications network optimized for point to point data communications and is characterized by at least two dimensions. The compute nodes are organized into at least one operational group of compute nodes for collective parallel operations of the parallel computer. One compute node of the operational group assigned to be a logical root. Broadcasting a message in a parallel computer includes: establishing a Hamiltonian path along all of the compute nodes in at least one plane of the data communications network and in the operational group; and broadcasting, by the logical root to the remaining compute nodes, the logical root's message along the established Hamiltonian path.
Force user's manual: A portable, parallel FORTRAN
NASA Technical Reports Server (NTRS)
Jordan, Harry F.; Benten, Muhammad S.; Arenstorf, Norbert S.; Ramanan, Aruna V.
1990-01-01
The use of Force, a parallel, portable FORTRAN on shared memory parallel computers is described. Force simplifies writing code for parallel computers and, once the parallel code is written, it is easily ported to computers on which Force is installed. Although Force is nearly the same for all computers, specific details are included for the Cray-2, Cray-YMP, Convex 220, Flex/32, Encore, Sequent, Alliant computers on which it is installed.
Bärnreuther, Peter; Czakon, Michał; Mitov, Alexander
2012-09-28
We compute the next-to-next-to-leading order QCD corrections to the partonic reaction that dominates top-pair production at the Tevatron. This is the first ever next-to-next-to-leading order calculation of an observable with more than two colored partons and/or massive fermions at hadron colliders. Augmenting our fixed order calculation with soft-gluon resummation through next-to-next-to-leading logarithmic accuracy, we observe that the predicted total inclusive cross section exhibits a very small perturbative uncertainty, estimated at ±2.7%. We expect that once all subdominant partonic reactions are accounted for, and work in this direction is ongoing, the perturbative theoretical uncertainty for this observable could drop below ±2%. Our calculation demonstrates the power of our computational approach and proves it can be successfully applied to all processes at hadron colliders for which high-precision analyses are needed.
NASA Astrophysics Data System (ADS)
Bärnreuther, Peter; Czakon, Michał; Mitov, Alexander
2012-09-01
We compute the next-to-next-to-leading order QCD corrections to the partonic reaction that dominates top-pair production at the Tevatron. This is the first ever next-to-next-to-leading order calculation of an observable with more than two colored partons and/or massive fermions at hadron colliders. Augmenting our fixed order calculation with soft-gluon resummation through next-to-next-to-leading logarithmic accuracy, we observe that the predicted total inclusive cross section exhibits a very small perturbative uncertainty, estimated at ±2.7%. We expect that once all subdominant partonic reactions are accounted for, and work in this direction is ongoing, the perturbative theoretical uncertainty for this observable could drop below ±2%. Our calculation demonstrates the power of our computational approach and proves it can be successfully applied to all processes at hadron colliders for which high-precision analyses are needed.
Spin polarized photons from an axially charged plasma at weak coupling: Complete leading order
Mamo, Kiminad A.; Yee, Ho-Ung
2016-03-24
In the presence of (approximately conserved) axial charge in the QCD plasma at finite temperature, the emitted photons are spin aligned, which is a unique P- and CP-odd signature of axial charge in the photon emission observables. We compute this “P-odd photon emission rate” in a weak coupling regime at a high temperature limit to complete leading order in the QCD coupling constant: the leading log as well as the constant under the log. As in the P-even total emission rate in the literature, the computation of the P-odd emission rate at leading order consists of three parts: (1) Comptonmore » and pair annihilation processes with hard momentum exchange, (2) soft t- and u-channel contributions with hard thermal loop resummation, (3) Landau-Pomeranchuk-Migdal resummation of collinear bremsstrahlung and pair annihilation. In conclusion, we present analytical and numerical evaluations of these contributions to our P-odd photon emission rate observable.« less
Exploratory Lattice QCD Study of the Rare Kaon Decay K^{+}→π^{+}νν[over ¯].
Bai, Ziyuan; Christ, Norman H; Feng, Xu; Lawson, Andrew; Portelli, Antonin; Sachrajda, Christopher T
2017-06-23
We report a first, complete lattice QCD calculation of the long-distance contribution to the K^{+}→π^{+}νν[over ¯] decay within the standard model. This is a second-order weak process involving two four-Fermi operators that is highly sensitive to new physics and being studied by the NA62 experiment at CERN. While much of this decay comes from perturbative, short-distance physics, there is a long-distance part, perhaps as large as the planned experimental error, which involves nonperturbative phenomena. The calculation presented here, with unphysical quark masses, demonstrates that this contribution can be computed using lattice methods by overcoming three technical difficulties: (i) a short-distance divergence that results when the two weak operators approach each other, (ii) exponentially growing, unphysical terms that appear in Euclidean, second-order perturbation theory, and (iii) potentially large finite-volume effects. A follow-on calculation with physical quark masses and controlled systematic errors will be possible with the next generation of computers.
One-loop calculations in Supersymmetric Lattice QCD
NASA Astrophysics Data System (ADS)
Costa, M.; Panagopoulos, H.
2017-03-01
We study the self energies of all particles which appear in a lattice regularization of supersymmetric QCD (N = 1). We compute, perturbatively to one-loop, the relevant two-point Green's functions using both the dimensional and the lattice regularizations. Our lattice formulation employs the Wilson fermion acrion for the gluino and quark fields. The gauge group that we consider is SU(Nc) while the number of colors, Nc and the number of flavors, Nf , are kept as generic parameters. We have also searched for relations among the propagators which are computed from our one-loop results. We have obtained analytic expressions for the renormalization functions of the quark field (Zψ), gluon field (Zu), gluino field (Zλ) and squark field (ZA±). We present here results from dimensional regularization, relegating to a forthcoming publication [1] our results along with a more complete list of references. Part of the lattice study regards also the renormalization of quark bilinear operators which, unlike the nonsupersymmetric case, exhibit a rich pattern of operator mixing at the quantum level.
Exploratory Lattice QCD Study of the Rare Kaon Decay K+→π+ν ν ¯
NASA Astrophysics Data System (ADS)
Bai, Ziyuan; Christ, Norman H.; Feng, Xu; Lawson, Andrew; Portelli, Antonin; Sachrajda, Christopher T.; Rbc-Ukqcd Collaboration
2017-06-01
We report a first, complete lattice QCD calculation of the long-distance contribution to the K+→π+ν ν ¯ decay within the standard model. This is a second-order weak process involving two four-Fermi operators that is highly sensitive to new physics and being studied by the NA62 experiment at CERN. While much of this decay comes from perturbative, short-distance physics, there is a long-distance part, perhaps as large as the planned experimental error, which involves nonperturbative phenomena. The calculation presented here, with unphysical quark masses, demonstrates that this contribution can be computed using lattice methods by overcoming three technical difficulties: (i) a short-distance divergence that results when the two weak operators approach each other, (ii) exponentially growing, unphysical terms that appear in Euclidean, second-order perturbation theory, and (iii) potentially large finite-volume effects. A follow-on calculation with physical quark masses and controlled systematic errors will be possible with the next generation of computers.
Architecture Adaptive Computing Environment
NASA Technical Reports Server (NTRS)
Dorband, John E.
2006-01-01
Architecture Adaptive Computing Environment (aCe) is a software system that includes a language, compiler, and run-time library for parallel computing. aCe was developed to enable programmers to write programs, more easily than was previously possible, for a variety of parallel computing architectures. Heretofore, it has been perceived to be difficult to write parallel programs for parallel computers and more difficult to port the programs to different parallel computing architectures. In contrast, aCe is supportable on all high-performance computing architectures. Currently, it is supported on LINUX clusters. aCe uses parallel programming constructs that facilitate writing of parallel programs. Such constructs were used in single-instruction/multiple-data (SIMD) programming languages of the 1980s, including Parallel Pascal, Parallel Forth, C*, *LISP, and MasPar MPL. In aCe, these constructs are extended and implemented for both SIMD and multiple- instruction/multiple-data (MIMD) architectures. Two new constructs incorporated in aCe are those of (1) scalar and virtual variables and (2) pre-computed paths. The scalar-and-virtual-variables construct increases flexibility in optimizing memory utilization in various architectures. The pre-computed-paths construct enables the compiler to pre-compute part of a communication operation once, rather than computing it every time the communication operation is performed.
Leptonic-decay-constant ratio f(K+)/f(π+) from lattice QCD with physical light quarks.
Bazavov, A; Bernard, C; DeTar, C; Foley, J; Freeman, W; Gottlieb, Steven; Heller, U M; Hetrick, J E; Kim, J; Laiho, J; Levkova, L; Lightman, M; Osborn, J; Qiu, S; Sugar, R L; Toussaint, D; Van de Water, R S; Zhou, R
2013-04-26
A calculation of the ratio of leptonic decay constants f(K+)/f(π+) makes possible a precise determination of the ratio of Cabibbo-Kobayashi-Maskawa (CKM) matrix elements |V(us)|/|V(ud)| in the standard model, and places a stringent constraint on the scale of new physics that would lead to deviations from unitarity in the first row of the CKM matrix. We compute f(K+)/f(π+) numerically in unquenched lattice QCD using gauge-field ensembles recently generated that include four flavors of dynamical quarks: up, down, strange, and charm. We analyze data at four lattice spacings a ≈ 0.06, 0.09, 0.12, and 0.15 fm with simulated pion masses down to the physical value 135 MeV. We obtain f(K+)/f(π+) = 1.1947(26)(37), where the errors are statistical and total systematic, respectively. This is our first physics result from our N(f) = 2+1+1 ensembles, and the first calculation of f(K+)/f(π+) from lattice-QCD simulations at the physical point. Our result is the most precise lattice-QCD determination of f(K+)/f(π+), with an error comparable to the current world average. When combined with experimental measurements of the leptonic branching fractions, it leads to a precise determination of |V(us)|/|V(ud)| = 0.2309(9)(4) where the errors are theoretical and experimental, respectively.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of manymore » computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.« less
Distributing an executable job load file to compute nodes in a parallel computer
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gooding, Thomas M.
Distributing an executable job load file to compute nodes in a parallel computer, the parallel computer comprising a plurality of compute nodes, including: determining, by a compute node in the parallel computer, whether the compute node is participating in a job; determining, by the compute node in the parallel computer, whether a descendant compute node is participating in the job; responsive to determining that the compute node is participating in the job or that the descendant compute node is participating in the job, communicating, by the compute node to a parent compute node, an identification of a data communications linkmore » over which the compute node receives data from the parent compute node; constructing a class route for the job, wherein the class route identifies all compute nodes participating in the job; and broadcasting the executable load file for the job along the class route for the job.« less
USSR Report, Cybernetics, Computers and Automation Technology
1985-08-27
ions penetrate the semiconductor’s lattice and change its electrical properties at the penetration sites. Electron/ion facilities consisting of...Design Systems], AVTOMETRIYA, 1984, No 4. 56 k. -^ qCD (l[^]^[Bm]-*{MCp^ ■srn ,(4) :2) (2(&L_ -Serf Key: 1. Application program 3. Link module
Diphoton production at the LHC: a QCD study up to NNLO
NASA Astrophysics Data System (ADS)
Catani, Stefano; Cieri, Leandro; de Florian, Daniel; Ferrera, Giancarlo; Grazzini, Massimiliano
2018-04-01
We consider the production of prompt-photon pairs at the LHC and we report on a study of QCD radiative corrections up to the next-to-next-to-leading order (NNLO). We present a detailed comparison of next-to-leading order (NLO) results obtained within the standard and smooth cone isolation criteria, by studying the dependence on the isolation parameters. We highlight the role of different partonic subprocesses within the two isolation criteria, and we show that they produce large radiative corrections for both criteria. Smooth cone isolation is a consistent procedure to compute QCD radiative corrections at NLO and beyond. If photon isolation is sufficiently tight, we show that the NLO results for the two isolation procedures are consistent with each other within their perturbative uncertainties. We then extend our study to NNLO by using smooth cone isolation. We discuss the impact of the NNLO corrections and the corresponding perturbative uncertainties for both fiducial cross sections and distributions, and we comment on the comparison with some LHC data. Throughout our study we remark on the main features that are produced by the kinematical selection cuts that are applied to the photons. In particular, we examine soft-gluon singularities that appear in the perturbative computations of the invariant mass distribution of the photon pair, the transverse-momentum spectra of the photons, and the fiducial cross section with asymmetric and symmetric photon transverse-momentum cuts, and we present their behaviour in analytic form.
Topological susceptibility in finite temperature (2 +1 )-flavor QCD using gradient flow
NASA Astrophysics Data System (ADS)
Taniguchi, Yusuke; Kanaya, Kazuyuki; Suzuki, Hiroshi; Umeda, Takashi; WHOT-QCD Collaboration
2017-03-01
We compute the topological charge and its susceptibility in finite temperature (2 +1 )-flavor QCD on the lattice applying a gradient flow method. With the Iwasaki gauge action and nonperturbatively O (a ) -improved Wilson quarks, we perform simulations on a fine lattice with a ≃0.07 fm at a heavy u , d quark mass with mπ/mρ≃0.63 , but approximately physical s quark mass with mηss/mϕ≃0.74 . In a temperature range from T ≃174 MeV (Nt=16 ) to 697 MeV (Nt=4 ), we study two topics on the topological susceptibility. One is a comparison of gluonic and fermionic definitions of the topological susceptibility. Because the two definitions are related by chiral Ward-Takahashi identities, their equivalence is not trivial for lattice quarks which violate the chiral symmetry explicitly at finite lattice spacings. The gradient flow method enables us to compute them without being bothered by the chiral violation. We find a good agreement between the two definitions with Wilson quarks. The other is a comparison with a prediction of the dilute instanton gas approximation, which is relevant in a study of axions as a candidate of the dark matter in the evolution of the Universe. We find that the topological susceptibility shows a decrease in T which is consistent with the predicted χt(T )∝(T /Tpc)-8 for three-flavor QCD even at low temperature Tpc
Baryon interactions from lattice QCD with physical masses — strangeness S = -1 sector —
NASA Astrophysics Data System (ADS)
Nemura, Hidekatsu; Aoki, Sinya; Doi, Takumi; Gongyo, Shinya; Hatsuda, Tetsuo; Ikeda, Yoichi; Inoue, Takashi; Iritani, Takumi; Ishii, Noriyoshi; Miyamoto, Takaya; Sasaki, Kenji
2018-03-01
We present our recent results of baryon interactions with strangeness S = -1 based on Nambu-Bethe-Salpeter (NBS) correlation functions calculated fromlattice QCD with almost physical quark masses corresponding to (mk,mk) ≈ (146, 525) MeV and large volume (La)4 ≈ (96a)4 ≈ (8.1 fm)4. In order to perform a comprehensive study of baryon interactions, a large number of NBS correlation functions from NN to ΞΞ are calculated simultaneously by using large scale computer resources. In this contribution, we focus on the strangeness S = -1 channels of the hyperon interactions by means of HAL QCD method. Four sets of three potentials (the 3S1 - 3 D1 central, 3S1 - 3 D1 tensor, and the 1S0 central potentials) are presented for the ∑N - ∑N (the isospin I = 3/2) diagonal, the ∧N - ∧N diagonal, the ∧N → ∑N transition, and the ∑N - ∑N (I = 1/2) diagonal interactions. Scattering phase shifts for ∑N (I = 3/2) system are presented.
Stochastic reconstructions of spectral functions: Application to lattice QCD
NASA Astrophysics Data System (ADS)
Ding, H.-T.; Kaczmarek, O.; Mukherjee, Swagato; Ohno, H.; Shu, H.-T.
2018-05-01
We present a detailed study of the applications of two stochastic approaches, stochastic optimization method (SOM) and stochastic analytical inference (SAI), to extract spectral functions from Euclidean correlation functions. SOM has the advantage that it does not require prior information. On the other hand, SAI is a more generalized method based on Bayesian inference. Under mean field approximation SAI reduces to the often-used maximum entropy method (MEM) and for a specific choice of the prior SAI becomes equivalent to SOM. To test the applicability of these two stochastic methods to lattice QCD, firstly, we apply these methods to various reasonably chosen model correlation functions and present detailed comparisons of the reconstructed spectral functions obtained from SOM, SAI and MEM. Next, we present similar studies for charmonia correlation functions obtained from lattice QCD computations using clover-improved Wilson fermions on large, fine, isotropic lattices at 0.75 and 1.5 Tc, Tc being the deconfinement transition temperature of a pure gluon plasma. We find that SAI and SOM give consistent results to MEM at these two temperatures.
Jet production in the CoLoRFulNNLO method: Event shapes in electron-positron collisions
NASA Astrophysics Data System (ADS)
Del Duca, Vittorio; Duhr, Claude; Kardos, Adam; Somogyi, Gábor; Szőr, Zoltán; Trócsányi, Zoltán; Tulipánt, Zoltán
2016-10-01
We present the CoLoRFulNNLO method to compute higher order radiative corrections to jet cross sections in perturbative QCD. We apply our method to the computation of event shape observables in electron-positron collisions at NNLO accuracy and validate our code by comparing our predictions to previous results in the literature. We also calculate for the first time jet cone energy fraction at NNLO.
Matching pursuit parallel decomposition of seismic data
NASA Astrophysics Data System (ADS)
Li, Chuanhui; Zhang, Fanchang
2017-07-01
In order to improve the computation speed of matching pursuit decomposition of seismic data, a matching pursuit parallel algorithm is designed in this paper. We pick a fixed number of envelope peaks from the current signal in every iteration according to the number of compute nodes and assign them to the compute nodes on average to search the optimal Morlet wavelets in parallel. With the help of parallel computer systems and Message Passing Interface, the parallel algorithm gives full play to the advantages of parallel computing to significantly improve the computation speed of the matching pursuit decomposition and also has good expandability. Besides, searching only one optimal Morlet wavelet by every compute node in every iteration is the most efficient implementation.
Computer hardware fault administration
Archer, Charles J.; Megerian, Mark G.; Ratterman, Joseph D.; Smith, Brian E.
2010-09-14
Computer hardware fault administration carried out in a parallel computer, where the parallel computer includes a plurality of compute nodes. The compute nodes are coupled for data communications by at least two independent data communications networks, where each data communications network includes data communications links connected to the compute nodes. Typical embodiments carry out hardware fault administration by identifying a location of a defective link in the first data communications network of the parallel computer and routing communications data around the defective link through the second data communications network of the parallel computer.
Data communications in a parallel active messaging interface of a parallel computer
Archer, Charles J; Blocksome, Michael A; Ratterman, Joseph D; Smith, Brian E
2014-02-11
Data communications in a parallel active messaging interface ('PAMI') or a parallel computer, the parallel computer including a plurality of compute nodes that execute a parallel application, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution of a compute node, including specification of a client, a context, and a task, the compute nodes and the endpoints coupled for data communications instruction, the instruction characterized by instruction type, the instruction specifying a transmission of transfer data from the origin endpoint to a target endpoint and transmitting, in accordance witht the instruction type, the transfer data from the origin endpoin to the target endpoint.
Phenomenology of single-inclusive jet production with jet radius and threshold resummation
NASA Astrophysics Data System (ADS)
Liu, Xiaohui; Moch, Sven-Olaf; Ringer, Felix
2018-03-01
We perform a detailed study of inclusive jet production cross sections at the LHC and compare the QCD theory predictions based on the recently developed formalism for threshold and jet radius joint resummation at next-to-leading logarithmic accuracy to inclusive jet data collected by the CMS Collaboration at √{S }=7 and 13 TeV. We compute the cross sections at next-to-leading order in QCD with and without the joint resummation for different choices of jet radii R and observe that the joint resummation leads to crucial improvements in the description of the data. Comprehensive studies with different parton distribution functions demonstrate the necessity of considering the joint resummation in fits of those functions based on the LHC jet data.
The QCD form factor of heavy quarks at NNLO
NASA Astrophysics Data System (ADS)
Gluza, J.; Mitov, A.; Moch, S.; Riemann, T.
2009-07-01
We present an analytical calculation of the two-loop QCD corrections to the electromagnetic form factor of heavy quarks. The two-loop contributions to the form factor are reduced to linear combinations of master integrals, which are computed through higher orders in the parameter of dimensional regularization epsilon = (4-D)/2. Our result includes all terms of order epsilon at two loops and extends the previous literature. We apply the exponentiation of the heavy-quark form factor to derive new improved three-loop expansions in the high-energy limit. We also discuss the implications for predictions of massive n-parton amplitudes based on massless results in the limit, where the quark mass is small compared to all kinematical invariants.
Non-perturbative determination of cV, ZV and ZS/ZP in Nf = 3 lattice QCD
NASA Astrophysics Data System (ADS)
Heitger, Jochen; Joswig, Fabian; Vladikas, Anastassios; Wittemeier, Christian
2018-03-01
We report on non-perturbative computations of the improvement coefficient cV and the renormalization factor ZV of the vector current in three-flavour O(a) improved lattice QCD with Wilson quarks and tree-level Symanzik improved gauge action. To reduce finite quark mass effects, our improvement and normalization conditions exploit massive chiral Ward identities formulated in the Schrödinger functional setup, which also allow deriving a new method to extract the ratio ZS/ZP of scalar to pseudoscalar renormalization constants. We present preliminary results of a numerical evaluation of ZV and cV along a line of constant physics with gauge couplings corresponding to lattice spacings of about 0:09 fm and below, relevant for phenomenological applications.
QCD Coupling from a Nonperturbative Determination of the Three-Flavor Λ Parameter.
Bruno, Mattia; Brida, Mattia Dalla; Fritzsch, Patrick; Korzec, Tomasz; Ramos, Alberto; Schaefer, Stefan; Simma, Hubert; Sint, Stefan; Sommer, Rainer
2017-09-08
We present a lattice determination of the Λ parameter in three-flavor QCD and the strong coupling at the Z pole mass. Computing the nonperturbative running of the coupling in the range from 0.2 to 70 GeV, and using experimental input values for the masses and decay constants of the pion and the kaon, we obtain Λ_{MS[over ¯]}^{(3)}=341(12) MeV. The nonperturbative running up to very high energies guarantees that systematic effects associated with perturbation theory are well under control. Using the four-loop prediction for Λ_{MS[over ¯]}^{(5)}/Λ_{MS[over ¯]}^{(3)} yields α_{MS[over ¯]}^{(5)}(m_{Z})=0.11852(84).
Two-loop integrals for CP-even heavy quarkonium production and decays: elliptic sectors
NASA Astrophysics Data System (ADS)
Chen, Long-Bin; Jiang, Jun; Qiao, Cong-Feng
2018-04-01
By employing the differential equations, we compute analytically the elliptic sectors of two-loop master integrals appearing in the NNLO QCD corrections to CP-even heavy quarkonium exclusive production and decays, which turns out to be the last and toughest part in the relevant calculation. The integrals are found can be expressed as Goncharov polylogarithms and iterative integrals over elliptic functions. The master integrals may be applied to some other NNLO QCD calculations about heavy quarkonium exclusive production, like {γ}^{\\ast}γ \\to Q\\overline{Q} , {e}+{e}-\\to γ +Q\\overline{Q} , and H/{Z}^0\\to γ +Q\\overline{Q} , heavy quarkonium exclusive decays, and also the CP-even heavy quarkonium inclusive production and decays.
High-performance computing — an overview
NASA Astrophysics Data System (ADS)
Marksteiner, Peter
1996-08-01
An overview of high-performance computing (HPC) is given. Different types of computer architectures used in HPC are discussed: vector supercomputers, high-performance RISC processors, various parallel computers like symmetric multiprocessors, workstation clusters, massively parallel processors. Software tools and programming techniques used in HPC are reviewed: vectorizing compilers, optimization and vector tuning, optimization for RISC processors; parallel programming techniques like shared-memory parallelism, message passing and data parallelism; and numerical libraries.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dulat, Falko; Lionetti, Simone; Mistlberger, Bernhard
We present an analytic computation of the Higgs production cross section in the gluon fusion channel, which is differential in the components of the Higgs momentum and inclusive in the associated partonic radiation through NNLO in perturbative QCD. Our computation includes the necessary higher order terms in the dimensional regulator beyond the finite part that are required for renormalisation and collinear factorisation at N 3LO. We outline in detail the computational methods which we employ. We present numerical predictions for realistic final state observables, specifically distributions for the decay products of the Higgs boson in the γγ decay channel.
The 2nd Symposium on the Frontiers of Massively Parallel Computations
NASA Technical Reports Server (NTRS)
Mills, Ronnie (Editor)
1988-01-01
Programming languages, computer graphics, neural networks, massively parallel computers, SIMD architecture, algorithms, digital terrain models, sort computation, simulation of charged particle transport on the massively parallel processor and image processing are among the topics discussed.
Lattice QCD in rotating frames.
Yamamoto, Arata; Hirono, Yuji
2013-08-23
We formulate lattice QCD in rotating frames to study the physics of QCD matter under rotation. We construct the lattice QCD action with the rotational metric and apply it to the Monte Carlo simulation. As the first application, we calculate the angular momenta of gluons and quarks in the rotating QCD vacuum. This new framework is useful to analyze various rotation-related phenomena in QCD.
Data communications in a parallel active messaging interface of a parallel computer
Archer, Charles J; Blocksome, Michael A; Ratterman, Joseph D; Smith, Brian E
2013-11-12
Data communications in a parallel active messaging interface (`PAMI`) of a parallel computer composed of compute nodes that execute a parallel application, each compute node including application processors that execute the parallel application and at least one management processor dedicated to gathering information regarding data communications. The PAMI is composed of data communications endpoints, each endpoint composed of a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes and the endpoints coupled for data communications through the PAMI and through data communications resources. Embodiments function by gathering call site statistics describing data communications resulting from execution of data communications instructions and identifying in dependence upon the call cite statistics a data communications algorithm for use in executing a data communications instruction at a call site in the parallel application.
Parallel Computation of the Jacobian Matrix for Nonlinear Equation Solvers Using MATLAB
NASA Technical Reports Server (NTRS)
Rose, Geoffrey K.; Nguyen, Duc T.; Newman, Brett A.
2017-01-01
Demonstrating speedup for parallel code on a multicore shared memory PC can be challenging in MATLAB due to underlying parallel operations that are often opaque to the user. This can limit potential for improvement of serial code even for the so-called embarrassingly parallel applications. One such application is the computation of the Jacobian matrix inherent to most nonlinear equation solvers. Computation of this matrix represents the primary bottleneck in nonlinear solver speed such that commercial finite element (FE) and multi-body-dynamic (MBD) codes attempt to minimize computations. A timing study using MATLAB's Parallel Computing Toolbox was performed for numerical computation of the Jacobian. Several approaches for implementing parallel code were investigated while only the single program multiple data (spmd) method using composite objects provided positive results. Parallel code speedup is demonstrated but the goal of linear speedup through the addition of processors was not achieved due to PC architecture.
Performance Evaluation in Network-Based Parallel Computing
NASA Technical Reports Server (NTRS)
Dezhgosha, Kamyar
1996-01-01
Network-based parallel computing is emerging as a cost-effective alternative for solving many problems which require use of supercomputers or massively parallel computers. The primary objective of this project has been to conduct experimental research on performance evaluation for clustered parallel computing. First, a testbed was established by augmenting our existing SUNSPARCs' network with PVM (Parallel Virtual Machine) which is a software system for linking clusters of machines. Second, a set of three basic applications were selected. The applications consist of a parallel search, a parallel sort, a parallel matrix multiplication. These application programs were implemented in C programming language under PVM. Third, we conducted performance evaluation under various configurations and problem sizes. Alternative parallel computing models and workload allocations for application programs were explored. The performance metric was limited to elapsed time or response time which in the context of parallel computing can be expressed in terms of speedup. The results reveal that the overhead of communication latency between processes in many cases is the restricting factor to performance. That is, coarse-grain parallelism which requires less frequent communication between processes will result in higher performance in network-based computing. Finally, we are in the final stages of installing an Asynchronous Transfer Mode (ATM) switch and four ATM interfaces (each 155 Mbps) which will allow us to extend our study to newer applications, performance metrics, and configurations.
Lattice QCD inputs to the CKM unitarity triangle analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Laiho, Jack; Department of Physics and Astronomy, University of Glasgow, Glasgow, G128 QQ; Lunghi, E.
2010-02-01
We perform a global fit to the Cabibbo-Kobayashi-Maskawa unitarity triangle using the latest experimental and theoretical constraints. Our emphasis is on the hadronic weak matrix elements that enter the analysis, which must be computed using lattice QCD or other nonperturbative methods. Realistic lattice QCD calculations which include the effects of the dynamical up, down, and strange quarks are now available for all of the standard inputs to the global fit. We therefore present lattice averages for all of the necessary hadronic weak matrix elements. We attempt to account for correlations between lattice QCD results in a reasonable but conservative manner:more » whenever there are reasons to believe that an error is correlated between two lattice calculations, we take the degree of correlation to be 100%. These averages are suitable for use as inputs both in the global Cabibbo-Kobayashi-Maskawa unitarity triangle fit and other phenomenological analyses. In order to illustrate the impact of the lattice averages, we make standard model predictions for the parameters B-circumflex{sub K}, |V{sub cb}|, and |V{sub ub}|/|V{sub cb}|. We find a (2-3){sigma} tension in the unitarity triangle, depending upon whether we use the inclusive or exclusive determination of |V{sub cb}|. If we interpret the tension as a sign of new physics in either neutral kaon or B mixing, we find that the scenario with new physics in kaon mixing is preferred by present data.« less
Lattice QCD Inputs to the CKM Unitarity Triangle Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Van de Water, R.; Lunghi, E; Laiho, J
2010-02-02
We perform a global fit to the Cabibbo-Kobayashi-Maskawa unitarity triangle using the latest experimental and theoretical constraints. Our emphasis is on the hadronic weak matrix elements that enter the analysis, which must be computed using lattice QCD or other nonperturbative methods. Realistic lattice QCD calculations which include the effects of the dynamical up, down, and strange quarks are now available for all of the standard inputs to the global fit. We therefore present lattice averages for all of the necessary hadronic weak matrix elements. We attempt to account for correlations between lattice QCD results in a reasonable but conservative manner:more » whenever there are reasons to believe that an error is correlated between two lattice calculations, we take the degree of correlation to be 100%. These averages are suitable for use as inputs both in the global Cabibbo-Kobayashi-Maskawa unitarity triangle fit and other phenomenological analyses. In order to illustrate the impact of the lattice averages, we make standard model predictions for the parameters B{sub K}, |V{sub cb}|, and |V{sub ub}|/|Vcb|. We find a (2-3){sigma} tension in the unitarity triangle, depending upon whether we use the inclusive or exclusive determination of |V{sub cb}|. If we interpret the tension as a sign of new physics in either neutral kaon or B mixing, we find that the scenario with new physics in kaon mixing is preferred by present data.« less
Parallel Computing Using Web Servers and "Servlets".
ERIC Educational Resources Information Center
Lo, Alfred; Bloor, Chris; Choi, Y. K.
2000-01-01
Describes parallel computing and presents inexpensive ways to implement a virtual parallel computer with multiple Web servers. Highlights include performance measurement of parallel systems; models for using Java and intranet technology including single server, multiple clients and multiple servers, single client; and a comparison of CGI (common…
A class of parallel algorithms for computation of the manipulator inertia matrix
NASA Technical Reports Server (NTRS)
Fijany, Amir; Bejczy, Antal K.
1989-01-01
Parallel and parallel/pipeline algorithms for computation of the manipulator inertia matrix are presented. An algorithm based on composite rigid-body spatial inertia method, which provides better features for parallelization, is used for the computation of the inertia matrix. Two parallel algorithms are developed which achieve the time lower bound in computation. Also described is the mapping of these algorithms with topological variation on a two-dimensional processor array, with nearest-neighbor connection, and with cardinality variation on a linear processor array. An efficient parallel/pipeline algorithm for the linear array was also developed, but at significantly higher efficiency.
Parallel computing of a climate model on the dawn 1000 by domain decomposition method
NASA Astrophysics Data System (ADS)
Bi, Xunqiang
1997-12-01
In this paper the parallel computing of a grid-point nine-level atmospheric general circulation model on the Dawn 1000 is introduced. The model was developed by the Institute of Atmospheric Physics (IAP), Chinese Academy of Sciences (CAS). The Dawn 1000 is a MIMD massive parallel computer made by National Research Center for Intelligent Computer (NCIC), CAS. A two-dimensional domain decomposition method is adopted to perform the parallel computing. The potential ways to increase the speed-up ratio and exploit more resources of future massively parallel supercomputation are also discussed.
Parallel Computing Strategies for Irregular Algorithms
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Oliker, Leonid; Shan, Hongzhang; Biegel, Bryan (Technical Monitor)
2002-01-01
Parallel computing promises several orders of magnitude increase in our ability to solve realistic computationally-intensive problems, but relies on their efficient mapping and execution on large-scale multiprocessor architectures. Unfortunately, many important applications are irregular and dynamic in nature, making their effective parallel implementation a daunting task. Moreover, with the proliferation of parallel architectures and programming paradigms, the typical scientist is faced with a plethora of questions that must be answered in order to obtain an acceptable parallel implementation of the solution algorithm. In this paper, we consider three representative irregular applications: unstructured remeshing, sparse matrix computations, and N-body problems, and parallelize them using various popular programming paradigms on a wide spectrum of computer platforms ranging from state-of-the-art supercomputers to PC clusters. We present the underlying problems, the solution algorithms, and the parallel implementation strategies. Smart load-balancing, partitioning, and ordering techniques are used to enhance parallel performance. Overall results demonstrate the complexity of efficiently parallelizing irregular algorithms.
Heavy quark form factors at two loops
NASA Astrophysics Data System (ADS)
Ablinger, J.; Behring, A.; Blümlein, J.; Falcioni, G.; De Freitas, A.; Marquard, P.; Rana, N.; Schneider, C.
2018-05-01
We compute the two-loop QCD corrections to the heavy quark form factors in the case of the vector, axial-vector, scalar and pseudoscalar currents up to second order in the dimensional parameter ɛ =(4 -D )/2 . These terms are required in the renormalization of the higher-order corrections to these form factors.
Phase diagram and thermal properties of strong-interaction matter
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gao, Fei; Chen, Jing; Liu, Yu-Xin
2016-05-20
We introduce a novel method for computing the (μ, T)-dependent pressure in continuum QCD, from which we obtain a complex phase diagram and predictions for thermal properties of the dressed-quark component of the system, providing the in-medium behavior of the related trace anomaly, speed of sound, latent heat, and heat capacity.
Lattice QCD and the unitarity triangle
DOE Office of Scientific and Technical Information (OSTI.GOV)
Andreas S Kronfeld
2001-12-03
Theoretical and computational advances in lattice calculations are reviewed, with focus on examples relevant to the unitarity triangle of the CKM matrix. Recent progress in semi-leptonic form factors for B {yields} {pi}/v and B {yields} D*lv, as well as the parameter {zeta} in B{sup 0}-{bar B}{sup 0} mixing, are highlighted.
Interpretation of the new Ω _c0 states via their mass and width
NASA Astrophysics Data System (ADS)
Agaev, S. S.; Azizi, K.; Sundu, H.
2017-06-01
The masses and pole residues of the ground and first radially excited Ω _c0 states with spin-parities JP=1/2+, 3/2+, as well as P-wave Ω _c0 with JP=1/2-, 3/2- are calculated by means of the two-point QCD sum rules. The strong decays of Ω _c0 baryons are also studied and the widths of these decay channels are computed. The relevant computations are performed in the context of the full QCD sum rules on the light cone. The results obtained for the masses and widths are confronted with recent experimental data of the LHCb Collaboration, which allow us to interpret Ω _c(3000)0, Ω _c(3050)0, and Ω _c(3119)0 as the excited css baryons with the quantum numbers (1P, 1/2-), (1P, 3/2-), and (2S, 3/2+), respectively. The (2S, 1/2+) state can be assigned either to the Ω _c(3066)0 state or the Ω _c(3090)0 excited baryon.
NASA Astrophysics Data System (ADS)
von Hippel, Georg; Rae, Thomas D.; Shintani, Eigo; Wittig, Hartmut
2017-01-01
We study the performance of all-mode-averaging (AMA) when used in conjunction with a locally deflated SAP-preconditioned solver, determining how to optimize the local block sizes and number of deflation fields in order to minimize the computational cost for a given level of overall statistical accuracy. We find that AMA enables a reduction of the statistical error on nucleon charges by a factor of around two at the same cost when compared to the standard method. As a demonstration, we compute the axial, scalar and tensor charges of the nucleon in Nf = 2 lattice QCD with non-perturbatively O(a)-improved Wilson quarks, using O(10,000) measurements to pursue the signal out to source-sink separations of ts ∼ 1.5 fm. Our results suggest that the axial charge is suffering from a significant amount (5-10%) of excited-state contamination at source-sink separations of up to ts ∼ 1.2 fm, whereas the excited-state contamination in the scalar and tensor charges seems to be small.
N3LO corrections to jet production in deep inelastic scattering using the Projection-to-Born method
NASA Astrophysics Data System (ADS)
Currie, J.; Gehrmann, T.; Glover, E. W. N.; Huss, A.; Niehues, J.; Vogt, A.
2018-05-01
Computations of higher-order QCD corrections for processes with exclusive final states require a subtraction method for real-radiation contributions. We present the first-ever generalisation of a subtraction method for third-order (N3LO) QCD corrections. The Projection-to-Born method is used to combine inclusive N3LO coefficient functions with an exclusive second-order (NNLO) calculation for a final state with an extra jet. The input requirements, advantages, and potential applications of the method are discussed, and validations at lower orders are performed. As a test case, we compute the N3LO corrections to kinematical distributions and production rates for single-jet production in deep inelastic scattering in the laboratory frame, and compare them with data from the ZEUS experiment at HERA. The corrections are small in the central rapidity region, where they stabilize the predictions to sub per-cent level. The corrections increase substantially towards forward rapidity where large logarithmic effects are expected, thereby yielding an improved description of the data in this region.
Parallel solution of sparse one-dimensional dynamic programming problems
NASA Technical Reports Server (NTRS)
Nicol, David M.
1989-01-01
Parallel computation offers the potential for quickly solving large computational problems. However, it is often a non-trivial task to effectively use parallel computers. Solution methods must sometimes be reformulated to exploit parallelism; the reformulations are often more complex than their slower serial counterparts. We illustrate these points by studying the parallelization of sparse one-dimensional dynamic programming problems, those which do not obviously admit substantial parallelization. We propose a new method for parallelizing such problems, develop analytic models which help us to identify problems which parallelize well, and compare the performance of our algorithm with existing algorithms on a multiprocessor.
Jackin, Boaz Jessie; Watanabe, Shinpei; Ootsu, Kanemitsu; Ohkawa, Takeshi; Yokota, Takashi; Hayasaki, Yoshio; Yatagai, Toyohiko; Baba, Takanobu
2018-04-20
A parallel computation method for large-size Fresnel computer-generated hologram (CGH) is reported. The method was introduced by us in an earlier report as a technique for calculating Fourier CGH from 2D object data. In this paper we extend the method to compute Fresnel CGH from 3D object data. The scale of the computation problem is also expanded to 2 gigapixels, making it closer to real application requirements. The significant feature of the reported method is its ability to avoid communication overhead and thereby fully utilize the computing power of parallel devices. The method exhibits three layers of parallelism that favor small to large scale parallel computing machines. Simulation and optical experiments were conducted to demonstrate the workability and to evaluate the efficiency of the proposed technique. A two-times improvement in computation speed has been achieved compared to the conventional method, on a 16-node cluster (one GPU per node) utilizing only one layer of parallelism. A 20-times improvement in computation speed has been estimated utilizing two layers of parallelism on a very large-scale parallel machine with 16 nodes, where each node has 16 GPUs.
On the two-loop virtual QCD corrections to Higgs boson pair production in the standard model
Degrassi, Giuseppe; Giardino, Pier Paolo; Gröber, Ramona
2016-07-21
Here, we compute the next-to-leading order virtual QCD corrections to Higgs-pair production via gluon fusion. We also present analytic results for the two-loop contributions to the spin-0 and spin-2 form factors in the amplitude. The reducible contributions, given by the double-triangle diagrams, are evaluated exactly while the two-loop irreducible diagrams are evaluated by an asymptotic expansion in heavy top-quark mass up to and including terms of O(1/mmore » $$8\\atop{t}$$). We estimate that mass effects can reduce the hadronic cross section by at most 10 %, assuming that the finite top-quark mass effects are of similar size in the entire range of partonic energies.« less
Bulk viscous corrections to screening and damping in QCD at high temperatures
DOE Office of Scientific and Technical Information (OSTI.GOV)
Du, Qianqian; Dumitru, Adrian; Guo, Yun
2017-01-01
Non-equilibrium corrections to the distribution functions of quarks and gluons in a hot and dense QCD medium modify the \\hard thermal loops" (HTL). The HTLs determine the retarded, advanced, and symmetric (time-ordered) propagators for gluons with soft momenta as well as the Debye screening and Landau damping mass scales. Here, we compute such corrections to a thermal as well as to a non-thermal fixed point. The screening and damping mass scales are sensitive to the bulk pressure and hence to (pseudo-) critical dynamical scaling of the bulk viscosity in the vicinity of a second-order critical point. This could be reectedmore » in the properties of quarkonium bound states in the deconfined phase and in the dynamics of soft gluon fields.« less
Short-distance matrix elements for $D$-meson mixing for 2+1 lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chang, Chia Cheng
2015-01-01
We study the short-distance hadronic matrix elements for D-meson mixing with partially quenched N f = 2+1 lattice QCD. We use a large set of the MIMD Lattice Computation Collaboration's gauge configurations with a 2 tadpole-improved staggered sea quarks and tadpole-improved Lüscher-Weisz gluons. We use the a 2 tadpole-improved action for valence light quarks and the Sheikoleslami-Wohlert action with the Fermilab interpretation for the valence charm quark. Our calculation covers the complete set of five operators needed to constrain new physics models for D-meson mixing. We match our matrix elements to the MS-NDR scheme evaluated at 3 GeV. We reportmore » values for the Beneke-Buchalla-Greub-Lenz-Nierste choice of evanescent operators.« less
Nucleon, $$\\Delta$$ and $$\\Omega$$ excited states in $$N_f=2+1$$ lattice QCD
John Bulava; Edwards, Robert G.; Engelson, Eric; ...
2010-07-22
The energies of the excited states of the Nucleon,more » $$\\Delta$$ and $$\\Omega$$ are computed in lattice QCD, using two light quarks and one strange quark on anisotropic lattices. The calculation is performed at three values of the light quark mass, corresponding to pion masses $$m_{\\pi}$$ = 392(4), 438(3) and 521(3) MeV. We employ the variational method with a large basis of interpolating operators enabling six energies in each irreducible representation of the lattice to be distinguished clearly. We compare our calculation with the low-lying experimental spectrum, with which we find reasonable agreement in the pattern of states. In addition, the need to include operators that couple to the expected multi-hadron states in the spectrum is clearly identified.« less
Resonant π + γ → π + π 0 amplitude from Quantum Chromodynamics
Briceño, Raúl A.; Dudek, Jozef J.; Edwards, Robert G.; ...
2015-12-08
We present the first ab initio calculation of a radiative transition of a hadronic resonance within Quantum Chromodynamics (QCD). We compute the amplitude formore » $$\\pi\\pi \\to \\pi\\gamma^\\star$$, as a function of the energy of the $$\\pi\\pi$$ pair and the virtuality of the photon, in the kinematic regime where $$\\pi\\pi$$ couples strongly to the unstable $$\\rho$$ resonance. This exploratory calculation is performed using a lattice discretization of QCD with quark masses corresponding to $$m_\\pi \\approx 400$$ MeV. As a result, we obtain a description of the energy dependence of the transition amplitude, constrained at 48 kinematic points, that we can analytically continue to the $$\\rho$$ pole and identify from its residue the $$\\rho \\to \\pi\\gamma^\\star$$ form-factor.« less
Two loop QCD vertices at the symmetric point
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gracey, J. A.
2011-10-15
We compute the triple gluon, quark-gluon and ghost-gluon vertices of QCD at the symmetric subtraction point at two loops in the MS scheme. In addition we renormalize each of the three vertices in their respective momentum subtraction schemes, MOMggg, MOMq and MOMh. The conversion functions of all the wave functions, coupling constant and gauge parameter renormalization constants of each of the schemes relative to MS are determined analytically. These are then used to derive the three loop anomalous dimensions of the gluon, quark, Faddeev-Popov ghost and gauge parameter as well as the {beta} function in an arbitrary linear covariant gaugemore » for each MOM scheme. There is good agreement of the latter with earlier Landau gauge numerical estimates of Chetyrkin and Seidensticker.« less
MPI_XSTAR: MPI-based Parallelization of the XSTAR Photoionization Program
NASA Astrophysics Data System (ADS)
Danehkar, Ashkbiz; Nowak, Michael A.; Lee, Julia C.; Smith, Randall K.
2018-02-01
We describe a program for the parallel implementation of multiple runs of XSTAR, a photoionization code that is used to predict the physical properties of an ionized gas from its emission and/or absorption lines. The parallelization program, called MPI_XSTAR, has been developed and implemented in the C++ language by using the Message Passing Interface (MPI) protocol, a conventional standard of parallel computing. We have benchmarked parallel multiprocessing executions of XSTAR, using MPI_XSTAR, against a serial execution of XSTAR, in terms of the parallelization speedup and the computing resource efficiency. Our experience indicates that the parallel execution runs significantly faster than the serial execution, however, the efficiency in terms of the computing resource usage decreases with increasing the number of processors used in the parallel computing.
Data communications in a parallel active messaging interface of a parallel computer
Archer, Charles J; Blocksome, Michael A; Ratterman, Joseph D; Smith, Brian E
2013-10-29
Data communications in a parallel active messaging interface (`PAMI`) of a parallel computer, the parallel computer including a plurality of compute nodes that execute a parallel application, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes and the endpoints coupled for data communications through the PAMI and through data communications resources, including receiving in an origin endpoint of the PAMI a data communications instruction, the instruction characterized by an instruction type, the instruction specifying a transmission of transfer data from the origin endpoint to a target endpoint and transmitting, in accordance with the instruction type, the transfer data from the origin endpoint to the target endpoint.
A CFD Heterogeneous Parallel Solver Based on Collaborating CPU and GPU
NASA Astrophysics Data System (ADS)
Lai, Jianqi; Tian, Zhengyu; Li, Hua; Pan, Sha
2018-03-01
Since Graphic Processing Unit (GPU) has a strong ability of floating-point computation and memory bandwidth for data parallelism, it has been widely used in the areas of common computing such as molecular dynamics (MD), computational fluid dynamics (CFD) and so on. The emergence of compute unified device architecture (CUDA), which reduces the complexity of compiling program, brings the great opportunities to CFD. There are three different modes for parallel solution of NS equations: parallel solver based on CPU, parallel solver based on GPU and heterogeneous parallel solver based on collaborating CPU and GPU. As we can see, GPUs are relatively rich in compute capacity but poor in memory capacity and the CPUs do the opposite. We need to make full use of the GPUs and CPUs, so a CFD heterogeneous parallel solver based on collaborating CPU and GPU has been established. Three cases are presented to analyse the solver’s computational accuracy and heterogeneous parallel efficiency. The numerical results agree well with experiment results, which demonstrate that the heterogeneous parallel solver has high computational precision. The speedup on a single GPU is more than 40 for laminar flow, it decreases for turbulent flow, but it still can reach more than 20. What’s more, the speedup increases as the grid size becomes larger.
Higgs-differential cross section at NNLO in dimensional regularisation
Dulat, Falko; Lionetti, Simone; Mistlberger, Bernhard; ...
2017-07-05
We present an analytic computation of the Higgs production cross section in the gluon fusion channel, which is differential in the components of the Higgs momentum and inclusive in the associated partonic radiation through NNLO in perturbative QCD. Our computation includes the necessary higher order terms in the dimensional regulator beyond the finite part that are required for renormalisation and collinear factorisation at N 3LO. We outline in detail the computational methods which we employ. We present numerical predictions for realistic final state observables, specifically distributions for the decay products of the Higgs boson in the γγ decay channel.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gooding, Thomas M.
Distributing an executable job load file to compute nodes in a parallel computer, the parallel computer comprising a plurality of compute nodes, including: determining, by a compute node in the parallel computer, whether the compute node is participating in a job; determining, by the compute node in the parallel computer, whether a descendant compute node is participating in the job; responsive to determining that the compute node is participating in the job or that the descendant compute node is participating in the job, communicating, by the compute node to a parent compute node, an identification of a data communications linkmore » over which the compute node receives data from the parent compute node; constructing a class route for the job, wherein the class route identifies all compute nodes participating in the job; and broadcasting the executable load file for the job along the class route for the job.« less
NASA Technical Reports Server (NTRS)
Dongarra, Jack (Editor); Messina, Paul (Editor); Sorensen, Danny C. (Editor); Voigt, Robert G. (Editor)
1990-01-01
Attention is given to such topics as an evaluation of block algorithm variants in LAPACK and presents a large-grain parallel sparse system solver, a multiprocessor method for the solution of the generalized Eigenvalue problem on an interval, and a parallel QR algorithm for iterative subspace methods on the CM2. A discussion of numerical methods includes the topics of asynchronous numerical solutions of PDEs on parallel computers, parallel homotopy curve tracking on a hypercube, and solving Navier-Stokes equations on the Cedar Multi-Cluster system. A section on differential equations includes a discussion of a six-color procedure for the parallel solution of elliptic systems using the finite quadtree structure, data parallel algorithms for the finite element method, and domain decomposition methods in aerodynamics. Topics dealing with massively parallel computing include hypercube vs. 2-dimensional meshes and massively parallel computation of conservation laws. Performance and tools are also discussed.
Zanderighi, Giulia
2018-04-26
Modern QCD - Lecture 1 Starting from the QCD Lagrangian we will revisit some basic QCD concepts and derive fundamental properties like gauge invariance and isospin symmetry and will discuss the Feynman rules of the theory. We will then focus on the gauge group of QCD and derive the Casimirs CF and CA and some useful color identities.
Lee, Jae H.; Yao, Yushu; Shrestha, Uttam; Gullberg, Grant T.; Seo, Youngho
2014-01-01
The primary goal of this project is to implement the iterative statistical image reconstruction algorithm, in this case maximum likelihood expectation maximum (MLEM) used for dynamic cardiac single photon emission computed tomography, on Spark/GraphX. This involves porting the algorithm to run on large-scale parallel computing systems. Spark is an easy-to- program software platform that can handle large amounts of data in parallel. GraphX is a graph analytic system running on top of Spark to handle graph and sparse linear algebra operations in parallel. The main advantage of implementing MLEM algorithm in Spark/GraphX is that it allows users to parallelize such computation without any expertise in parallel computing or prior knowledge in computer science. In this paper we demonstrate a successful implementation of MLEM in Spark/GraphX and present the performance gains with the goal to eventually make it useable in clinical setting. PMID:27081299
Lee, Jae H; Yao, Yushu; Shrestha, Uttam; Gullberg, Grant T; Seo, Youngho
2014-11-01
The primary goal of this project is to implement the iterative statistical image reconstruction algorithm, in this case maximum likelihood expectation maximum (MLEM) used for dynamic cardiac single photon emission computed tomography, on Spark/GraphX. This involves porting the algorithm to run on large-scale parallel computing systems. Spark is an easy-to- program software platform that can handle large amounts of data in parallel. GraphX is a graph analytic system running on top of Spark to handle graph and sparse linear algebra operations in parallel. The main advantage of implementing MLEM algorithm in Spark/GraphX is that it allows users to parallelize such computation without any expertise in parallel computing or prior knowledge in computer science. In this paper we demonstrate a successful implementation of MLEM in Spark/GraphX and present the performance gains with the goal to eventually make it useable in clinical setting.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, Stanley J.; de Teramond, Guy F.; Deur, Alexandre P.
2015-09-01
The valence Fock-state wavefunctions of the light-front QCD Hamiltonian satisfy a relativistic equation of motion with an effective confining potential U which systematically incorporates the effects of higher quark and gluon Fock states. If one requires that the effective action which underlies the QCD Lagrangian remains conformally invariant and extends the formalism of de Alfaro, Fubini and Furlan to light front Hamiltonian theory, the potential U has a unique form of a harmonic oscillator potential, and a mass gap arises. The result is a nonperturbative relativistic light-front quantum mechanical wave equation which incorporates color confinement and other essential spectroscopic andmore » dynamical features of hadron physics, including a massless pion for zero quark mass and linear Regge trajectories with the same slope in the radial quantum number n and orbital angular momentum L. Only one mass parameter κ appears. Light-front holography thus provides a precise relation between the bound-state amplitudes in the fifth dimension of AdS space and the boost-invariant light-front wavefunctions describing the internal structure of hadrons in physical space-time. We also show how the mass scale κ underlying confinement and hadron masses determines the scale Λ {ovr MS} controlling the evolution of the perturbative QCD coupling. The relation between scales is obtained by matching the nonperturbative dynamics, as described by an effective conformal theory mapped to the light-front and its embedding in AdS space, to the perturbative QCD regime computed to four-loop order. The result is an effective coupling defined at all momenta. The predicted value Λ {ovr MS}=0.328±0.034 GeV is in agreement with the world average 0.339±0.010 GeV. The analysis applies to any renormalization scheme.« less
Wu, Xiao-Lin; Sun, Chuanyu; Beissinger, Timothy M; Rosa, Guilherme Jm; Weigel, Kent A; Gatti, Natalia de Leon; Gianola, Daniel
2012-09-25
Most Bayesian models for the analysis of complex traits are not analytically tractable and inferences are based on computationally intensive techniques. This is true of Bayesian models for genome-enabled selection, which uses whole-genome molecular data to predict the genetic merit of candidate animals for breeding purposes. In this regard, parallel computing can overcome the bottlenecks that can arise from series computing. Hence, a major goal of the present study is to bridge the gap to high-performance Bayesian computation in the context of animal breeding and genetics. Parallel Monte Carlo Markov chain algorithms and strategies are described in the context of animal breeding and genetics. Parallel Monte Carlo algorithms are introduced as a starting point including their applications to computing single-parameter and certain multiple-parameter models. Then, two basic approaches for parallel Markov chain Monte Carlo are described: one aims at parallelization within a single chain; the other is based on running multiple chains, yet some variants are discussed as well. Features and strategies of the parallel Markov chain Monte Carlo are illustrated using real data, including a large beef cattle dataset with 50K SNP genotypes. Parallel Markov chain Monte Carlo algorithms are useful for computing complex Bayesian models, which does not only lead to a dramatic speedup in computing but can also be used to optimize model parameters in complex Bayesian models. Hence, we anticipate that use of parallel Markov chain Monte Carlo will have a profound impact on revolutionizing the computational tools for genomic selection programs.
2012-01-01
Background Most Bayesian models for the analysis of complex traits are not analytically tractable and inferences are based on computationally intensive techniques. This is true of Bayesian models for genome-enabled selection, which uses whole-genome molecular data to predict the genetic merit of candidate animals for breeding purposes. In this regard, parallel computing can overcome the bottlenecks that can arise from series computing. Hence, a major goal of the present study is to bridge the gap to high-performance Bayesian computation in the context of animal breeding and genetics. Results Parallel Monte Carlo Markov chain algorithms and strategies are described in the context of animal breeding and genetics. Parallel Monte Carlo algorithms are introduced as a starting point including their applications to computing single-parameter and certain multiple-parameter models. Then, two basic approaches for parallel Markov chain Monte Carlo are described: one aims at parallelization within a single chain; the other is based on running multiple chains, yet some variants are discussed as well. Features and strategies of the parallel Markov chain Monte Carlo are illustrated using real data, including a large beef cattle dataset with 50K SNP genotypes. Conclusions Parallel Markov chain Monte Carlo algorithms are useful for computing complex Bayesian models, which does not only lead to a dramatic speedup in computing but can also be used to optimize model parameters in complex Bayesian models. Hence, we anticipate that use of parallel Markov chain Monte Carlo will have a profound impact on revolutionizing the computational tools for genomic selection programs. PMID:23009363
Parallel approach in RDF query processing
NASA Astrophysics Data System (ADS)
Vajgl, Marek; Parenica, Jan
2017-07-01
Parallel approach is nowadays a very cheap solution to increase computational power due to possibility of usage of multithreaded computational units. This hardware became typical part of nowadays personal computers or notebooks and is widely spread. This contribution deals with experiments how evaluation of computational complex algorithm of the inference over RDF data can be parallelized over graphical cards to decrease computational time.
A comparative study of serial and parallel aeroelastic computations of wings
NASA Technical Reports Server (NTRS)
Byun, Chansup; Guruswamy, Guru P.
1994-01-01
A procedure for computing the aeroelasticity of wings on parallel multiple-instruction, multiple-data (MIMD) computers is presented. In this procedure, fluids are modeled using Euler equations, and structures are modeled using modal or finite element equations. The procedure is designed in such a way that each discipline can be developed and maintained independently by using a domain decomposition approach. In the present parallel procedure, each computational domain is scalable. A parallel integration scheme is used to compute aeroelastic responses by solving fluid and structural equations concurrently. The computational efficiency issues of parallel integration of both fluid and structural equations are investigated in detail. This approach, which reduces the total computational time by a factor of almost 2, is demonstrated for a typical aeroelastic wing by using various numbers of processors on the Intel iPSC/860.
Research in parallel computing
NASA Technical Reports Server (NTRS)
Ortega, James M.; Henderson, Charles
1994-01-01
This report summarizes work on parallel computations for NASA Grant NAG-1-1529 for the period 1 Jan. - 30 June 1994. Short summaries on highly parallel preconditioners, target-specific parallel reductions, and simulation of delta-cache protocols are provided.
Light-Front Holography, Light-Front Wavefunctions, and Novel QCD Phenomena
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, Stanley J.; /SLAC /Southern Denmark U., CP3-Origins; de Teramond, Guy F.
2012-02-16
Light-Front Holography is one of the most remarkable features of the AdS/CFT correspondence. In spite of its present limitations it provides important physical insights into the nonperturbative regime of QCD and its transition to the perturbative domain. This novel framework allows hadronic amplitudes in a higher dimensional anti-de Sitter (AdS) space to be mapped to frame-independent light-front wavefunctions of hadrons in physical space-time. The model leads to an effective confining light-front QCD Hamiltonian and a single-variable light-front Schroedinger equation which determines the eigenspectrum and the light-front wavefunctions of hadrons for general spin and orbital angular momentum. The coordinate z inmore » AdS space is uniquely identified with a Lorentz-invariant coordinate {zeta} which measures the separation of the constituents within a hadron at equal light-front time and determines the off-shell dynamics of the bound-state wavefunctions, and thus the fall-off as a function of the invariant mass of the constituents. The soft-wall holographic model modified by a positive-sign dilaton metric, leads to a remarkable one-parameter description of nonperturbative hadron dynamics - a semi-classical frame-independent first approximation to the spectra and light-front wavefunctions of meson and baryons. The model predicts a Regge spectrum of linear trajectories with the same slope in the leading orbital angular momentum L of hadrons and the radial quantum number n. The hadron eigensolutions projected on the free Fock basis provides the complete set of valence and non-valence light-front Fock state wavefunctions {Psi}{sub n/H} (x{sub i}, k{sub {perpendicular}i}, {lambda}{sub i}) which describe the hadron's momentum and spin distributions needed to compute the direct measures of hadron structure at the quark and gluon level, such as elastic and transition form factors, distribution amplitudes, structure functions, generalized parton distributions and transverse momentum distributions. The effective confining potential also creates quark-antiquark pairs from the amplitude q {yields} q{bar q}q. Thus in holographic QCD higher Fock states can have any number of extra q{bar q} pairs. We discuss the relevance of higher Fock-states for describing the detailed structure of space and time-like form factors. The AdS/QCD model can be systematically improved by using its complete orthonormal solutions to diagonalize the full QCD light-front Hamiltonian or by applying the Lippmann-Schwinger method in order to systematically include the QCD interaction terms. A new perspective on quark and gluon condensates is also obtained.« less
Contemporary continuum QCD approaches to excited hadrons
NASA Astrophysics Data System (ADS)
El-Bennich, Bruno; Rojas, Eduardo
2016-03-01
Amongst the bound states produced by the strong interaction, radially excited meson and nucleon states offer an important phenomenological window into the long-range behavior of the coupling constant in Quantum Chromodynamics. We here report on some technical details related to the computation of the bound state's eigenvalue spectrum in the framework of Bethe-Salpeter and Faddeev equations.
Parallel computations and control of adaptive structures
NASA Technical Reports Server (NTRS)
Park, K. C.; Alvin, Kenneth F.; Belvin, W. Keith; Chong, K. P. (Editor); Liu, S. C. (Editor); Li, J. C. (Editor)
1991-01-01
The equations of motion for structures with adaptive elements for vibration control are presented for parallel computations to be used as a software package for real-time control of flexible space structures. A brief introduction of the state-of-the-art parallel computational capability is also presented. Time marching strategies are developed for an effective use of massive parallel mapping, partitioning, and the necessary arithmetic operations. An example is offered for the simulation of control-structure interaction on a parallel computer and the impact of the approach presented for applications in other disciplines than aerospace industry is assessed.
Design of a massively parallel computer using bit serial processing elements
NASA Technical Reports Server (NTRS)
Aburdene, Maurice F.; Khouri, Kamal S.; Piatt, Jason E.; Zheng, Jianqing
1995-01-01
A 1-bit serial processor designed for a parallel computer architecture is described. This processor is used to develop a massively parallel computational engine, with a single instruction-multiple data (SIMD) architecture. The computer is simulated and tested to verify its operation and to measure its performance for further development.
AdS/QCD and Applications of Light-Front Holography
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, Stanley J.; /SLAC /Southern Denmark U., CP3-Origins; Cao, Fu-Guang
2012-02-16
Light-Front Holography leads to a rigorous connection between hadronic amplitudes in a higher dimensional anti-de Sitter (AdS) space and frame-independent light-front wavefunctions of hadrons in 3 + 1 physical space-time, thus providing a compelling physical interpretation of the AdS/CFT correspondence principle and AdS/QCD, a useful framework which describes the correspondence between theories in a modified AdS5 background and confining field theories in physical space-time. To a first semiclassical approximation, where quantum loops and quark masses are not included, this approach leads to a single-variable light-front Schroedinger equation which determines the eigenspectrum and the light-front wavefunctions of hadrons for general spinmore » and orbital angular momentum. The coordinate z in AdS space is uniquely identified with a Lorentz-invariant coordinate {zeta} which measures the separation of the constituents within a hadron at equal light-front time. The internal structure of hadrons is explicitly introduced and the angular momentum of the constituents plays a key role. We give an overview of the light-front holographic approach to strongly coupled QCD. In particular, we study the photon-to-meson transition form factors (TFFs) F{sub M{gamma}}(Q{sup 2}) for {gamma}{gamma}* {yields} M using light-front holographic methods. The results for the TFFs for the {eta} and {eta}' mesons are also presented. Some novel features of QCD are discussed, including the consequences of confinement for quark and gluon condensates. A method for computing the hadronization of quark and gluon jets at the amplitude level is outlined.« less
Parreño, Assumpta; Savage, Martin J.; Tiburzi, Brian C.; ...
2017-06-23
We used lattice QCD calculations with background magnetic fields to determine the magnetic moments of the octet baryons. Computations are performed at the physical value of the strange quark mass, and two values of the light quark mass, one corresponding to the SU(3) flavor-symmetric point, where the pion mass is m π ~ 800 MeV, and the other corresponding to a pion mass m π ~ 450 MeV. The moments are found to exhibit only mild pion-mass dependence when expressed in terms of appropriately chosen magneton units---the natural baryon magneton. This suggests that simple extrapolations can be used to determinemore » magnetic moments at the physical point, and extrapolated results are found to agree with experiment within uncertainties. A curious pattern is revealed among the anomalous baryon magnetic moments which is linked to the constituent quark model, however, careful scrutiny exposes additional features. Relations expected to hold in the large-N c limit of QCD are studied; and, in one case, the quark model prediction is significantly closer to the extracted values than the large-N c prediction. The magnetically coupled Λ-Σ 0 system is treated in detail at the SU(3) F point, with the lattice QCD results comparing favorably with predictions based on SU(3) F symmetry. Our analysis enables the first extraction of the isovector transition magnetic polarizability. The possibility that large magnetic fields stabilize strange matter is explored, but such a scenario is found to be unlikely.« less
1986-12-01
17 III. Analysis of Parallel Design ................................................ 18 Parallel Abstract Data ...Types ........................................... 18 Abstract Data Type .................................................. 19 Parallel ADT...22 Data -Structure Design ........................................... 23 Object-Oriented Design
Exploring Partonic Structure of Hadrons Using ab initio Lattice QCD Calculations.
Ma, Yan-Qing; Qiu, Jian-Wei
2018-01-12
Following our previous proposal, we construct a class of good "lattice cross sections" (LCSs), from which we can study the partonic structure of hadrons from ab initio lattice QCD calculations. These good LCSs, on the one hand, can be calculated directly in lattice QCD, and on the other hand, can be factorized into parton distribution functions (PDFs) with calculable coefficients, in the same way as QCD factorization for factorizable hadronic cross sections. PDFs could be extracted from QCD global analysis of the lattice QCD generated data of LCSs. We also show that the proposed functions for lattice QCD calculation of PDFs in the literature are special cases of these good LCSs.
Höche, Stefan; Krauss, Frank; Maierhöfer, Philipp; ...
2015-06-26
We present differential cross sections for the production of top-quark pairs in conjunction with up to two jets, computed at next-to-leading order in perturbative QCD and consistently merged with a parton shower in the SHERPA+OPENLOOPS framework. Top quark decays including spin correlation effects are taken into account at leading order accuracy. The calculation yields a unified description of top-pair plus multi-jet production, and detailed results are presented for various key observables at the Large Hadron Collider. As a result, a large improvement with respect to the multi-jet merging approach at leading order is found for the total transverse energy spectrum,more » which plays a prominent role in searches for physics beyond the Standard Model.« less
Heavy quark free energy in QCD and in gauge theories with gravity duals
NASA Astrophysics Data System (ADS)
Noronha, Jorge
2010-09-01
Recent lattice results in pure glue SU(3) theory at high temperatures have shown that the expectation value of the renormalized Polyakov loop approaches its asymptotic limit at high temperatures from above. We show that this implies that the “heavy quark free energy” obtained from the renormalized loop computed on the lattice does not behave like a true thermodynamic free energy. While this should be expected to occur in asymptotically free gauge theories such as QCD, we use the gauge/string duality to show that in a large class of strongly coupled gauge theories with nontrivial UV fixed points the Polyakov loop reaches its asymptotic value from above only if the dimension of the relevant operator used to deform the conformal field theory is greater than or equal to 3.
Light-Front Hamiltonian Approach to the Bound-State Problem in Quantum Electrodynamics
NASA Astrophysics Data System (ADS)
Jones, Billy D.
1997-10-01
Why is the study of the Lamb shift in hydrogen, which at the level of detail found in this paper was largely completed by Bethe in 1947, of any real interest today? While completing such a calculation using new techniques may be very interesting for formal and academic reasons, our primary motivation is to lay groundwork for precision bound-state calculations in QCD. The Lamb shift provides an excellent pedagogical tool for illustrating light-front Hamiltonian techniques, which are not widely known; but more importantly it presents three of the central dynamical and computational problems that we must face to make these techniques useful for solving QCD: How does a constituent picture emerge in a gauge field theory? How do bound-state energy scales emerge non-perturbatively? How does rotational symmetry emerge in a non-perturbative light-front calculation?
Neutron Electric Dipole Moment from Gauge-String Duality.
Bartolini, Lorenzo; Bigazzi, Francesco; Bolognesi, Stefano; Cotrone, Aldo L; Manenti, Andrea
2017-03-03
We compute the electric dipole moment of nucleons in the large N_{c} QCD model by Witten, Sakai, and Sugimoto with N_{f}=2 degenerate massive flavors. Baryons in the model are instantonic solitons of an effective five-dimensional action describing the whole tower of mesonic fields. We find that the dipole electromagnetic form factor of the nucleons, induced by a finite topological θ angle, exhibits complete vector meson dominance. We are able to evaluate the contribution of each vector meson to the final result-a small number of modes are relevant to obtain an accurate estimate. Extrapolating the model parameters to real QCD data, the neutron electric dipole moment is evaluated to be d_{n}=1.8×10^{-16}θ e cm. The electric dipole moment of the proton is exactly the opposite.
Isoscalar ππ Scattering and the σ Meson Resonance from QCD.
Briceño, Raul A; Dudek, Jozef J; Edwards, Robert G; Wilson, David J
2017-01-13
We present for the first time a determination of the energy dependence of the isoscalar ππ elastic scattering phase shift within a first-principles numerical lattice approach to QCD. Hadronic correlation functions are computed including all required quark propagation diagrams, and from these the discrete spectrum of states in the finite volume defined by the lattice boundary is extracted. From the volume dependence of the spectrum, we obtain the S-wave phase shift up to the KK[over ¯] threshold. Calculations are performed at two values of the u, d quark mass corresponding to m_{π}=236,391 MeV, and the resulting amplitudes are described in terms of a σ meson which evolves from a bound state below the ππ threshold at the heavier quark mass to a broad resonance at the lighter quark mass.
Benchmark results in the 2D lattice Thirring model with a chemical potential
NASA Astrophysics Data System (ADS)
Ayyar, Venkitesh; Chandrasekharan, Shailesh; Rantaharju, Jarno
2018-03-01
We study the two-dimensional lattice Thirring model in the presence of a fermion chemical potential. Our model is asymptotically free and contains massive fermions that mimic a baryon and light bosons that mimic pions. Hence, it is a useful toy model for QCD, especially since it, too, suffers from a sign problem in the auxiliary field formulation in the presence of a fermion chemical potential. In this work, we formulate the model in both the world line and fermion-bag representations and show that the sign problem can be completely eliminated with open boundary conditions when the fermions are massless. Hence, we are able accurately compute a variety of interesting quantities in the model, and these results could provide benchmarks for other methods that are being developed to solve the sign problem in QCD.
Connected and disconnected contractions in pion-pion scattering
NASA Astrophysics Data System (ADS)
Acharya, Neramballi Ripunjay; Guo, Feng-Kun; Meißner, Ulf-G.; Seng, Chien-Yeah
2017-09-01
We show that the interplay of chiral effective field theory and lattice QCD can be used in the evaluation of so-called disconnected diagrams, which appear in the study of the isoscalar and isovector channels of pion-pion scattering and have long been a major challenge for the lattice community. By means of partially-quenched chiral perturbation theory, we distinguish and analyze the effects from different types of contraction diagrams to the pion-pion scattering amplitude, including its scattering lengths and the energy-dependence of its imaginary part. Our results may be used to test the current degree of accuracy of lattice calculation in the handling of disconnected diagrams, as well as to set criteria for the future improvement of relevant lattice computational techniques that may play a critical role in the study of other interesting QCD matrix elements.
Hypercluster Parallel Processor
NASA Technical Reports Server (NTRS)
Blech, Richard A.; Cole, Gary L.; Milner, Edward J.; Quealy, Angela
1992-01-01
Hypercluster computer system includes multiple digital processors, operation of which coordinated through specialized software. Configurable according to various parallel-computing architectures of shared-memory or distributed-memory class, including scalar computer, vector computer, reduced-instruction-set computer, and complex-instruction-set computer. Designed as flexible, relatively inexpensive system that provides single programming and operating environment within which one can investigate effects of various parallel-computing architectures and combinations on performance in solution of complicated problems like those of three-dimensional flows in turbomachines. Hypercluster software and architectural concepts are in public domain.
NASA Technical Reports Server (NTRS)
Hsia, T. C.; Lu, G. Z.; Han, W. H.
1987-01-01
In advanced robot control problems, on-line computation of inverse Jacobian solution is frequently required. Parallel processing architecture is an effective way to reduce computation time. A parallel processing architecture is developed for the inverse Jacobian (inverse differential kinematic equation) of the PUMA arm. The proposed pipeline/parallel algorithm can be inplemented on an IC chip using systolic linear arrays. This implementation requires 27 processing cells and 25 time units. Computation time is thus significantly reduced.
Studies of QCD structure in high-energy collisions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nadolsky, Pavel M.
2016-06-26
”Studies of QCD structure in high-energy collisions” is a research project in theoretical particle physics at Southern Methodist University funded by US DOE Award DE-SC0013681. The award furnished bridge funding for one year (2015/04/15-2016/03/31) between the periods funded by Nadolsky’s DOE Early Career Research Award DE-SC0003870 (in 2010-2015) and a DOE grant DE-SC0010129 for SMU Department of Physics (starting in April 2016). The primary objective of the research is to provide theoretical predictions for Run-2 of the CERN Large Hadron Collider (LHC). The LHC physics program relies on state-of-the-art predictions in the field of quantum chromodynamics. The main effort ofmore » our group went into the global analysis of parton distribution functions (PDFs) employed by the bulk of LHC computations. Parton distributions describe internal structure of protons during ultrarelivistic collisions. A new generation of CTEQ parton distribution functions (PDFs), CT14, was released in summer 2015 and quickly adopted by the HEP community. The new CT14 parametrizations of PDFs were obtained using benchmarked NNLO calculations and latest data from LHC and Tevatron experiments. The group developed advanced methods for the PDF analysis and estimation of uncertainties in LHC predictions associated with the PDFs. We invented and refined a new ’meta-parametrization’ technique that streamlines usage of PDFs in Higgs boson production and other numerous LHC processes, by combining PDFs from various groups using multivariate stochastic sampling. In 2015, the PDF4LHC working group recommended to LHC experimental collaborations to use ’meta-parametrizations’ as a standard technique for computing PDF uncertainties. Finally, to include new QCD processes into the global fits, our group worked on several (N)NNLO calculations.« less
Perturbative corrections to B → D form factors in QCD
NASA Astrophysics Data System (ADS)
Wang, Yu-Ming; Wei, Yan-Bing; Shen, Yue-Long; Lü, Cai-Dian
2017-06-01
We compute perturbative QCD corrections to B → D form factors at leading power in Λ/ m b , at large hadronic recoil, from the light-cone sum rules (LCSR) with B-meson distribution amplitudes in HQET. QCD factorization for the vacuum-to- B-meson correlation function with an interpolating current for the D-meson is demonstrated explicitly at one loop with the power counting scheme {m}_c˜ O(√{Λ {m}_b}) . The jet functions encoding information of the hard-collinear dynamics in the above-mentioned correlation function are complicated by the appearance of an additional hard-collinear scale m c , compared to the counterparts entering the factorization formula of the vacuum-to- B-meson correction function for the construction of B → π from factors. Inspecting the next-to-leading-logarithmic sum rules for the form factors of B → Dℓν indicates that perturbative corrections to the hard-collinear functions are more profound than that for the hard functions, with the default theory inputs, in the physical kinematic region. We further compute the subleading power correction induced by the three-particle quark-gluon distribution amplitudes of the B-meson at tree level employing the background gluon field approach. The LCSR predictions for the semileptonic B → Dℓν form factors are then extrapolated to the entire kinematic region with the z-series parametrization. Phenomenological implications of our determinations for the form factors f BD +,0 ( q 2) are explored by investigating the (differential) branching fractions and the R( D) ratio of B → Dℓν and by determining the CKM matrix element |V cb | from the total decay rate of B → Dμν μ .
A scalable parallel black oil simulator on distributed memory parallel computers
NASA Astrophysics Data System (ADS)
Wang, Kun; Liu, Hui; Chen, Zhangxin
2015-11-01
This paper presents our work on developing a parallel black oil simulator for distributed memory computers based on our in-house parallel platform. The parallel simulator is designed to overcome the performance issues of common simulators that are implemented for personal computers and workstations. The finite difference method is applied to discretize the black oil model. In addition, some advanced techniques are employed to strengthen the robustness and parallel scalability of the simulator, including an inexact Newton method, matrix decoupling methods, and algebraic multigrid methods. A new multi-stage preconditioner is proposed to accelerate the solution of linear systems from the Newton methods. Numerical experiments show that our simulator is scalable and efficient, and is capable of simulating extremely large-scale black oil problems with tens of millions of grid blocks using thousands of MPI processes on parallel computers.
Automatic Generation of OpenMP Directives and Its Application to Computational Fluid Dynamics Codes
NASA Technical Reports Server (NTRS)
Yan, Jerry; Jin, Haoqiang; Frumkin, Michael; Yan, Jerry (Technical Monitor)
2000-01-01
The shared-memory programming model is a very effective way to achieve parallelism on shared memory parallel computers. As great progress was made in hardware and software technologies, performance of parallel programs with compiler directives has demonstrated large improvement. The introduction of OpenMP directives, the industrial standard for shared-memory programming, has minimized the issue of portability. In this study, we have extended CAPTools, a computer-aided parallelization toolkit, to automatically generate OpenMP-based parallel programs with nominal user assistance. We outline techniques used in the implementation of the tool and discuss the application of this tool on the NAS Parallel Benchmarks and several computational fluid dynamics codes. This work demonstrates the great potential of using the tool to quickly port parallel programs and also achieve good performance that exceeds some of the commercial tools.
NASA Astrophysics Data System (ADS)
Murni, Bustamam, A.; Ernastuti, Handhika, T.; Kerami, D.
2017-07-01
Calculation of the matrix-vector multiplication in the real-world problems often involves large matrix with arbitrary size. Therefore, parallelization is needed to speed up the calculation process that usually takes a long time. Graph partitioning techniques that have been discussed in the previous studies cannot be used to complete the parallelized calculation of matrix-vector multiplication with arbitrary size. This is due to the assumption of graph partitioning techniques that can only solve the square and symmetric matrix. Hypergraph partitioning techniques will overcome the shortcomings of the graph partitioning technique. This paper addresses the efficient parallelization of matrix-vector multiplication through hypergraph partitioning techniques using CUDA GPU-based parallel computing. CUDA (compute unified device architecture) is a parallel computing platform and programming model that was created by NVIDIA and implemented by the GPU (graphics processing unit).
[Series: Medical Applications of the PHITS Code (2): Acceleration by Parallel Computing].
Furuta, Takuya; Sato, Tatsuhiko
2015-01-01
Time-consuming Monte Carlo dose calculation becomes feasible owing to the development of computer technology. However, the recent development is due to emergence of the multi-core high performance computers. Therefore, parallel computing becomes a key to achieve good performance of software programs. A Monte Carlo simulation code PHITS contains two parallel computing functions, the distributed-memory parallelization using protocols of message passing interface (MPI) and the shared-memory parallelization using open multi-processing (OpenMP) directives. Users can choose the two functions according to their needs. This paper gives the explanation of the two functions with their advantages and disadvantages. Some test applications are also provided to show their performance using a typical multi-core high performance workstation.
Evolving binary classifiers through parallel computation of multiple fitness cases.
Cagnoni, Stefano; Bergenti, Federico; Mordonini, Monica; Adorni, Giovanni
2005-06-01
This paper describes two versions of a novel approach to developing binary classifiers, based on two evolutionary computation paradigms: cellular programming and genetic programming. Such an approach achieves high computation efficiency both during evolution and at runtime. Evolution speed is optimized by allowing multiple solutions to be computed in parallel. Runtime performance is optimized explicitly using parallel computation in the case of cellular programming or implicitly taking advantage of the intrinsic parallelism of bitwise operators on standard sequential architectures in the case of genetic programming. The approach was tested on a digit recognition problem and compared with a reference classifier.
NASA Astrophysics Data System (ADS)
Georgiev, K.; Zlatev, Z.
2010-11-01
The Danish Eulerian Model (DEM) is an Eulerian model for studying the transport of air pollutants on large scale. Originally, the model was developed at the National Environmental Research Institute of Denmark. The model computational domain covers Europe and some neighbour parts belong to the Atlantic Ocean, Asia and Africa. If DEM model is to be applied by using fine grids, then its discretization leads to a huge computational problem. This implies that such a model as DEM must be run only on high-performance computer architectures. The implementation and tuning of such a complex large-scale model on each different computer is a non-trivial task. Here, some comparison results of running of this model on different kind of vector (CRAY C92A, Fujitsu, etc.), parallel computers with distributed memory (IBM SP, CRAY T3E, Beowulf clusters, Macintosh G4 clusters, etc.), parallel computers with shared memory (SGI Origin, SUN, etc.) and parallel computers with two levels of parallelism (IBM SMP, IBM BlueGene/P, clusters of multiprocessor nodes, etc.) will be presented. The main idea in the parallel version of DEM is domain partitioning approach. Discussions according to the effective use of the cache and hierarchical memories of the modern computers as well as the performance, speed-ups and efficiency achieved will be done. The parallel code of DEM, created by using MPI standard library, appears to be highly portable and shows good efficiency and scalability on different kind of vector and parallel computers. Some important applications of the computer model output are presented in short.
A Debugger for Computational Grid Applications
NASA Technical Reports Server (NTRS)
Hood, Robert; Jost, Gabriele; Biegel, Bryan (Technical Monitor)
2001-01-01
This viewgraph presentation gives an overview of a debugger for computational grid applications. Details are given on NAS parallel tools groups (including parallelization support tools, evaluation of various parallelization strategies, and distributed and aggregated computing), debugger dependencies, scalability, initial implementation, the process grid, and information on Globus.
Application of a Scalable, Parallel, Unstructured-Grid-Based Navier-Stokes Solver
NASA Technical Reports Server (NTRS)
Parikh, Paresh
2001-01-01
A parallel version of an unstructured-grid based Navier-Stokes solver, USM3Dns, previously developed for efficient operation on a variety of parallel computers, has been enhanced to incorporate upgrades made to the serial version. The resultant parallel code has been extensively tested on a variety of problems of aerospace interest and on two sets of parallel computers to understand and document its characteristics. An innovative grid renumbering construct and use of non-blocking communication are shown to produce superlinear computing performance. Preliminary results from parallelization of a recently introduced "porous surface" boundary condition are also presented.
How to Build an AppleSeed: A Parallel Macintosh Cluster for Numerically Intensive Computing
NASA Astrophysics Data System (ADS)
Decyk, V. K.; Dauger, D. E.
We have constructed a parallel cluster consisting of a mixture of Apple Macintosh G3 and G4 computers running the Mac OS, and have achieved very good performance on numerically intensive, parallel plasma particle-incell simulations. A subset of the MPI message-passing library was implemented in Fortran77 and C. This library enabled us to port code, without modification, from other parallel processors to the Macintosh cluster. Unlike Unix-based clusters, no special expertise in operating systems is required to build and run the cluster. This enables us to move parallel computing from the realm of experts to the main stream of computing.
Parallel simulation of tsunami inundation on a large-scale supercomputer
NASA Astrophysics Data System (ADS)
Oishi, Y.; Imamura, F.; Sugawara, D.
2013-12-01
An accurate prediction of tsunami inundation is important for disaster mitigation purposes. One approach is to approximate the tsunami wave source through an instant inversion analysis using real-time observation data (e.g., Tsushima et al., 2009) and then use the resulting wave source data in an instant tsunami inundation simulation. However, a bottleneck of this approach is the large computational cost of the non-linear inundation simulation and the computational power of recent massively parallel supercomputers is helpful to enable faster than real-time execution of a tsunami inundation simulation. Parallel computers have become approximately 1000 times faster in 10 years (www.top500.org), and so it is expected that very fast parallel computers will be more and more prevalent in the near future. Therefore, it is important to investigate how to efficiently conduct a tsunami simulation on parallel computers. In this study, we are targeting very fast tsunami inundation simulations on the K computer, currently the fastest Japanese supercomputer, which has a theoretical peak performance of 11.2 PFLOPS. One computing node of the K computer consists of 1 CPU with 8 cores that share memory, and the nodes are connected through a high-performance torus-mesh network. The K computer is designed for distributed-memory parallel computation, so we have developed a parallel tsunami model. Our model is based on TUNAMI-N2 model of Tohoku University, which is based on a leap-frog finite difference method. A grid nesting scheme is employed to apply high-resolution grids only at the coastal regions. To balance the computation load of each CPU in the parallelization, CPUs are first allocated to each nested layer in proportion to the number of grid points of the nested layer. Using CPUs allocated to each layer, 1-D domain decomposition is performed on each layer. In the parallel computation, three types of communication are necessary: (1) communication to adjacent neighbours for the finite difference calculation, (2) communication between adjacent layers for the calculations to connect each layer, and (3) global communication to obtain the time step which satisfies the CFL condition in the whole domain. A preliminary test on the K computer showed the parallel efficiency on 1024 cores was 57% relative to 64 cores. We estimate that the parallel efficiency will be considerably improved by applying a 2-D domain decomposition instead of the present 1-D domain decomposition in future work. The present parallel tsunami model was applied to the 2011 Great Tohoku tsunami. The coarsest resolution layer covers a 758 km × 1155 km region with a 405 m grid spacing. A nesting of five layers was used with the resolution ratio of 1/3 between nested layers. The finest resolution region has 5 m resolution and covers most of the coastal region of Sendai city. To complete 2 hours of simulation time, the serial (non-parallel) computation took approximately 4 days on a workstation. To complete the same simulation on 1024 cores of the K computer, it took 45 minutes which is more than two times faster than real-time. This presentation discusses the updated parallel computational performance and the efficient use of the K computer when considering the characteristics of the tsunami inundation simulation model in relation to the characteristics and capabilities of the K computer.
NASA Astrophysics Data System (ADS)
Narison, Stephan
2004-05-01
About Stephan Narison; Outline of the book; Preface; Acknowledgements; Part I. General Introduction: 1. A short flash on particle physics; 2. The pre-QCD era; 3. The QCD story; 4. Field theory ingredients; Part II. QCD Gauge Theory: 5. Lagrangian and gauge invariance; 6. Quantization using path integral; 7. QCD and its global invariance; Part III. MS scheme for QCD and QED: Introduction; 8. Dimensional regularization; 9. The MS renormalization scheme; 10. Renormalization of operators using the background field method; 11. The renormalization group; 12. Other renormalization schemes; 13. MS scheme for QED; 14. High-precision low-energy QED tests; Part IV. Deep Inelastic Scattering at Hadron Colliders: 15. OPE for deep inelastic scattering; 16. Unpolarized lepton-hadron scattering; 17. The Altarelli-Parisi equation; 18. More on unpolarized deep inelastic scatterings; 19. Polarized deep-inelastic processes; 20. Drell-Yan process; 21. One 'prompt photon' inclusive production; Part V. Hard Processes in e+e- Collisions: Introduction; 22. One hadron inclusive production; 23. gg scatterings and the 'spin' of the photon; 24. QCD jets; 25. Total inclusive hadron productions; Part VI. Summary of QCD Tests and as Measurements; Part VII. Power Corrections in QCD: 26. Introduction; 27. The SVZ expansion; 28. Technologies for evaluating Wilson coefficients; 29. Renormalons; 30. Beyond the SVZ expansion; Part VIII. QCD Two-Point Functions: 31. References guide to original works; 32. (Pseudo)scalar correlators; 33. (Axial-)vector two-point functions; 34. Tensor-quark correlator; 35. Baryonic correlators; 36. Four-quark correlators; 37. Gluonia correlators; 38. Hybrid correlators; 39. Correlators in x-space; Part IX. QCD Non-Perturbative Methods: 40. Introduction; 41. Lattice gauge theory; 42. Chiral perturbation theory; 43. Models of the QCD effective action; 44. Heavy quark effective theory; 45. Potential approaches to quarkonia; 46. On monopole and confinement; Part X. QCD Spectral Sum Rules: 47. Introduction; 48. Theoretical foundations; 49. Survey of QCD spectral sum rules; 50. Weinberg and DMO sum rules; 51. The QCD coupling as; 52. The QCD condensates; 53. Light and heavy quark masses, etc.; 54. Hadron spectroscopy; 55. D, B and Bc exclusive weak decays; 56. B0(s)-B0(s) mixing, kaon CP violation; 57. Thermal behaviour of QCD; 58. More on spectral sum rules; Part XI. Appendix A: physical constants and unites; Appendix B: weight factors for SU(N)c; Appendix C: coordinates and momenta; Appendix D: Dirac equation and matrices; Appendix E: Feynman rules; Appendix F: Feynman integrals; Appendix G: useful formulae for the sum rules; Bibliography; Index.
NASA Astrophysics Data System (ADS)
Narison, Stephan
2007-07-01
About Stephan Narison; Outline of the book; Preface; Acknowledgements; Part I. General Introduction: 1. A short flash on particle physics; 2. The pre-QCD era; 3. The QCD story; 4. Field theory ingredients; Part II. QCD Gauge Theory: 5. Lagrangian and gauge invariance; 6. Quantization using path integral; 7. QCD and its global invariance; Part III. MS scheme for QCD and QED: Introduction; 8. Dimensional regularization; 9. The MS renormalization scheme; 10. Renormalization of operators using the background field method; 11. The renormalization group; 12. Other renormalization schemes; 13. MS scheme for QED; 14. High-precision low-energy QED tests; Part IV. Deep Inelastic Scattering at Hadron Colliders: 15. OPE for deep inelastic scattering; 16. Unpolarized lepton-hadron scattering; 17. The Altarelli-Parisi equation; 18. More on unpolarized deep inelastic scatterings; 19. Polarized deep-inelastic processes; 20. Drell-Yan process; 21. One 'prompt photon' inclusive production; Part V. Hard Processes in e+e- Collisions: Introduction; 22. One hadron inclusive production; 23. gg scatterings and the 'spin' of the photon; 24. QCD jets; 25. Total inclusive hadron productions; Part VI. Summary of QCD Tests and as Measurements; Part VII. Power Corrections in QCD: 26. Introduction; 27. The SVZ expansion; 28. Technologies for evaluating Wilson coefficients; 29. Renormalons; 30. Beyond the SVZ expansion; Part VIII. QCD Two-Point Functions: 31. References guide to original works; 32. (Pseudo)scalar correlators; 33. (Axial-)vector two-point functions; 34. Tensor-quark correlator; 35. Baryonic correlators; 36. Four-quark correlators; 37. Gluonia correlators; 38. Hybrid correlators; 39. Correlators in x-space; Part IX. QCD Non-Perturbative Methods: 40. Introduction; 41. Lattice gauge theory; 42. Chiral perturbation theory; 43. Models of the QCD effective action; 44. Heavy quark effective theory; 45. Potential approaches to quarkonia; 46. On monopole and confinement; Part X. QCD Spectral Sum Rules: 47. Introduction; 48. Theoretical foundations; 49. Survey of QCD spectral sum rules; 50. Weinberg and DMO sum rules; 51. The QCD coupling as; 52. The QCD condensates; 53. Light and heavy quark masses, etc.; 54. Hadron spectroscopy; 55. D, B and Bc exclusive weak decays; 56. B0(s)-B0(s) mixing, kaon CP violation; 57. Thermal behaviour of QCD; 58. More on spectral sum rules; Part XI. Appendix A: physical constants and unites; Appendix B: weight factors for SU(N)c; Appendix C: coordinates and momenta; Appendix D: Dirac equation and matrices; Appendix E: Feynman rules; Appendix F: Feynman integrals; Appendix G: useful formulae for the sum rules; Bibliography; Index.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Boz, Tamer; Skullerud, Jon-Ivar; Centre for the Subatomic Structure of Matter, Adelaide University, Adelaide, SA 5005
2016-01-22
QCD at high chemical potential has interesting properties such as deconfinement of quarks. Two-color QCD, which enables numerical simulations on the lattice, constitutes a laboratory to study QCD at high chemical potential. Among the interesting properties of two-color QCD at high density is the diquark condensation, for which we present recent results obtained on a finer lattice compared to previous studies. The quark propagator in two-color QCD at non-zero chemical potential is referred to as the Gor’kov propagator. We express the Gor’kov propagator in terms of form factors and present recent lattice simulation results.
Progress in vacuum susceptibilities and their applications to the chiral phase transition of QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cui, Zhu-Fang, E-mail: phycui@nju.edu.cn; State Key Laboratory of Theoretical Physics, Institute of Theoretical Physics, CAS, Beijing, 100190; Hou, Feng-Yao
2015-07-15
The QCD vacuum condensates and various vacuum susceptibilities are all important parameters which characterize the nonperturbative properties of the QCD vacuum. In the QCD sum rules external field formula, various QCD vacuum susceptibilities play important roles in determining the properties of hadrons. In this paper, we review the recent progress in studies of vacuum susceptibilities together with their applications to the chiral phase transition of QCD. The results of the tensor, the vector, the axial–vector, the scalar, and the pseudo-scalar vacuum susceptibilities are shown in detail in the framework of Dyson–Schwinger equations.
Lee, Wei-Po; Hsiao, Yu-Ting; Hwang, Wei-Che
2014-01-16
To improve the tedious task of reconstructing gene networks through testing experimentally the possible interactions between genes, it becomes a trend to adopt the automated reverse engineering procedure instead. Some evolutionary algorithms have been suggested for deriving network parameters. However, to infer large networks by the evolutionary algorithm, it is necessary to address two important issues: premature convergence and high computational cost. To tackle the former problem and to enhance the performance of traditional evolutionary algorithms, it is advisable to use parallel model evolutionary algorithms. To overcome the latter and to speed up the computation, it is advocated to adopt the mechanism of cloud computing as a promising solution: most popular is the method of MapReduce programming model, a fault-tolerant framework to implement parallel algorithms for inferring large gene networks. This work presents a practical framework to infer large gene networks, by developing and parallelizing a hybrid GA-PSO optimization method. Our parallel method is extended to work with the Hadoop MapReduce programming model and is executed in different cloud computing environments. To evaluate the proposed approach, we use a well-known open-source software GeneNetWeaver to create several yeast S. cerevisiae sub-networks and use them to produce gene profiles. Experiments have been conducted and the results have been analyzed. They show that our parallel approach can be successfully used to infer networks with desired behaviors and the computation time can be largely reduced. Parallel population-based algorithms can effectively determine network parameters and they perform better than the widely-used sequential algorithms in gene network inference. These parallel algorithms can be distributed to the cloud computing environment to speed up the computation. By coupling the parallel model population-based optimization method and the parallel computational framework, high quality solutions can be obtained within relatively short time. This integrated approach is a promising way for inferring large networks.
2014-01-01
Background To improve the tedious task of reconstructing gene networks through testing experimentally the possible interactions between genes, it becomes a trend to adopt the automated reverse engineering procedure instead. Some evolutionary algorithms have been suggested for deriving network parameters. However, to infer large networks by the evolutionary algorithm, it is necessary to address two important issues: premature convergence and high computational cost. To tackle the former problem and to enhance the performance of traditional evolutionary algorithms, it is advisable to use parallel model evolutionary algorithms. To overcome the latter and to speed up the computation, it is advocated to adopt the mechanism of cloud computing as a promising solution: most popular is the method of MapReduce programming model, a fault-tolerant framework to implement parallel algorithms for inferring large gene networks. Results This work presents a practical framework to infer large gene networks, by developing and parallelizing a hybrid GA-PSO optimization method. Our parallel method is extended to work with the Hadoop MapReduce programming model and is executed in different cloud computing environments. To evaluate the proposed approach, we use a well-known open-source software GeneNetWeaver to create several yeast S. cerevisiae sub-networks and use them to produce gene profiles. Experiments have been conducted and the results have been analyzed. They show that our parallel approach can be successfully used to infer networks with desired behaviors and the computation time can be largely reduced. Conclusions Parallel population-based algorithms can effectively determine network parameters and they perform better than the widely-used sequential algorithms in gene network inference. These parallel algorithms can be distributed to the cloud computing environment to speed up the computation. By coupling the parallel model population-based optimization method and the parallel computational framework, high quality solutions can be obtained within relatively short time. This integrated approach is a promising way for inferring large networks. PMID:24428926
Parallel CE/SE Computations via Domain Decomposition
NASA Technical Reports Server (NTRS)
Himansu, Ananda; Jorgenson, Philip C. E.; Wang, Xiao-Yen; Chang, Sin-Chung
2000-01-01
This paper describes the parallelization strategy and achieved parallel efficiency of an explicit time-marching algorithm for solving conservation laws. The Space-Time Conservation Element and Solution Element (CE/SE) algorithm for solving the 2D and 3D Euler equations is parallelized with the aid of domain decomposition. The parallel efficiency of the resultant algorithm on a Silicon Graphics Origin 2000 parallel computer is checked.
The b Quark Fragmentation Function, From LEP to TeVatron
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ben-haim, Eli
2004-12-01
The b quark fragmentation distribution has been measured, using data registered by the DELPHI experiment at the Z pole, in the years 1994-1995. The measurement made use of 176000 inclusively reconstructed B meson candidates. The errors of this measurement are dominated by systematic effects, the principal ones being related to the energy calibration. The distribution has been established in a nine bin histogram. Its mean value has been found to be
Parallel Algorithms for Least Squares and Related Computations.
1991-03-22
for dense computations in linear algebra . The work has recently been published in a general reference book on parallel algorithms by SIAM. AFO SR...written his Ph.D. dissertation with the principal investigator. (See publication 6.) • Parallel Algorithms for Dense Linear Algebra Computations. Our...and describe and to put into perspective a selection of the more important parallel algorithms for numerical linear algebra . We give a major new
Reliability models for dataflow computer systems
NASA Technical Reports Server (NTRS)
Kavi, K. M.; Buckles, B. P.
1985-01-01
The demands for concurrent operation within a computer system and the representation of parallelism in programming languages have yielded a new form of program representation known as data flow (DENN 74, DENN 75, TREL 82a). A new model based on data flow principles for parallel computations and parallel computer systems is presented. Necessary conditions for liveness and deadlock freeness in data flow graphs are derived. The data flow graph is used as a model to represent asynchronous concurrent computer architectures including data flow computers.
Parallel computing method for simulating hydrological processesof large rivers under climate change
NASA Astrophysics Data System (ADS)
Wang, H.; Chen, Y.
2016-12-01
Climate change is one of the proverbial global environmental problems in the world.Climate change has altered the watershed hydrological processes in time and space distribution, especially in worldlarge rivers.Watershed hydrological process simulation based on physically based distributed hydrological model can could have better results compared with the lumped models.However, watershed hydrological process simulation includes large amount of calculations, especially in large rivers, thus needing huge computing resources that may not be steadily available for the researchers or at high expense, this seriously restricted the research and application. To solve this problem, the current parallel method are mostly parallel computing in space and time dimensions.They calculate the natural features orderly thatbased on distributed hydrological model by grid (unit, a basin) from upstream to downstream.This articleproposes ahigh-performancecomputing method of hydrological process simulation with high speedratio and parallel efficiency.It combinedthe runoff characteristics of time and space of distributed hydrological model withthe methods adopting distributed data storage, memory database, distributed computing, parallel computing based on computing power unit.The method has strong adaptability and extensibility,which means it canmake full use of the computing and storage resources under the condition of limited computing resources, and the computing efficiency can be improved linearly with the increase of computing resources .This method can satisfy the parallel computing requirements ofhydrological process simulation in small, medium and large rivers.
High Performance Biocomputation
2005-03-01
in some other fields (e.g. computational hydrodynamics, lattice quantum chroniodynamics, etc.) but appears wholly inappropriate here as pointed out...restrict the overall conformational space by putting the system on a lattice . These have been used to great effect to study folding kinetics. These...many important problems to be worked on, not a single unique challenge (contrast this to QCD , for example). " almost all problems require significant
One-dimensional QCD in thimble regularization
NASA Astrophysics Data System (ADS)
Di Renzo, F.; Eruzzi, G.
2018-01-01
QCD in 0 +1 dimensions is numerically solved via thimble regularization. In the context of this toy model, a general formalism is presented for S U (N ) theories. The sign problem that the theory displays is a genuine one, stemming from a (quark) chemical potential. Three stationary points are present in the original (real) domain of integration, so that contributions from all the thimbles associated to them are to be taken into account: we show how semiclassical computations can provide hints on the regions of parameter space where this is absolutely crucial. Known analytical results for the chiral condensate and the Polyakov loop are correctly reproduced: this is in particular trivial at high values of the number of flavors Nf. In this regime we notice that the single thimble dominance scenario takes place (the dominant thimble is the one associated to the identity). At low values of Nf computations can be more difficult. It is important to stress that this is not at all a consequence of the original sign problem (not even via the residual phase). The latter is always under control, while accidental, delicate cancelations of contributions coming from different thimbles can be in place in (restricted) regions of the parameter space.
QCD next-to-leading-order predictions matched to parton showers for vector-like quark models.
Fuks, Benjamin; Shao, Hua-Sheng
2017-01-01
Vector-like quarks are featured by a wealth of beyond the Standard Model theories and are consequently an important goal of many LHC searches for new physics. Those searches, as well as most related phenomenological studies, however, rely on predictions evaluated at the leading-order accuracy in QCD and consider well-defined simplified benchmark scenarios. Adopting an effective bottom-up approach, we compute next-to-leading-order predictions for vector-like-quark pair production and single production in association with jets, with a weak or with a Higgs boson in a general new physics setup. We additionally compute vector-like-quark contributions to the production of a pair of Standard Model bosons at the same level of accuracy. For all processes under consideration, we focus both on total cross sections and on differential distributions, most these calculations being performed for the first time in our field. As a result, our work paves the way to precise extraction of experimental limits on vector-like quarks thanks to an accurate control of the shapes of the relevant observables and emphasise the extra handles that could be provided by novel vector-like-quark probes never envisaged so far.
Hyper-scaling relations in the conformal window from dynamic AdS/QCD
NASA Astrophysics Data System (ADS)
Evans, Nick; Scott, Marc
2014-09-01
Dynamic AdS/QCD is a holographic model of strongly coupled gauge theories with the dynamics included through the running anomalous dimension of the quark bilinear, γ. We apply it to describe the physics of massive quarks in the conformal window of SU(Nc) gauge theories with Nf fundamental flavors, assuming the perturbative two-loop running for γ. We show that to find regular, holographic renormalization group flows in the infrared, the decoupling of the quark flavors at the scale of the mass is important, and enact it through suitable boundary conditions when the flavors become on shell. We can then compute the quark condensate and the mesonic spectrum (Mρ,Mπ,Mσ) and decay constants. We compute their scaling dependence on the quark mass for a number of examples. The model matches perturbative expectations for large quark mass and naïve dimensional analysis (including the anomalous dimensions) for small quark mass. The model allows study of the intermediate regime where there is an additional scale from the running of the coupling, and we present results for the deviation of scalings from assuming only the single scale of the mass.
NASA Astrophysics Data System (ADS)
Wilczek, Frank
Introduction Symmetry and the Phenomena of QCD Apparent and Actual Symmetries Asymptotic Freedom Confinement Chiral Symmetry Breaking Chiral Anomalies and Instantons High Temperature QCD: Asymptotic Properties Significance of High Temperature QCD Numerical Indications for Quasi-Free Behavior Ideas About Quark-Gluon Plasma Screening Versus Confinement Models of Chiral Symmetry Breaking More Refined Numerical Experiments High-Temperature QCD: Phase Transitions Yoga of Phase Transitions and Order Parameters Application to Glue Theories Application to Chiral Transitions Close Up on Two Flavors A Genuine Critical Point! (?) High-Density QCD: Methods Hopes, Doubts, and Fruition Another Renormalization Group Pairing Theory Taming the Magnetic Singularity High-Density QCD: Color-Flavor Locking and Quark-Hadron Continuity Gauge Symmetry (Non)Breaking Symmetry Accounting Elementary Excitations A Modified Photon Quark-Hadron Continuity Remembrance of Things Past More Quarks Fewer Quarks and Reality
Archer, Charles J; Blocksome, Michael E; Ratterman, Joseph D; Smith, Brian E
2014-02-11
Endpoint-based parallel data processing in a parallel active messaging interface ('PAMI') of a parallel computer, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes coupled for data communications through the PAMI, including establishing a data communications geometry, the geometry specifying, for tasks representing processes of execution of the parallel application, a set of endpoints that are used in collective operations of the PAMI including a plurality of endpoints for one of the tasks; receiving in endpoints of the geometry an instruction for a collective operation; and executing the instruction for a collective opeartion through the endpoints in dependence upon the geometry, including dividing data communications operations among the plurality of endpoints for one of the tasks.
Archer, Charles J.; Blocksome, Michael A.; Ratterman, Joseph D.; Smith, Brian E.
2014-08-12
Endpoint-based parallel data processing in a parallel active messaging interface (`PAMI`) of a parallel computer, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes coupled for data communications through the PAMI, including establishing a data communications geometry, the geometry specifying, for tasks representing processes of execution of the parallel application, a set of endpoints that are used in collective operations of the PAMI including a plurality of endpoints for one of the tasks; receiving in endpoints of the geometry an instruction for a collective operation; and executing the instruction for a collective operation through the endpoints in dependence upon the geometry, including dividing data communications operations among the plurality of endpoints for one of the tasks.
NASA Astrophysics Data System (ADS)
Decyk, Viktor K.; Dauger, Dean E.
We have constructed a parallel cluster consisting of Apple Macintosh G4 computers running both Classic Mac OS as well as the Unix-based Mac OS X, and have achieved very good performance on numerically intensive, parallel plasma particle-in-cell simulations. Unlike other Unix-based clusters, no special expertise in operating systems is required to build and run the cluster. This enables us to move parallel computing from the realm of experts to the mainstream of computing.
Recent development in lattice QCD studies for three-nucleon forces
NASA Astrophysics Data System (ADS)
Doi, Takumi; HAL QCD Collaboration
2014-09-01
The direct determination of nuclear forces from QCD has been one of the most desirable challenges in nuclear physics. Recently, a first-principles lattice QCD determination is becoming possible by a novel theoretical method, HAL QCD method, in which Nambu-Bethe-Salpeter (NBS) wave functions are utilized. In this talk, I will focus on the study of three-nucleon forces in HAL QCD method by presenting the recent theoretical/numerical development.
NASA Astrophysics Data System (ADS)
Kim, Seyong; Petreczky, Peter; Rothkopf, Alexander
2015-03-01
We investigate the properties of S - and P -wave bottomonium states in the vicinity of the deconfinement transition temperature. The light degrees of freedom are represented by dynamical lattice quantum chromodynamics (QCD) configurations of the HotQCD collaboration with Nf=2 +1 flavors. Bottomonium correlators are obtained from bottom quark propagators, computed in nonrelativistic QCD under the background of these gauge field configurations. The spectral functions for the 3S1 (ϒ ) and 3P1 (χb 1) channel are extracted from the Euclidean time correlators using a novel Bayesian approach in the temperature region 140 MeV ≤T ≤249 MeV and the results are contrasted to those from the standard maximum entropy method. We find that the new Bayesian approach is far superior to the maximum entropy method. It enables us to study reliably the presence or absence of the lowest state signal in the spectral function of a certain channel, even under the limitations present in the finite temperature setup. We find that χb 1 survives up to T =249 MeV , the highest temperature considered in our study, and put stringent constraints on the size of the medium modification of ϒ and χb 1 states.
NASA Astrophysics Data System (ADS)
Harrison, Judd; Davies, Christine T. H.; Wingate, Matthew; Hpqcd Collaboration
2018-03-01
We present results of a lattice QCD calculation of B →D* and Bs→Ds* axial vector matrix elements with both states at rest. These zero recoil matrix elements provide the normalization necessary to infer a value for the CKM matrix element |Vc b| from experimental measurements of B¯ 0→D*+ℓ-ν ¯ and B¯s0→Ds*+ℓ-ν¯ decay. Results are derived from correlation functions computed with highly improved staggered quarks (HISQ) for light, strange, and charm quark propagators, and nonrelativistic QCD for the bottom quark propagator. The calculation of correlation functions employs MILC Collaboration ensembles over a range of three lattice spacings. These gauge field configurations include sea quark effects of charm, strange, and equal-mass up and down quarks. We use ensembles with physically light up and down quarks, as well as heavier values. Our main results are FB→D *(1 )=0.895 ±0.01 0stat±0.024sys and FBs→Ds*(1 )=0.883 ±0.01 2stat±0.02 8sys . We discuss the consequences for |Vc b| in light of recent investigations into the extrapolation of experimental data to zero recoil.
APFEL: A PDF evolution library with QED corrections
NASA Astrophysics Data System (ADS)
Bertone, Valerio; Carrazza, Stefano; Rojo, Juan
2014-06-01
Quantum electrodynamics and electroweak corrections are important ingredients for many theoretical predictions at the LHC. This paper documents APFEL, a new PDF evolution package that allows for the first time to perform DGLAP evolution up to NNLO in QCD and to LO in QED, in the variable-flavor-number scheme and with either pole or MS bar heavy quark masses. APFEL consistently accounts for the QED corrections to the evolution of quark and gluon PDFs and for the contribution from the photon PDF in the proton. The coupled QCD ⊗ QED equations are solved in x-space by means of higher order interpolation, followed by Runge-Kutta solution of the resulting discretized evolution equations. APFEL is based on an innovative and flexible methodology for the sequential solution of the QCD and QED evolution equations and their combination. In addition to PDF evolution, APFEL provides a module that computes Deep-Inelastic Scattering structure functions in the FONLL general-mass variable-flavor-number scheme up to O(αs2) . All the functionalities of APFEL can be accessed via a Graphical User Interface, supplemented with a variety of plotting tools for PDFs, parton luminosities and structure functions. Written in FORTRAN 77, APFEL can also be used via the C/C++ and Python interfaces, and is publicly available from the HepForge repository.
QCD topological susceptibility from the nonlocal chiral quark model
NASA Astrophysics Data System (ADS)
Nam, Seung-Il; Kao, Chung-Wen
2017-06-01
We investigate the quantum chromodynamics (QCD) topological susceptibility χ by using the semi-bosonized nonlocal chiral-quark model (SB-NLχQM) for the leading large- N c contributions. This model is based on the liquid-instanton QCD-vacuum configuration, in which SU(3) flavor symmetry is explicitly broken by the finite current-quark mass ( m u,d, m s) ≈ (5, 135) MeV. To compute χ, we derive the local topological charge-density operator Q t( x) from the effective action of SB-NLχQM. We verify that the derived expression for χ in our model satisfies the Witten- Veneziano (WV) and the Leutwyler-Smilga (LS) formulae, and the Crewther theorem in the chiral limit by construction. Once the average instanton size and the inter-instanton distance are fixed with ρ¯ = 1/3 fm and R¯ = 1 fm, respectively, all the other parameters are determined self-consistently within the model. We obtain χ = (167.67MeV)4, which is comparable with the empirical value χ = (175±5MeV)4 whereas it turns out that χ QL = (194.30MeV)4 in the quenched limit. Thus, we conclude that the value of χ will be reduced around 10 20% by the dynamical-quark contribution.
{lambda}{sub b}{yields}p, {lambda} transition form factors from QCD light-cone sum rules
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang Yuming; Lue Caidian; Shen Yuelong
2009-10-01
Light-cone sum rules for the {lambda}{sub b}{yields}p, {lambda} transition form factors are derived from the correlation functions expanded by the twist of the distribution amplitudes of the {lambda}{sub b} baryon. In terms of the {lambda}{sub b} three-quark distribution amplitude models constrained by the QCD theory, we calculate the form factors at small momentum transfers and compare the results with those estimated in the conventional light-cone sum rules (LCSR) and perturbative QCD approaches. Our results indicate that the two different versions of sum rules can lead to the consistent numbers of form factors responsible for {lambda}{sub b}{yields}p transition. The {lambda}{sub b}{yields}{lambda}more » transition form factors from LCSR with the asymptotic {lambda} baryon distribution amplitudes are found to be almost 1 order larger than those obtained in the {lambda}{sub b}-baryon LCSR, implying that the preasymptotic corrections to the baryonic distribution amplitudes are of great importance. Moreover, the SU(3) symmetry breaking effects between the form factors f{sub 1}{sup {lambda}{sub b}}{sup {yields}}{sup p} and f{sub 1}{sup {lambda}{sub b}}{sup {yields}}{sup {lambda}} are computed as 28{sub -8}{sup +14}% in the framework of {lambda}{sub b}-baryon LCSR.« less
NASA Astrophysics Data System (ADS)
Wei, Xiaohui; Li, Weishan; Tian, Hailong; Li, Hongliang; Xu, Haixiao; Xu, Tianfu
2015-07-01
The numerical simulation of multiphase flow and reactive transport in the porous media on complex subsurface problem is a computationally intensive application. To meet the increasingly computational requirements, this paper presents a parallel computing method and architecture. Derived from TOUGHREACT that is a well-established code for simulating subsurface multi-phase flow and reactive transport problems, we developed a high performance computing THC-MP based on massive parallel computer, which extends greatly on the computational capability for the original code. The domain decomposition method was applied to the coupled numerical computing procedure in the THC-MP. We designed the distributed data structure, implemented the data initialization and exchange between the computing nodes and the core solving module using the hybrid parallel iterative and direct solver. Numerical accuracy of the THC-MP was verified through a CO2 injection-induced reactive transport problem by comparing the results obtained from the parallel computing and sequential computing (original code). Execution efficiency and code scalability were examined through field scale carbon sequestration applications on the multicore cluster. The results demonstrate successfully the enhanced performance using the THC-MP on parallel computing facilities.
A parallel Jacobson-Oksman optimization algorithm. [parallel processing (computers)
NASA Technical Reports Server (NTRS)
Straeter, T. A.; Markos, A. T.
1975-01-01
A gradient-dependent optimization technique which exploits the vector-streaming or parallel-computing capabilities of some modern computers is presented. The algorithm, derived by assuming that the function to be minimized is homogeneous, is a modification of the Jacobson-Oksman serial minimization method. In addition to describing the algorithm, conditions insuring the convergence of the iterates of the algorithm and the results of numerical experiments on a group of sample test functions are presented. The results of these experiments indicate that this algorithm will solve optimization problems in less computing time than conventional serial methods on machines having vector-streaming or parallel-computing capabilities.
Methods of parallel computation applied on granular simulations
NASA Astrophysics Data System (ADS)
Martins, Gustavo H. B.; Atman, Allbens P. F.
2017-06-01
Every year, parallel computing has becoming cheaper and more accessible. As consequence, applications were spreading over all research areas. Granular materials is a promising area for parallel computing. To prove this statement we study the impact of parallel computing in simulations of the BNE (Brazil Nut Effect). This property is due the remarkable arising of an intruder confined to a granular media when vertically shaken against gravity. By means of DEM (Discrete Element Methods) simulations, we study the code performance testing different methods to improve clock time. A comparison between serial and parallel algorithms, using OpenMP® is also shown. The best improvement was obtained by optimizing the function that find contacts using Verlet's cells.
Parallel computation using boundary elements in solid mechanics
NASA Technical Reports Server (NTRS)
Chien, L. S.; Sun, C. T.
1990-01-01
The inherent parallelism of the boundary element method is shown. The boundary element is formulated by assuming the linear variation of displacements and tractions within a line element. Moreover, MACSYMA symbolic program is employed to obtain the analytical results for influence coefficients. Three computational components are parallelized in this method to show the speedup and efficiency in computation. The global coefficient matrix is first formed concurrently. Then, the parallel Gaussian elimination solution scheme is applied to solve the resulting system of equations. Finally, and more importantly, the domain solutions of a given boundary value problem are calculated simultaneously. The linear speedups and high efficiencies are shown for solving a demonstrated problem on Sequent Symmetry S81 parallel computing system.
Parallel Algorithms for the Exascale Era
DOE Office of Scientific and Technical Information (OSTI.GOV)
Robey, Robert W.
New parallel algorithms are needed to reach the Exascale level of parallelism with millions of cores. We look at some of the research developed by students in projects at LANL. The research blends ideas from the early days of computing while weaving in the fresh approach brought by students new to the field of high performance computing. We look at reproducibility of global sums and why it is important to parallel computing. Next we look at how the concept of hashing has led to the development of more scalable algorithms suitable for next-generation parallel computers. Nearly all of this workmore » has been done by undergraduates and published in leading scientific journals.« less
NASA Technical Reports Server (NTRS)
Ortega, J. M.
1986-01-01
Various graduate research activities in the field of computer science are reported. Among the topics discussed are: (1) failure probabilities in multi-version software; (2) Gaussian Elimination on parallel computers; (3) three dimensional Poisson solvers on parallel/vector computers; (4) automated task decomposition for multiple robot arms; (5) multi-color incomplete cholesky conjugate gradient methods on the Cyber 205; and (6) parallel implementation of iterative methods for solving linear equations.
A high-speed linear algebra library with automatic parallelism
NASA Technical Reports Server (NTRS)
Boucher, Michael L.
1994-01-01
Parallel or distributed processing is key to getting highest performance workstations. However, designing and implementing efficient parallel algorithms is difficult and error-prone. It is even more difficult to write code that is both portable to and efficient on many different computers. Finally, it is harder still to satisfy the above requirements and include the reliability and ease of use required of commercial software intended for use in a production environment. As a result, the application of parallel processing technology to commercial software has been extremely small even though there are numerous computationally demanding programs that would significantly benefit from application of parallel processing. This paper describes DSSLIB, which is a library of subroutines that perform many of the time-consuming computations in engineering and scientific software. DSSLIB combines the high efficiency and speed of parallel computation with a serial programming model that eliminates many undesirable side-effects of typical parallel code. The result is a simple way to incorporate the power of parallel processing into commercial software without compromising maintainability, reliability, or ease of use. This gives significant advantages over less powerful non-parallel entries in the market.
NASA Astrophysics Data System (ADS)
Jiang, Yuning; Kang, Jinfeng; Wang, Xinan
2017-03-01
Resistive switching memory (RRAM) is considered as one of the most promising devices for parallel computing solutions that may overcome the von Neumann bottleneck of today’s electronic systems. However, the existing RRAM-based parallel computing architectures suffer from practical problems such as device variations and extra computing circuits. In this work, we propose a novel parallel computing architecture for pattern recognition by implementing k-nearest neighbor classification on metal-oxide RRAM crossbar arrays. Metal-oxide RRAM with gradual RESET behaviors is chosen as both the storage and computing components. The proposed architecture is tested by the MNIST database. High speed (~100 ns per example) and high recognition accuracy (97.05%) are obtained. The influence of several non-ideal device properties is also discussed, and it turns out that the proposed architecture shows great tolerance to device variations. This work paves a new way to achieve RRAM-based parallel computing hardware systems with high performance.
Symplectic molecular dynamics simulations on specially designed parallel computers.
Borstnik, Urban; Janezic, Dusanka
2005-01-01
We have developed a computer program for molecular dynamics (MD) simulation that implements the Split Integration Symplectic Method (SISM) and is designed to run on specialized parallel computers. The MD integration is performed by the SISM, which analytically treats high-frequency vibrational motion and thus enables the use of longer simulation time steps. The low-frequency motion is treated numerically on specially designed parallel computers, which decreases the computational time of each simulation time step. The combination of these approaches means that less time is required and fewer steps are needed and so enables fast MD simulations. We study the computational performance of MD simulation of molecular systems on specialized computers and provide a comparison to standard personal computers. The combination of the SISM with two specialized parallel computers is an effective way to increase the speed of MD simulations up to 16-fold over a single PC processor.
The current matrix elements from HAL QCD method
NASA Astrophysics Data System (ADS)
Watanabe, Kai; Ishii, Noriyoshi
2018-03-01
HAL QCD method is a method to construct a potential (HAL QCD potential) that reproduces the NN scattering phase shift faithful to the QCD. The HAL QCD potential is obtained from QCD by eliminating the degrees of freedom of quarks and gluons and leaving only two particular hadrons. Therefor, in the effective quantum mechanics of two nucleons defined by HAL QCD potential, the conserved current consists not only of the nucleon current but also an extra current originating from the potential (two-body current). Though the form of the two-body current is closely related to the potential, it is not straight forward to extract the former from the latter. In this work, we derive the the current matrix element formula in the quantum mechanics defined by the HAL QCD potential. As a first step, we focus on the non-relativistic case. To give an explicit example, we consider a second quantized non-relativistic two-channel coupling model which we refer to as the original model. From the original model, the HAL QCD potential for the open channel is constructed by eliminating the closed channel in the elastic two-particle scattering region. The current matrix element formula is derived by demanding the effective quantum mechanics defined by the HAL QCD potential to respond to the external field in the same way as the original two-channel coupling model.
Parallelization of fine-scale computation in Agile Multiscale Modelling Methodology
NASA Astrophysics Data System (ADS)
Macioł, Piotr; Michalik, Kazimierz
2016-10-01
Nowadays, multiscale modelling of material behavior is an extensively developed area. An important obstacle against its wide application is high computational demands. Among others, the parallelization of multiscale computations is a promising solution. Heterogeneous multiscale models are good candidates for parallelization, since communication between sub-models is limited. In this paper, the possibility of parallelization of multiscale models based on Agile Multiscale Methodology framework is discussed. A sequential, FEM based macroscopic model has been combined with concurrently computed fine-scale models, employing a MatCalc thermodynamic simulator. The main issues, being investigated in this work are: (i) the speed-up of multiscale models with special focus on fine-scale computations and (ii) on decreasing the quality of computations enforced by parallel execution. Speed-up has been evaluated on the basis of Amdahl's law equations. The problem of `delay error', rising from the parallel execution of fine scale sub-models, controlled by the sequential macroscopic sub-model is discussed. Some technical aspects of combining third-party commercial modelling software with an in-house multiscale framework and a MPI library are also discussed.
Parallel algorithms for mapping pipelined and parallel computations
NASA Technical Reports Server (NTRS)
Nicol, David M.
1988-01-01
Many computational problems in image processing, signal processing, and scientific computing are naturally structured for either pipelined or parallel computation. When mapping such problems onto a parallel architecture it is often necessary to aggregate an obvious problem decomposition. Even in this context the general mapping problem is known to be computationally intractable, but recent advances have been made in identifying classes of problems and architectures for which optimal solutions can be found in polynomial time. Among these, the mapping of pipelined or parallel computations onto linear array, shared memory, and host-satellite systems figures prominently. This paper extends that work first by showing how to improve existing serial mapping algorithms. These improvements have significantly lower time and space complexities: in one case a published O(nm sup 3) time algorithm for mapping m modules onto n processors is reduced to an O(nm log m) time complexity, and its space requirements reduced from O(nm sup 2) to O(m). Run time complexity is further reduced with parallel mapping algorithms based on these improvements, which run on the architecture for which they create the mappings.
Superconformal Algebraic Approach to Hadron Structure
DOE Office of Scientific and Technical Information (OSTI.GOV)
de Teramond, Guy F.; Brodsky, Stanley J.; Deur, Alexandre
2017-03-01
Fundamental aspects of nonperturbative QCD dynamics which are not obvious from its classical Lagrangian, such as the emergence of a mass scale and confinement, the existence of a zero mass bound state, the appearance of universal Regge trajectories and the breaking of chiral symmetry are incorporated from the onset in an effective theory based on superconformal quantum mechanics and its embedding in a higher dimensional gravitational theory. In addition, superconformal quantum mechanics gives remarkable connections between the light meson and nucleon spectra. This new approach to hadron physics is also suitable to describe nonperturbative QCD observables based on structure functions,more » such as GPDs, which are not amenable to a first-principle computation. The formalism is also successful in the description of form factors, the nonperturbative behavior of the strong coupling and diffractive processes. We also discuss in this article how the framework can be extended rather successfully to the heavy-light hadron sector.« less
Constraints on the [Formula: see text] form factor from analyticity and unitarity.
Ananthanarayan, B; Caprini, I; Kubis, B
Motivated by the discrepancies noted recently between the theoretical calculations of the electromagnetic [Formula: see text] form factor and certain experimental data, we investigate this form factor using analyticity and unitarity in a framework known as the method of unitarity bounds. We use a QCD correlator computed on the spacelike axis by operator product expansion and perturbative QCD as input, and exploit unitarity and the positivity of its spectral function, including the two-pion contribution that can be reliably calculated using high-precision data on the pion form factor. From this information, we derive upper and lower bounds on the modulus of the [Formula: see text] form factor in the elastic region. The results provide a significant check on those obtained with standard dispersion relations, confirming the existence of a disagreement with experimental data in the region around [Formula: see text].
Single Top Production at Next-to-Leading Order in the Standard Model Effective Field Theory.
Zhang, Cen
2016-04-22
Single top production processes at hadron colliders provide information on the relation between the top quark and the electroweak sector of the standard model. We compute the next-to-leading order QCD corrections to the three main production channels: t-channel, s-channel, and tW associated production, in the standard model including operators up to dimension six. The calculation can be matched to parton shower programs and can therefore be directly used in experimental analyses. The QCD corrections are found to significantly impact the extraction of the current limits on the operators, because both of an improved accuracy and a better precision of the theoretical predictions. In addition, the distributions of some of the key discriminating observables are modified in a nontrivial way, which could change the interpretation of measurements in terms of UV complete models.
Electromagnetic corrections to the hadronic vacuum polarization of the photon within QEDL and QEDM
NASA Astrophysics Data System (ADS)
Bussone, Andrea; Della Morte, Michele; Janowski, Tadeusz
2018-03-01
We compute the leading QED corrections to the hadronic vacuum polarization (HVP) of the photon, relevant for the determination of leptonic anomalous magnetic moments, al. We work in the electroquenched approximation and use dynamical QCD configurations generated by the CLS initiative with two degenerate flavors of nonperturbatively O(a)-improved Wilson fermions. We consider QEDL and QEDM to deal with the finite-volume zero modes. We compare results for the Wilson loops with exact analytical determinations. In addition we make sure that the volumes and photon masses used in QEDM are such that the correct dispersion relation is reproduced by the energy levels extracted from the charged pions two-point functions. Finally we compare results for pion masses and the HVP between QEDL and QEDM. For the vacuum polarization, corrections with respect to the pure QCD case, at fixed pion masses, turn out to be at the percent level.
Multi-boson block factorization of fermions
NASA Astrophysics Data System (ADS)
Giusti, Leonardo; Cè, Marco; Schaefer, Stefan
2018-03-01
The numerical computations of many quantities of theoretical and phenomenological interest are plagued by statistical errors which increase exponentially with the distance of the sources in the relevant correlators. Notable examples are baryon masses and matrix elements, the hadronic vacuum polarization and the light-by-light scattering contributions to the muon g - 2, and the form factors of semileptonic B decays. Reliable and precise determinations of these quantities are very difficult if not impractical with state-of-the-art standard Monte Carlo integration schemes. I will review a recent proposal for factorizing the fermion determinant in lattice QCD that leads to a local action in the gauge field and in the auxiliary boson fields. Once combined with the corresponding factorization of the quark propagator, it paves the way for multi-level Monte Carlo integration in the presence of fermions opening new perspectives in lattice QCD. Exploratory results on the impact on the above mentioned observables will be presented.
ππ P-wave resonant scattering from lattice QCD
NASA Astrophysics Data System (ADS)
Paul, Srijit; Alexandrou, Constantia; Leskovec, Luka; Meinel, Stefan; Negele, John W.; Petschlies, Marcus; Pochinsky, Andrew; Rendon Suzuki, Jesus Gumaro; Syritsyn, Sergey
2018-03-01
We present a high-statistics analysis of the ρ resonance in ππ scattering, using 2 + 1 flavors of clover fermions at a pion mass of approximately 320 MeV and a lattice size of approximately 3:6 fm. The computation of the two-point functions are carried out using combinations of forward, sequential, and stochastic propagators. For the extraction of the ρ-resonance parameters, we compare different fit methods and demonstrate their consistency. For the ππ scattering phase shift, we consider different Breit-Wigner parametrizations and also investigate possible nonresonant contributions. We find that the minimal Breit-Wigner model is suffcient to describe our data, and obtain amρ = 0:4609(16)stat(14)sys and gρππ = 5:69(13)stat(16)sys. In our comparison with other lattice QCD results, we consider the dimensionless ratios amρ/amN and amπ/amN to avoid scale setting ambiguities.
Isospin splittings in the light-baryon octet from lattice QCD and QED.
Borsanyi, Sz; Dürr, S; Fodor, Z; Frison, J; Hoelbling, C; Katz, S D; Krieg, S; Kurth, Th; Lellouch, L; Lippert, Th; Portelli, A; Ramos, A; Sastre, A; Szabo, K
2013-12-20
While electromagnetic and up-down quark mass difference effects on octet baryon masses are very small, they have important consequences. The stability of the hydrogen atom against beta decay is a prominent example. Here, we include these effects by adding them to valence quarks in a lattice QCD calculation based on Nf=2+1 simulations with five lattice spacings down to 0.054 fm, lattice sizes up to 6 fm, and average up-down quark masses all the way down to their physical value. This allows us to gain control over all systematic errors, except for the one associated with neglecting electromagnetism in the sea. We compute the octet baryon isomultiplet mass splittings, as well as the individual contributions from electromagnetism and the up-down quark mass difference. Our results for the total splittings are in good agreement with experiment.
Coupled π π , K K ¯ scattering in P -wave and the ρ resonance from lattice QCD
Wilson, David J.; Briceño, Raúl A.; Dudek, Jozef J.; ...
2015-11-02
In this study, we determine elastic and coupled-channel amplitudes for isospin-1 meson-meson scattering inmore » $P$-wave, by calculating correlation functions using lattice QCD with light quark masses such that $$m_\\pi = 236$$ MeV in a cubic volume of $$\\sim (4 \\,\\mathrm{fm})^3$$. Variational analyses of large matrices of correlation functions computed using operator constructions resembling $$\\pi\\pi$$, $$K\\overline{K}$$ and $$q\\bar{q}$$, in several moving frames and several lattice irreducible representations, leads to discrete energy spectra from which scattering amplitudes are extracted. In the elastic $$\\pi\\pi$$ scattering region we obtain a detailed energy-dependence for the phase-shift, corresponding to a $$\\rho$$ resonance, and we extend the analysis into the coupled-channel $$K\\overline{K}$$ region for the first time, finding a small coupling between the channels.« less
Parton distribution functions with QED corrections in the valon model
NASA Astrophysics Data System (ADS)
Mottaghizadeh, Marzieh; Taghavi Shahri, Fatemeh; Eslami, Parvin
2017-10-01
The parton distribution functions (PDFs) with QED corrections are obtained by solving the QCD ⊗QED DGLAP evolution equations in the framework of the "valon" model at the next-to-leading-order QCD and the leading-order QED approximations. Our results for the PDFs with QED corrections in this phenomenological model are in good agreement with the newly related CT14QED global fits code [Phys. Rev. D 93, 114015 (2016), 10.1103/PhysRevD.93.114015] and APFEL (NNPDF2.3QED) program [Comput. Phys. Commun. 185, 1647 (2014), 10.1016/j.cpc.2014.03.007] in a wide range of x =[10-5,1 ] and Q2=[0.283 ,108] GeV2 . The model calculations agree rather well with those codes. In the latter, we proposed a new method for studying the symmetry breaking of the sea quark distribution functions inside the proton.
Impressions of the Meson Spectrum: Hybrids & Exotics, present and future
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pennington, Michael R.
2016-03-25
It has long been expected that the spectrum of hadrons in QCD would be far richer and extensive than experiment has so far revealed. While there have been experimental hints of this richness for some time, it is really only in the last few years that dramatic progress has been seen in the exploration both experimentally and in calculations on the lattice. Precision studies enabled by new technology both with detectors and high performance computations are converging on an understanding of the spectrum in strong coupling QCD. These methodologies are laying the foundation for a decade of potential discovery thatmore » electro and photoproduction experiments at Jefferson Lab, which when combined with key results on B and charmonium decays from both e+e? and pp colliders, should turn mere impressions of the light meson spectrum into a high definition picture.« less
Isoscalar π π Scattering and the σ Meson Resonance from QCD
Briceño, Raul A.; Dudek, Jozef J.; Edwards, Robert G.; ...
2017-01-09
Here, we present for the first time a determination of the energy dependence of the isoscalar ππ elastic scattering phase-shift within a first-principles numerical lattice approach to QCD. We also compute the hadronic correlation functions including all required quark propagation diagrams. From these the discrete spectrum of states in the finite volume defined by the lattice boundary is extracted. From the volume dependence of the spectrum we obtain the S-wave phase-shift up to the Kmore » $$\\bar{K}$$ threshold. Calculations are performed at two values of the u, d quark mass corresponding to m π = 236, 391 MeV and the resulting amplitudes are described in terms of a σ meson which evolves from a bound-state below ππ threshold at the heavier quark mass, to a broad resonance at the lighter quark mass.« less
Bottomonium suppression using a lattice QCD vetted potential
NASA Astrophysics Data System (ADS)
Krouppa, Brandon; Rothkopf, Alexander; Strickland, Michael
2018-01-01
We estimate bottomonium yields in relativistic heavy-ion collisions using a lattice QCD vetted, complex-valued, heavy-quark potential embedded in a realistic, hydrodynamically evolving medium background. We find that the lattice-vetted functional form and temperature dependence of the proper heavy-quark potential dramatically reduces the dependence of the yields on parameters other than the temperature evolution, strengthening the picture of bottomonium as QGP thermometer. Our results also show improved agreement between computed yields and experimental data produced in RHIC 200 GeV /nucleon collisions. For LHC 2.76 TeV /nucleon collisions, the excited states, whose suppression has been used as a vital sign for quark-gluon-plasma production in a heavy-ion collision, are reproduced better than previous perturbatively-motivated potential models; however, at the highest LHC energies our estimates for bottomonium suppression begin to underestimate the data. Possible paths to remedy this situation are discussed.
Fragmentation contributions to J / ψ photoproduction at HERA
Bodwin, Geoffrey T.; Chung, Hee Sok; Kim, U-Rae; ...
2015-10-28
Here, we compute leading-power fragmentation corrections to J/ψ photoproduction at DESY HERA, making use of the nonrelativistic QCD factorization approach. Our calculations include parton production cross sections through order α 3 s, fragmentation functions though order α 2 s, and leading logarithms of the transverse momentum divided by the charm-quark mass to all orders in α s. We find that the leading-power fragmentation corrections, beyond those that are included through next-to-leading order in α s, are small relative to the fixed-order contributions through next-to-leading order in α s. Consequently, an important discrepancy remains between the experimental measurements of the J/ψmore » photoproduction cross section and predictions that make use of nonrelativistic-QCD long-distance matrix elements that are extracted from the J/ψ hadroproduction cross-section and polarization data.« less
Isospin Splittings in the Light-Baryon Octet from Lattice QCD and QED
NASA Astrophysics Data System (ADS)
Borsanyi, Sz.; Dürr, S.; Fodor, Z.; Frison, J.; Hoelbling, C.; Katz, S. D.; Krieg, S.; Kurth, Th.; Lellouch, L.; Lippert, Th.; Portelli, A.; Ramos, A.; Sastre, A.; Szabo, K.; Budapest-Marseille-Wuppertal Collaboration
2013-12-01
While electromagnetic and up-down quark mass difference effects on octet baryon masses are very small, they have important consequences. The stability of the hydrogen atom against beta decay is a prominent example. Here, we include these effects by adding them to valence quarks in a lattice QCD calculation based on Nf=2+1 simulations with five lattice spacings down to 0.054 fm, lattice sizes up to 6 fm, and average up-down quark masses all the way down to their physical value. This allows us to gain control over all systematic errors, except for the one associated with neglecting electromagnetism in the sea. We compute the octet baryon isomultiplet mass splittings, as well as the individual contributions from electromagnetism and the up-down quark mass difference. Our results for the total splittings are in good agreement with experiment.
Synthesizing parallel imaging applications using the CAP (computer-aided parallelization) tool
NASA Astrophysics Data System (ADS)
Gennart, Benoit A.; Mazzariol, Marc; Messerli, Vincent; Hersch, Roger D.
1997-12-01
Imaging applications such as filtering, image transforms and compression/decompression require vast amounts of computing power when applied to large data sets. These applications would potentially benefit from the use of parallel processing. However, dedicated parallel computers are expensive and their processing power per node lags behind that of the most recent commodity components. Furthermore, developing parallel applications remains a difficult task: writing and debugging the application is difficult (deadlocks), programs may not be portable from one parallel architecture to the other, and performance often comes short of expectations. In order to facilitate the development of parallel applications, we propose the CAP computer-aided parallelization tool which enables application programmers to specify at a high-level of abstraction the flow of data between pipelined-parallel operations. In addition, the CAP tool supports the programmer in developing parallel imaging and storage operations. CAP enables combining efficiently parallel storage access routines and image processing sequential operations. This paper shows how processing and I/O intensive imaging applications must be implemented to take advantage of parallelism and pipelining between data access and processing. This paper's contribution is (1) to show how such implementations can be compactly specified in CAP, and (2) to demonstrate that CAP specified applications achieve the performance of custom parallel code. The paper analyzes theoretically the performance of CAP specified applications and demonstrates the accuracy of the theoretical analysis through experimental measurements.
Magnetic bion condensation: A new mechanism of confinement and mass gap in four dimensions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Uensal, Mithat
In recent work, we derived the long-distance confining dynamics of certain QCD-like gauge theories formulated on small S{sup 1}xR{sup 3} based on symmetries, an index theorem, and Abelian duality. Here, we give the microscopic derivation. The solution reveals a new mechanism of confinement in QCD(adj) in the regime where we have control over both perturbative and nonperturbative aspects. In particular, consider SU(2) QCD(adj) theory with 1{<=}n{sub f}{<=}4 Majorana fermions, a theory which undergoes gauge symmetry breaking at small S{sup 1}. If the magnetic charge of the Bogomol'nyi-Prasad-Sommerfield (BPS) monopole is normalized to unity, we show that confinement occurs due tomore » condensation of objects with magnetic charge 2, not 1. Because of index theorems, we know that such an object cannot be a two identical monopole configuration. Its net topological charge must vanish, and hence it must be topologically indistinguishable from the perturbative vacuum. We construct such non-self-dual topological excitations, the magnetically charged, topologically null molecules of a BPS monopole and KK antimonopole, which we refer to as magnetic bions. An immediate puzzle with this proposal is the apparent Coulomb repulsion between the BPS-KK pair. An attraction which overcomes the Coulomb repulsion between the two is induced by 2n{sub f}-fermion exchange. Bion condensation is also the mechanism of confinement in N=1 SYM on the same four-manifold. The SU(N) generalization hints a possible hidden integrability behind nonsupersymmetric QCD of affine Toda type, and allows us to analytically compute the mass gap in the gauge sector. We currently do not know the extension to R{sup 4}.« less
CSM parallel structural methods research
NASA Technical Reports Server (NTRS)
Storaasli, Olaf O.
1989-01-01
Parallel structural methods, research team activities, advanced architecture computers for parallel computational structural mechanics (CSM) research, the FLEX/32 multicomputer, a parallel structural analyses testbed, blade-stiffened aluminum panel with a circular cutout and the dynamic characteristics of a 60 meter, 54-bay, 3-longeron deployable truss beam are among the topics discussed.
Parallelized direct execution simulation of message-passing parallel programs
NASA Technical Reports Server (NTRS)
Dickens, Phillip M.; Heidelberger, Philip; Nicol, David M.
1994-01-01
As massively parallel computers proliferate, there is growing interest in findings ways by which performance of massively parallel codes can be efficiently predicted. This problem arises in diverse contexts such as parallelizing computers, parallel performance monitoring, and parallel algorithm development. In this paper we describe one solution where one directly executes the application code, but uses a discrete-event simulator to model details of the presumed parallel machine such as operating system and communication network behavior. Because this approach is computationally expensive, we are interested in its own parallelization specifically the parallelization of the discrete-event simulator. We describe methods suitable for parallelized direct execution simulation of message-passing parallel programs, and report on the performance of such a system, Large Application Parallel Simulation Environment (LAPSE), we have built on the Intel Paragon. On all codes measured to date, LAPSE predicts performance well typically within 10 percent relative error. Depending on the nature of the application code, we have observed low slowdowns (relative to natively executing code) and high relative speedups using up to 64 processors.
Automatic Generation of Directive-Based Parallel Programs for Shared Memory Parallel Systems
NASA Technical Reports Server (NTRS)
Jin, Hao-Qiang; Yan, Jerry; Frumkin, Michael
2000-01-01
The shared-memory programming model is a very effective way to achieve parallelism on shared memory parallel computers. As great progress was made in hardware and software technologies, performance of parallel programs with compiler directives has demonstrated large improvement. The introduction of OpenMP directives, the industrial standard for shared-memory programming, has minimized the issue of portability. Due to its ease of programming and its good performance, the technique has become very popular. In this study, we have extended CAPTools, a computer-aided parallelization toolkit, to automatically generate directive-based, OpenMP, parallel programs. We outline techniques used in the implementation of the tool and present test results on the NAS parallel benchmarks and ARC3D, a CFD application. This work demonstrates the great potential of using computer-aided tools to quickly port parallel programs and also achieve good performance.
The science of computing - Parallel computation
NASA Technical Reports Server (NTRS)
Denning, P. J.
1985-01-01
Although parallel computation architectures have been known for computers since the 1920s, it was only in the 1970s that microelectronic components technologies advanced to the point where it became feasible to incorporate multiple processors in one machine. Concommitantly, the development of algorithms for parallel processing also lagged due to hardware limitations. The speed of computing with solid-state chips is limited by gate switching delays. The physical limit implies that a 1 Gflop operational speed is the maximum for sequential processors. A computer recently introduced features a 'hypercube' architecture with 128 processors connected in networks at 5, 6 or 7 points per grid, depending on the design choice. Its computing speed rivals that of supercomputers, but at a fraction of the cost. The added speed with less hardware is due to parallel processing, which utilizes algorithms representing different parts of an equation that can be broken into simpler statements and processed simultaneously. Present, highly developed computer languages like FORTRAN, PASCAL, COBOL, etc., rely on sequential instructions. Thus, increased emphasis will now be directed at parallel processing algorithms to exploit the new architectures.
Lattice analysis for the energy scale of QCD phenomena.
Yamamoto, Arata; Suganuma, Hideo
2008-12-12
We formulate a new framework in lattice QCD to study the relevant energy scale of QCD phenomena. By considering the Fourier transformation of link variable, we can investigate the intrinsic energy scale of a physical quantity nonperturbatively. This framework is broadly available for all lattice QCD calculations. We apply this framework for the quark-antiquark potential and meson masses in quenched lattice QCD. The gluonic energy scale relevant for the confinement is found to be less than 1 GeV in the Landau or Coulomb gauge.
QCDNUM: Fast QCD evolution and convolution
NASA Astrophysics Data System (ADS)
Botje, M.
2011-02-01
The QCDNUM program numerically solves the evolution equations for parton densities and fragmentation functions in perturbative QCD. Un-polarised parton densities can be evolved up to next-to-next-to-leading order in powers of the strong coupling constant, while polarised densities or fragmentation functions can be evolved up to next-to-leading order. Other types of evolution can be accessed by feeding alternative sets of evolution kernels into the program. A versatile convolution engine provides tools to compute parton luminosities, cross-sections in hadron-hadron scattering, and deep inelastic structure functions in the zero-mass scheme or in generalised mass schemes. Input to these calculations are either the QCDNUM evolved densities, or those read in from an external parton density repository. Included in the software distribution are packages to calculate zero-mass structure functions in un-polarised deep inelastic scattering, and heavy flavour contributions to these structure functions in the fixed flavour number scheme. Program summaryProgram title: QCDNUM version: 17.00 Catalogue identifier: AEHV_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEHV_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GNU Public Licence No. of lines in distributed program, including test data, etc.: 45 736 No. of bytes in distributed program, including test data, etc.: 911 569 Distribution format: tar.gz Programming language: Fortran-77 Computer: All Operating system: All RAM: Typically 3 Mbytes Classification: 11.5 Nature of problem: Evolution of the strong coupling constant and parton densities, up to next-to-next-to-leading order in perturbative QCD. Computation of observable quantities by Mellin convolution of the evolved densities with partonic cross-sections. Solution method: Parametrisation of the parton densities as linear or quadratic splines on a discrete grid, and evolution of the spline coefficients by solving (coupled) triangular matrix equations with a forward substitution algorithm. Fast computation of convolution integrals as weighted sums of spline coefficients, with weights derived from user-given convolution kernels. Restrictions: Accuracy and speed are determined by the density of the evolution grid. Running time: Less than 10 ms on a 2 GHz Intel Core 2 Duo processor to evolve the gluon density and 12 quark densities at next-to-next-to-leading order over a large kinematic range.
Parallel-Processing Test Bed For Simulation Software
NASA Technical Reports Server (NTRS)
Blech, Richard; Cole, Gary; Townsend, Scott
1996-01-01
Second-generation Hypercluster computing system is multiprocessor test bed for research on parallel algorithms for simulation in fluid dynamics, electromagnetics, chemistry, and other fields with large computational requirements but relatively low input/output requirements. Built from standard, off-shelf hardware readily upgraded as improved technology becomes available. System used for experiments with such parallel-processing concepts as message-passing algorithms, debugging software tools, and computational steering. First-generation Hypercluster system described in "Hypercluster Parallel Processor" (LEW-15283).
System-wide power management control via clock distribution network
Coteus, Paul W.; Gara, Alan; Gooding, Thomas M.; Haring, Rudolf A.; Kopcsay, Gerard V.; Liebsch, Thomas A.; Reed, Don D.
2015-05-19
An apparatus, method and computer program product for automatically controlling power dissipation of a parallel computing system that includes a plurality of processors. A computing device issues a command to the parallel computing system. A clock pulse-width modulator encodes the command in a system clock signal to be distributed to the plurality of processors. The plurality of processors in the parallel computing system receive the system clock signal including the encoded command, and adjusts power dissipation according to the encoded command.
Parallel Computing:. Some Activities in High Energy Physics
NASA Astrophysics Data System (ADS)
Willers, Ian
This paper examines some activities in High Energy Physics that utilise parallel computing. The topic includes all computing from the proposed SIMD front end detectors, the farming applications, high-powered RISC processors and the large machines in the computer centers. We start by looking at the motivation behind using parallelism for general purpose computing. The developments around farming are then described from its simplest form to the more complex system in Fermilab. Finally, there is a list of some developments that are happening close to the experiments.
Implementation of DFT application on ternary optical computer
NASA Astrophysics Data System (ADS)
Junjie, Peng; Youyi, Fu; Xiaofeng, Zhang; Shuai, Kong; Xinyu, Wei
2018-03-01
As its characteristics of huge number of data bits and low energy consumption, optical computing may be used in the applications such as DFT etc. which needs a lot of computation and can be implemented in parallel. According to this, DFT implementation methods in full parallel as well as in partial parallel are presented. Based on resources ternary optical computer (TOC), extensive experiments were carried out. Experimental results show that the proposed schemes are correct and feasible. They provide a foundation for further exploration of the applications on TOC that needs a large amount calculation and can be processed in parallel.
QCD dirac operator at nonzero chemical potential: lattice data and matrix model.
Akemann, Gernot; Wettig, Tilo
2004-03-12
Recently, a non-Hermitian chiral random matrix model was proposed to describe the eigenvalues of the QCD Dirac operator at nonzero chemical potential. This matrix model can be constructed from QCD by mapping it to an equivalent matrix model which has the same symmetries as QCD with chemical potential. Its microscopic spectral correlations are conjectured to be identical to those of the QCD Dirac operator. We investigate this conjecture by comparing large ensembles of Dirac eigenvalues in quenched SU(3) lattice QCD at a nonzero chemical potential to the analytical predictions of the matrix model. Excellent agreement is found in the two regimes of weak and strong non-Hermiticity, for several different lattice volumes.
NASA Technical Reports Server (NTRS)
Fijany, Amir (Inventor); Bejczy, Antal K. (Inventor)
1993-01-01
This is a real-time robotic controller and simulator which is a MIMD-SIMD parallel architecture for interfacing with an external host computer and providing a high degree of parallelism in computations for robotic control and simulation. It includes a host processor for receiving instructions from the external host computer and for transmitting answers to the external host computer. There are a plurality of SIMD microprocessors, each SIMD processor being a SIMD parallel processor capable of exploiting fine grain parallelism and further being able to operate asynchronously to form a MIMD architecture. Each SIMD processor comprises a SIMD architecture capable of performing two matrix-vector operations in parallel while fully exploiting parallelism in each operation. There is a system bus connecting the host processor to the plurality of SIMD microprocessors and a common clock providing a continuous sequence of clock pulses. There is also a ring structure interconnecting the plurality of SIMD microprocessors and connected to the clock for providing the clock pulses to the SIMD microprocessors and for providing a path for the flow of data and instructions between the SIMD microprocessors. The host processor includes logic for controlling the RRCS by interpreting instructions sent by the external host computer, decomposing the instructions into a series of computations to be performed by the SIMD microprocessors, using the system bus to distribute associated data among the SIMD microprocessors, and initiating activity of the SIMD microprocessors to perform the computations on the data by procedure call.
Exact parallel algorithms for some members of the traveling salesman problem family
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pekny, J.F.
1989-01-01
The traveling salesman problem and its many generalizations comprise one of the best known combinatorial optimization problem families. Most members of the family are NP-complete problems so that exact algorithms require an unpredictable and sometimes large computational effort. Parallel computers offer hope for providing the power required to meet these demands. A major barrier to applying parallel computers is the lack of parallel algorithms. The contributions presented in this thesis center around new exact parallel algorithms for the asymmetric traveling salesman problem (ATSP), prize collecting traveling salesman problem (PCTSP), and resource constrained traveling salesman problem (RCTSP). The RCTSP is amore » particularly difficult member of the family since finding a feasible solution is an NP-complete problem. An exact sequential algorithm is also presented for the directed hamiltonian cycle problem (DHCP). The DHCP algorithm is superior to current heuristic approaches and represents the first exact method applicable to large graphs. Computational results presented for each of the algorithms demonstrates the effectiveness of combining efficient algorithms with parallel computing methods. Performance statistics are reported for randomly generated ATSPs with 7,500 cities, PCTSPs with 200 cities, RCTSPs with 200 cities, DHCPs with 3,500 vertices, and assignment problems of size 10,000. Sequential results were collected on a Sun 4/260 engineering workstation, while parallel results were collected using a 14 and 100 processor BBN Butterfly Plus computer. The computational results represent the largest instances ever solved to optimality on any type of computer.« less
Use of parallel computing in mass processing of laser data
NASA Astrophysics Data System (ADS)
Będkowski, J.; Bratuś, R.; Prochaska, M.; Rzonca, A.
2015-12-01
The first part of the paper includes a description of the rules used to generate the algorithm needed for the purpose of parallel computing and also discusses the origins of the idea of research on the use of graphics processors in large scale processing of laser scanning data. The next part of the paper includes the results of an efficiency assessment performed for an array of different processing options, all of which were substantially accelerated with parallel computing. The processing options were divided into the generation of orthophotos using point clouds, coloring of point clouds, transformations, and the generation of a regular grid, as well as advanced processes such as the detection of planes and edges, point cloud classification, and the analysis of data for the purpose of quality control. Most algorithms had to be formulated from scratch in the context of the requirements of parallel computing. A few of the algorithms were based on existing technology developed by the Dephos Software Company and then adapted to parallel computing in the course of this research study. Processing time was determined for each process employed for a typical quantity of data processed, which helped confirm the high efficiency of the solutions proposed and the applicability of parallel computing to the processing of laser scanning data. The high efficiency of parallel computing yields new opportunities in the creation and organization of processing methods for laser scanning data.
Shen, Wenfeng; Wei, Daming; Xu, Weimin; Zhu, Xin; Yuan, Shizhong
2010-10-01
Biological computations like electrocardiological modelling and simulation usually require high-performance computing environments. This paper introduces an implementation of parallel computation for computer simulation of electrocardiograms (ECGs) in a personal computer environment with an Intel CPU of Core (TM) 2 Quad Q6600 and a GPU of Geforce 8800GT, with software support by OpenMP and CUDA. It was tested in three parallelization device setups: (a) a four-core CPU without a general-purpose GPU, (b) a general-purpose GPU plus 1 core of CPU, and (c) a four-core CPU plus a general-purpose GPU. To effectively take advantage of a multi-core CPU and a general-purpose GPU, an algorithm based on load-prediction dynamic scheduling was developed and applied to setting (c). In the simulation with 1600 time steps, the speedup of the parallel computation as compared to the serial computation was 3.9 in setting (a), 16.8 in setting (b), and 20.0 in setting (c). This study demonstrates that a current PC with a multi-core CPU and a general-purpose GPU provides a good environment for parallel computations in biological modelling and simulation studies. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.
Creating a Parallel Version of VisIt for Microsoft Windows
DOE Office of Scientific and Technical Information (OSTI.GOV)
Whitlock, B J; Biagas, K S; Rawson, P L
2011-12-07
VisIt is a popular, free interactive parallel visualization and analysis tool for scientific data. Users can quickly generate visualizations from their data, animate them through time, manipulate them, and save the resulting images or movies for presentations. VisIt was designed from the ground up to work on many scales of computers from modest desktops up to massively parallel clusters. VisIt is comprised of a set of cooperating programs. All programs can be run locally or in client/server mode in which some run locally and some run remotely on compute clusters. The VisIt program most able to harness today's computing powermore » is the VisIt compute engine. The compute engine is responsible for reading simulation data from disk, processing it, and sending results or images back to the VisIt viewer program. In a parallel environment, the compute engine runs several processes, coordinating using the Message Passing Interface (MPI) library. Each MPI process reads some subset of the scientific data and filters the data in various ways to create useful visualizations. By using MPI, VisIt has been able to scale well into the thousands of processors on large computers such as dawn and graph at LLNL. The advent of multicore CPU's has made parallelism the 'new' way to achieve increasing performance. With today's computers having at least 2 cores and in many cases up to 8 and beyond, it is more important than ever to deploy parallel software that can use that computing power not only on clusters but also on the desktop. We have created a parallel version of VisIt for Windows that uses Microsoft's MPI implementation (MSMPI) to process data in parallel on the Windows desktop as well as on a Windows HPC cluster running Microsoft Windows Server 2008. Initial desktop parallel support for Windows was deployed in VisIt 2.4.0. Windows HPC cluster support has been completed and will appear in the VisIt 2.5.0 release. We plan to continue supporting parallel VisIt on Windows so our users will be able to take full advantage of their multicore resources.« less
HeNCE: A Heterogeneous Network Computing Environment
Beguelin, Adam; Dongarra, Jack J.; Geist, George Al; ...
1994-01-01
Network computing seeks to utilize the aggregate resources of many networked computers to solve a single problem. In so doing it is often possible to obtain supercomputer performance from an inexpensive local area network. The drawback is that network computing is complicated and error prone when done by hand, especially if the computers have different operating systems and data formats and are thus heterogeneous. The heterogeneous network computing environment (HeNCE) is an integrated graphical environment for creating and running parallel programs over a heterogeneous collection of computers. It is built on a lower level package called parallel virtual machine (PVM).more » The HeNCE philosophy of parallel programming is to have the programmer graphically specify the parallelism of a computation and to automate, as much as possible, the tasks of writing, compiling, executing, debugging, and tracing the network computation. Key to HeNCE is a graphical language based on directed graphs that describe the parallelism and data dependencies of an application. Nodes in the graphs represent conventional Fortran or C subroutines and the arcs represent data and control flow. This article describes the present state of HeNCE, its capabilities, limitations, and areas of future research.« less
Tankam, Patrice; Santhanam, Anand P.; Lee, Kye-Sung; Won, Jungeun; Canavesi, Cristina; Rolland, Jannick P.
2014-01-01
Abstract. Gabor-domain optical coherence microscopy (GD-OCM) is a volumetric high-resolution technique capable of acquiring three-dimensional (3-D) skin images with histological resolution. Real-time image processing is needed to enable GD-OCM imaging in a clinical setting. We present a parallelized and scalable multi-graphics processing unit (GPU) computing framework for real-time GD-OCM image processing. A parallelized control mechanism was developed to individually assign computation tasks to each of the GPUs. For each GPU, the optimal number of amplitude-scans (A-scans) to be processed in parallel was selected to maximize GPU memory usage and core throughput. We investigated five computing architectures for computational speed-up in processing 1000×1000 A-scans. The proposed parallelized multi-GPU computing framework enables processing at a computational speed faster than the GD-OCM image acquisition, thereby facilitating high-speed GD-OCM imaging in a clinical setting. Using two parallelized GPUs, the image processing of a 1×1×0.6 mm3 skin sample was performed in about 13 s, and the performance was benchmarked at 6.5 s with four GPUs. This work thus demonstrates that 3-D GD-OCM data may be displayed in real-time to the examiner using parallelized GPU processing. PMID:24695868
Tankam, Patrice; Santhanam, Anand P; Lee, Kye-Sung; Won, Jungeun; Canavesi, Cristina; Rolland, Jannick P
2014-07-01
Gabor-domain optical coherence microscopy (GD-OCM) is a volumetric high-resolution technique capable of acquiring three-dimensional (3-D) skin images with histological resolution. Real-time image processing is needed to enable GD-OCM imaging in a clinical setting. We present a parallelized and scalable multi-graphics processing unit (GPU) computing framework for real-time GD-OCM image processing. A parallelized control mechanism was developed to individually assign computation tasks to each of the GPUs. For each GPU, the optimal number of amplitude-scans (A-scans) to be processed in parallel was selected to maximize GPU memory usage and core throughput. We investigated five computing architectures for computational speed-up in processing 1000×1000 A-scans. The proposed parallelized multi-GPU computing framework enables processing at a computational speed faster than the GD-OCM image acquisition, thereby facilitating high-speed GD-OCM imaging in a clinical setting. Using two parallelized GPUs, the image processing of a 1×1×0.6 mm3 skin sample was performed in about 13 s, and the performance was benchmarked at 6.5 s with four GPUs. This work thus demonstrates that 3-D GD-OCM data may be displayed in real-time to the examiner using parallelized GPU processing.
Seeing the forest for the trees: Networked workstations as a parallel processing computer
NASA Technical Reports Server (NTRS)
Breen, J. O.; Meleedy, D. M.
1992-01-01
Unlike traditional 'serial' processing computers in which one central processing unit performs one instruction at a time, parallel processing computers contain several processing units, thereby, performing several instructions at once. Many of today's fastest supercomputers achieve their speed by employing thousands of processing elements working in parallel. Few institutions can afford these state-of-the-art parallel processors, but many already have the makings of a modest parallel processing system. Workstations on existing high-speed networks can be harnessed as nodes in a parallel processing environment, bringing the benefits of parallel processing to many. While such a system can not rival the industry's latest machines, many common tasks can be accelerated greatly by spreading the processing burden and exploiting idle network resources. We study several aspects of this approach, from algorithms to select nodes to speed gains in specific tasks. With ever-increasing volumes of astronomical data, it becomes all the more necessary to utilize our computing resources fully.
Six Years of Parallel Computing at NAS (1987 - 1993): What Have we Learned?
NASA Technical Reports Server (NTRS)
Simon, Horst D.; Cooper, D. M. (Technical Monitor)
1994-01-01
In the fall of 1987 the age of parallelism at NAS began with the installation of a 32K processor CM-2 from Thinking Machines. In 1987 this was described as an "experiment" in parallel processing. In the six years since, NAS acquired a series of parallel machines, and conducted an active research and development effort focused on the use of highly parallel machines for applications in the computational aerosciences. In this time period parallel processing for scientific applications evolved from a fringe research topic into the one of main activities at NAS. In this presentation I will review the history of parallel computing at NAS in the context of the major progress, which has been made in the field in general. I will attempt to summarize the lessons we have learned so far, and the contributions NAS has made to the state of the art. Based on these insights I will comment on the current state of parallel computing (including the HPCC effort) and try to predict some trends for the next six years.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Archer, Charles J; Blocksome, Michael A; Cernohous, Bob R
Methods, apparatuses, and computer program products for endpoint-based parallel data processing with non-blocking collective instructions in a parallel active messaging interface (`PAMI`) of a parallel computer are provided. Embodiments include establishing by a parallel application a data communications geometry, the geometry specifying a set of endpoints that are used in collective operations of the PAMI, including associating with the geometry a list of collective algorithms valid for use with the endpoints of the geometry. Embodiments also include registering in each endpoint in the geometry a dispatch callback function for a collective operation and executing without blocking, through a single onemore » of the endpoints in the geometry, an instruction for the collective operation.« less
A parallel variable metric optimization algorithm
NASA Technical Reports Server (NTRS)
Straeter, T. A.
1973-01-01
An algorithm, designed to exploit the parallel computing or vector streaming (pipeline) capabilities of computers is presented. When p is the degree of parallelism, then one cycle of the parallel variable metric algorithm is defined as follows: first, the function and its gradient are computed in parallel at p different values of the independent variable; then the metric is modified by p rank-one corrections; and finally, a single univariant minimization is carried out in the Newton-like direction. Several properties of this algorithm are established. The convergence of the iterates to the solution is proved for a quadratic functional on a real separable Hilbert space. For a finite-dimensional space the convergence is in one cycle when p equals the dimension of the space. Results of numerical experiments indicate that the new algorithm will exploit parallel or pipeline computing capabilities to effect faster convergence than serial techniques.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Archer, Charles J; Blocksome, Michael A; Cernohous, Bob R
Endpoint-based parallel data processing with non-blocking collective instructions in a PAMI of a parallel computer is disclosed. The PAMI is composed of data communications endpoints, each including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task. The compute nodes are coupled for data communications through the PAMI. The parallel application establishes a data communications geometry specifying a set of endpoints that are used in collective operations of the PAMI by associating with the geometry a list of collective algorithms valid for use with themore » endpoints of the geometry; registering in each endpoint in the geometry a dispatch callback function for a collective operation; and executing without blocking, through a single one of the endpoints in the geometry, an instruction for the collective operation.« less
Template based parallel checkpointing in a massively parallel computer system
Archer, Charles Jens [Rochester, MN; Inglett, Todd Alan [Rochester, MN
2009-01-13
A method and apparatus for a template based parallel checkpoint save for a massively parallel super computer system using a parallel variation of the rsync protocol, and network broadcast. In preferred embodiments, the checkpoint data for each node is compared to a template checkpoint file that resides in the storage and that was previously produced. Embodiments herein greatly decrease the amount of data that must be transmitted and stored for faster checkpointing and increased efficiency of the computer system. Embodiments are directed to a parallel computer system with nodes arranged in a cluster with a high speed interconnect that can perform broadcast communication. The checkpoint contains a set of actual small data blocks with their corresponding checksums from all nodes in the system. The data blocks may be compressed using conventional non-lossy data compression algorithms to further reduce the overall checkpoint size.
NASA Technical Reports Server (NTRS)
Weed, Richard Allen; Sankar, L. N.
1994-01-01
An increasing amount of research activity in computational fluid dynamics has been devoted to the development of efficient algorithms for parallel computing systems. The increasing performance to price ratio of engineering workstations has led to research to development procedures for implementing a parallel computing system composed of distributed workstations. This thesis proposal outlines an ongoing research program to develop efficient strategies for performing three-dimensional flow analysis on distributed computing systems. The PVM parallel programming interface was used to modify an existing three-dimensional flow solver, the TEAM code developed by Lockheed for the Air Force, to function as a parallel flow solver on clusters of workstations. Steady flow solutions were generated for three different wing and body geometries to validate the code and evaluate code performance. The proposed research will extend the parallel code development to determine the most efficient strategies for unsteady flow simulations.
NASA Technical Reports Server (NTRS)
Weeks, Cindy Lou
1986-01-01
Experiments were conducted at NASA Ames Research Center to define multi-tasking software requirements for multiple-instruction, multiple-data stream (MIMD) computer architectures. The focus was on specifying solutions for algorithms in the field of computational fluid dynamics (CFD). The program objectives were to allow researchers to produce usable parallel application software as soon as possible after acquiring MIMD computer equipment, to provide researchers with an easy-to-learn and easy-to-use parallel software language which could be implemented on several different MIMD machines, and to enable researchers to list preferred design specifications for future MIMD computer architectures. Analysis of CFD algorithms indicated that extensions of an existing programming language, adaptable to new computer architectures, provided the best solution to meeting program objectives. The CoFORTRAN Language was written in response to these objectives and to provide researchers a means to experiment with parallel software solutions to CFD algorithms on machines with parallel architectures.
NASA Astrophysics Data System (ADS)
Hofierka, Jaroslav; Lacko, Michal; Zubal, Stanislav
2017-10-01
In this paper, we describe the parallelization of three complex and computationally intensive modules of GRASS GIS using the OpenMP application programming interface for multi-core computers. These include the v.surf.rst module for spatial interpolation, the r.sun module for solar radiation modeling and the r.sim.water module for water flow simulation. We briefly describe the functionality of the modules and parallelization approaches used in the modules. Our approach includes the analysis of the module's functionality, identification of source code segments suitable for parallelization and proper application of OpenMP parallelization code to create efficient threads processing the subtasks. We document the efficiency of the solutions using the airborne laser scanning data representing land surface in the test area and derived high-resolution digital terrain model grids. We discuss the performance speed-up and parallelization efficiency depending on the number of processor threads. The study showed a substantial increase in computation speeds on a standard multi-core computer while maintaining the accuracy of results in comparison to the output from original modules. The presented parallelization approach showed the simplicity and efficiency of the parallelization of open-source GRASS GIS modules using OpenMP, leading to an increased performance of this geospatial software on standard multi-core computers.
NASA Astrophysics Data System (ADS)
Kan, Guangyuan; He, Xiaoyan; Ding, Liuqian; Li, Jiren; Hong, Yang; Zuo, Depeng; Ren, Minglei; Lei, Tianjie; Liang, Ke
2018-01-01
Hydrological model calibration has been a hot issue for decades. The shuffled complex evolution method developed at the University of Arizona (SCE-UA) has been proved to be an effective and robust optimization approach. However, its computational efficiency deteriorates significantly when the amount of hydrometeorological data increases. In recent years, the rise of heterogeneous parallel computing has brought hope for the acceleration of hydrological model calibration. This study proposed a parallel SCE-UA method and applied it to the calibration of a watershed rainfall-runoff model, the Xinanjiang model. The parallel method was implemented on heterogeneous computing systems using OpenMP and CUDA. Performance testing and sensitivity analysis were carried out to verify its correctness and efficiency. Comparison results indicated that heterogeneous parallel computing-accelerated SCE-UA converged much more quickly than the original serial version and possessed satisfactory accuracy and stability for the task of fast hydrological model calibration.
Implementation of Parallel Dynamic Simulation on Shared-Memory vs. Distributed-Memory Environments
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jin, Shuangshuang; Chen, Yousu; Wu, Di
2015-12-09
Power system dynamic simulation computes the system response to a sequence of large disturbance, such as sudden changes in generation or load, or a network short circuit followed by protective branch switching operation. It consists of a large set of differential and algebraic equations, which is computational intensive and challenging to solve using single-processor based dynamic simulation solution. High-performance computing (HPC) based parallel computing is a very promising technology to speed up the computation and facilitate the simulation process. This paper presents two different parallel implementations of power grid dynamic simulation using Open Multi-processing (OpenMP) on shared-memory platform, and Messagemore » Passing Interface (MPI) on distributed-memory clusters, respectively. The difference of the parallel simulation algorithms and architectures of the two HPC technologies are illustrated, and their performances for running parallel dynamic simulation are compared and demonstrated.« less
NASA Astrophysics Data System (ADS)
Yang, Sheng-Chun; Lu, Zhong-Yuan; Qian, Hu-Jun; Wang, Yong-Lei; Han, Jie-Ping
2017-11-01
In this work, we upgraded the electrostatic interaction method of CU-ENUF (Yang, et al., 2016) which first applied CUNFFT (nonequispaced Fourier transforms based on CUDA) to the reciprocal-space electrostatic computation and made the computation of electrostatic interaction done thoroughly in GPU. The upgraded edition of CU-ENUF runs concurrently in a hybrid parallel way that enables the computation parallelizing on multiple computer nodes firstly, then further on the installed GPU in each computer. By this parallel strategy, the size of simulation system will be never restricted to the throughput of a single CPU or GPU. The most critical technical problem is how to parallelize a CUNFFT in the parallel strategy, which is conquered effectively by deep-seated research of basic principles and some algorithm skills. Furthermore, the upgraded method is capable of computing electrostatic interactions for both the atomistic molecular dynamics (MD) and the dissipative particle dynamics (DPD). Finally, the benchmarks conducted for validation and performance indicate that the upgraded method is able to not only present a good precision when setting suitable parameters, but also give an efficient way to compute electrostatic interactions for huge simulation systems. Program Files doi:http://dx.doi.org/10.17632/zncf24fhpv.1 Licensing provisions: GNU General Public License 3 (GPL) Programming language: C, C++, and CUDA C Supplementary material: The program is designed for effective electrostatic interactions of large-scale simulation systems, which runs on particular computers equipped with NVIDIA GPUs. It has been tested on (a) single computer node with Intel(R) Core(TM) i7-3770@ 3.40 GHz (CPU) and GTX 980 Ti (GPU), and (b) MPI parallel computer nodes with the same configurations. Nature of problem: For molecular dynamics simulation, the electrostatic interaction is the most time-consuming computation because of its long-range feature and slow convergence in simulation space, which approximately take up most of the total simulation time. Although the parallel method CU-ENUF (Yang et al., 2016) based on GPU has achieved a qualitative leap compared with previous methods in electrostatic interactions computation, the computation capability is limited to the throughput capacity of a single GPU for super-scale simulation system. Therefore, we should look for an effective method to handle the calculation of electrostatic interactions efficiently for a simulation system with super-scale size. Solution method: We constructed a hybrid parallel architecture, in which CPU and GPU are combined to accelerate the electrostatic computation effectively. Firstly, the simulation system is divided into many subtasks via domain-decomposition method. Then MPI (Message Passing Interface) is used to implement the CPU-parallel computation with each computer node corresponding to a particular subtask, and furthermore each subtask in one computer node will be executed in GPU in parallel efficiently. In this hybrid parallel method, the most critical technical problem is how to parallelize a CUNFFT (nonequispaced fast Fourier transform based on CUDA) in the parallel strategy, which is conquered effectively by deep-seated research of basic principles and some algorithm skills. Restrictions: The HP-ENUF is mainly oriented to super-scale system simulations, in which the performance superiority is shown adequately. However, for a small simulation system containing less than 106 particles, the mode of multiple computer nodes has no apparent efficiency advantage or even lower efficiency due to the serious network delay among computer nodes, than the mode of single computer node. References: (1) S.-C. Yang, H.-J. Qian, Z.-Y. Lu, Appl. Comput. Harmon. Anal. 2016, http://dx.doi.org/10.1016/j.acha.2016.04.009. (2) S.-C. Yang, Y.-L. Wang, G.-S. Jiao, H.-J. Qian, Z.-Y. Lu, J. Comput. Chem. 37 (2016) 378. (3) S.-C. Yang, Y.-L. Zhu, H.-J. Qian, Z.-Y. Lu, Appl. Chem. Res. Chin. Univ., 2017, http://dx.doi.org/10.1007/s40242-016-6354-5. (4) Y.-L. Zhu, H. Liu, Z.-W. Li, H.-J. Qian, G. Milano, Z.-Y. Lu, J. Comput. Chem. 34 (2013) 2197.
Zanderighi, Giulia
2018-05-21
Modern QCD - Lecture 2 We will start discussing the matter content of the theory and revisit the experimental measurements that led to the discovery of quarks. We will then consider a classic QCD observable, the R-ratio, and use it to illustrate the appearance of UV divergences and the need to renormalize the coupling constant of QCD. We will then discuss asymptotic freedom and confinement. Finally, we will examine a case where soft and collinear infrared divergences appear, will discuss the soft approximation in QCD and will introduce the concept of infrared safe jets.
NASA Astrophysics Data System (ADS)
Aoki, Sinya
2013-07-01
We review the potential method in lattice QCD, which has recently been proposed to extract nucleon-nucleon interactions via numerical simulations. We focus on the methodology of this approach by emphasizing the strategy of the potential method, the theoretical foundation behind it, and special numerical techniques. We compare the potential method with the standard finite volume method in lattice QCD, in order to make pros and cons of the approach clear. We also present several numerical results for nucleon-nucleon potentials.
Renormalization of Extended QCD2
NASA Astrophysics Data System (ADS)
Fukaya, Hidenori; Yamamura, Ryo
2015-10-01
Extended QCD (XQCD), proposed by Kaplan [D. B. Kaplan, arXiv:1306.5818], is an interesting reformulation of QCD with additional bosonic auxiliary fields. While its partition function is kept exactly the same as that of original QCD, XQCD naturally contains properties of low-energy hadronic models. We analyze the renormalization group flow of 2D (X)QCD, which is solvable in the limit of a large number of colors N_c, to understand what kind of roles the auxiliary degrees of freedom play and how the hadronic picture emerges in the low-energy region.
θ and the η ' in large N supersymmetric QCD
Dine, Michael; Draper, Patrick; Stephenson-Haskins, Laurel; ...
2017-05-22
Here, we study the large N θ dependence and the η' potential in supersymmetric QCD with small soft SUSY-breaking terms. Known exact results in SUSY QCD are found to reflect a variety of expectations from large N perturbation theory, including the presence of branches and the behavior of theories with matter (both with N f << N and N f ~ N ). But, there are also striking departures from ordinary QCD and the conventional large N description: instanton effects, when under control, are not exponentially suppressed at large N , and branched structure in supersymmetric QCD is always associatedmore » with approximate discrete symmetries. We suggest that these differences motivate further study of large N QCD on the lattice.« less
Matrix theory for baryons: an overview of holographic QCD for nuclear physics.
Aoki, Sinya; Hashimoto, Koji; Iizuka, Norihiro
2013-10-01
We provide, for non-experts, a brief overview of holographic QCD (quantum chromodynamics) and a review of the recent proposal (Hashimoto et al 2010 (arXiv:1003.4988[hep-th])) of a matrix-like description of multi-baryon systems in holographic QCD. Based on the matrix model, we derive the baryon interaction at short distances in multi-flavor holographic QCD. We show that there is a very universal repulsive core of inter-baryon forces for a generic number of flavors. This is consistent with a recent lattice QCD analysis for Nf = 2, 3 where the repulsive core looks universal. We also provide a comparison of our results with the lattice QCD and the operator product expansion analysis.
Exploring Partonic Structure of Hadrons Using ab initio Lattice QCD Calculations
Ma, Yan-Qing; Qiu, Jian-Wei
2018-01-10
Following our previous proposal, we construct a class of good "lattice cross sections" (LCSs), from which we can study the partonic structure of hadrons from ab initio lattice QCD calculations. These good LCSs, on the one hand, can be calculated directly in lattice QCD, and on the other hand, can be factorized into parton distribution functions (PDFs) with calculable coefficients, in the same way as QCD factorization for factorizable hadronic cross sections. PDFs could be extracted from QCD global analysis of the lattice QCD generated data of LCSs. In conclusion, we also show that the proposed functions for lattice QCDmore » calculation of PDFs in the literature are special cases of these good LCSs.« less
WTO — a deterministic approach to 4-fermion physics
NASA Astrophysics Data System (ADS)
Passarino, Giampiero
1996-09-01
The program WTO, which is designed for computing cross sections and other relevant observables in the e+e- annihilation into four fermions, is described. The various quantities are computed over both a completely inclusive experimental set-up and a realistic one, i.e. with cuts on the final state energies, final state angles, scattering angles and final state invariant masses. Initial state QED corrections are included by means of the structure function approach while final state QCD corrections are applicable in their naive formulation. A gauge restoring mechanism is included according to the Fermion-Loop scheme. The program structure is highly modular and particular care has been devoted to computing efficiency and speed.
Cooperative storage of shared files in a parallel computing system with dynamic block size
Bent, John M.; Faibish, Sorin; Grider, Gary
2015-11-10
Improved techniques are provided for parallel writing of data to a shared object in a parallel computing system. A method is provided for storing data generated by a plurality of parallel processes to a shared object in a parallel computing system. The method is performed by at least one of the processes and comprises: dynamically determining a block size for storing the data; exchanging a determined amount of the data with at least one additional process to achieve a block of the data having the dynamically determined block size; and writing the block of the data having the dynamically determined block size to a file system. The determined block size comprises, e.g., a total amount of the data to be stored divided by the number of parallel processes. The file system comprises, for example, a log structured virtual parallel file system, such as a Parallel Log-Structured File System (PLFS).
Efficient Parallel Kernel Solvers for Computational Fluid Dynamics Applications
NASA Technical Reports Server (NTRS)
Sun, Xian-He
1997-01-01
Distributed-memory parallel computers dominate today's parallel computing arena. These machines, such as Intel Paragon, IBM SP2, and Cray Origin2OO, have successfully delivered high performance computing power for solving some of the so-called "grand-challenge" problems. Despite initial success, parallel machines have not been widely accepted in production engineering environments due to the complexity of parallel programming. On a parallel computing system, a task has to be partitioned and distributed appropriately among processors to reduce communication cost and to attain load balance. More importantly, even with careful partitioning and mapping, the performance of an algorithm may still be unsatisfactory, since conventional sequential algorithms may be serial in nature and may not be implemented efficiently on parallel machines. In many cases, new algorithms have to be introduced to increase parallel performance. In order to achieve optimal performance, in addition to partitioning and mapping, a careful performance study should be conducted for a given application to find a good algorithm-machine combination. This process, however, is usually painful and elusive. The goal of this project is to design and develop efficient parallel algorithms for highly accurate Computational Fluid Dynamics (CFD) simulations and other engineering applications. The work plan is 1) developing highly accurate parallel numerical algorithms, 2) conduct preliminary testing to verify the effectiveness and potential of these algorithms, 3) incorporate newly developed algorithms into actual simulation packages. The work plan has well achieved. Two highly accurate, efficient Poisson solvers have been developed and tested based on two different approaches: (1) Adopting a mathematical geometry which has a better capacity to describe the fluid, (2) Using compact scheme to gain high order accuracy in numerical discretization. The previously developed Parallel Diagonal Dominant (PDD) algorithm and Reduced Parallel Diagonal Dominant (RPDD) algorithm have been carefully studied on different parallel platforms for different applications, and a NASA simulation code developed by Man M. Rai and his colleagues has been parallelized and implemented based on data dependency analysis. These achievements are addressed in detail in the paper.
Archer, Charles J [Rochester, MN; Blocksome, Michael A [Rochester, MN; Peters, Amanda A [Rochester, MN; Ratterman, Joseph D [Rochester, MN; Smith, Brian E [Rochester, MN
2012-01-10
Methods, apparatus, and products are disclosed for reducing power consumption while synchronizing a plurality of compute nodes during execution of a parallel application that include: beginning, by each compute node, performance of a blocking operation specified by the parallel application, each compute node beginning the blocking operation asynchronously with respect to the other compute nodes; reducing, for each compute node, power to one or more hardware components of that compute node in response to that compute node beginning the performance of the blocking operation; and restoring, for each compute node, the power to the hardware components having power reduced in response to all of the compute nodes beginning the performance of the blocking operation.
Archer, Charles J [Rochester, MN; Blocksome, Michael A [Rochester, MN; Peters, Amanda E [Cambridge, MA; Ratterman, Joseph D [Rochester, MN; Smith, Brian E [Rochester, MN
2012-04-17
Methods, apparatus, and products are disclosed for reducing power consumption while synchronizing a plurality of compute nodes during execution of a parallel application that include: beginning, by each compute node, performance of a blocking operation specified by the parallel application, each compute node beginning the blocking operation asynchronously with respect to the other compute nodes; reducing, for each compute node, power to one or more hardware components of that compute node in response to that compute node beginning the performance of the blocking operation; and restoring, for each compute node, the power to the hardware components having power reduced in response to all of the compute nodes beginning the performance of the blocking operation.
MPI implementation of PHOENICS: A general purpose computational fluid dynamics code
NASA Astrophysics Data System (ADS)
Simunovic, S.; Zacharia, T.; Baltas, N.; Spalding, D. B.
1995-03-01
PHOENICS is a suite of computational analysis programs that are used for simulation of fluid flow, heat transfer, and dynamical reaction processes. The parallel version of the solver EARTH for the Computational Fluid Dynamics (CFD) program PHOENICS has been implemented using Message Passing Interface (MPI) standard. Implementation of MPI version of PHOENICS makes this computational tool portable to a wide range of parallel machines and enables the use of high performance computing for large scale computational simulations. MPI libraries are available on several parallel architectures making the program usable across different architectures as well as on heterogeneous computer networks. The Intel Paragon NX and MPI versions of the program have been developed and tested on massively parallel supercomputers Intel Paragon XP/S 5, XP/S 35, and Kendall Square Research, and on the multiprocessor SGI Onyx computer at Oak Ridge National Laboratory. The preliminary testing results of the developed program have shown scalable performance for reasonably sized computational domains.
MPI implementation of PHOENICS: A general purpose computational fluid dynamics code
DOE Office of Scientific and Technical Information (OSTI.GOV)
Simunovic, S.; Zacharia, T.; Baltas, N.
1995-04-01
PHOENICS is a suite of computational analysis programs that are used for simulation of fluid flow, heat transfer, and dynamical reaction processes. The parallel version of the solver EARTH for the Computational Fluid Dynamics (CFD) program PHOENICS has been implemented using Message Passing Interface (MPI) standard. Implementation of MPI version of PHOENICS makes this computational tool portable to a wide range of parallel machines and enables the use of high performance computing for large scale computational simulations. MPI libraries are available on several parallel architectures making the program usable across different architectures as well as on heterogeneous computer networks. Themore » Intel Paragon NX and MPI versions of the program have been developed and tested on massively parallel supercomputers Intel Paragon XP/S 5, XP/S 35, and Kendall Square Research, and on the multiprocessor SGI Onyx computer at Oak Ridge National Laboratory. The preliminary testing results of the developed program have shown scalable performance for reasonably sized computational domains.« less
Biocellion: accelerating computer simulation of multicellular biological system models
Kang, Seunghwa; Kahan, Simon; McDermott, Jason; Flann, Nicholas; Shmulevich, Ilya
2014-01-01
Motivation: Biological system behaviors are often the outcome of complex interactions among a large number of cells and their biotic and abiotic environment. Computational biologists attempt to understand, predict and manipulate biological system behavior through mathematical modeling and computer simulation. Discrete agent-based modeling (in combination with high-resolution grids to model the extracellular environment) is a popular approach for building biological system models. However, the computational complexity of this approach forces computational biologists to resort to coarser resolution approaches to simulate large biological systems. High-performance parallel computers have the potential to address the computing challenge, but writing efficient software for parallel computers is difficult and time-consuming. Results: We have developed Biocellion, a high-performance software framework, to solve this computing challenge using parallel computers. To support a wide range of multicellular biological system models, Biocellion asks users to provide their model specifics by filling the function body of pre-defined model routines. Using Biocellion, modelers without parallel computing expertise can efficiently exploit parallel computers with less effort than writing sequential programs from scratch. We simulate cell sorting, microbial patterning and a bacterial system in soil aggregate as case studies. Availability and implementation: Biocellion runs on x86 compatible systems with the 64 bit Linux operating system and is freely available for academic use. Visit http://biocellion.com for additional information. Contact: seunghwa.kang@pnnl.gov PMID:25064572
Research in Parallel Algorithms and Software for Computational Aerosciences
NASA Technical Reports Server (NTRS)
Domel, Neal D.
1996-01-01
Phase I is complete for the development of a Computational Fluid Dynamics parallel code with automatic grid generation and adaptation for the Euler analysis of flow over complex geometries. SPLITFLOW, an unstructured Cartesian grid code developed at Lockheed Martin Tactical Aircraft Systems, has been modified for a distributed memory/massively parallel computing environment. The parallel code is operational on an SGI network, Cray J90 and C90 vector machines, SGI Power Challenge, and Cray T3D and IBM SP2 massively parallel machines. Parallel Virtual Machine (PVM) is the message passing protocol for portability to various architectures. A domain decomposition technique was developed which enforces dynamic load balancing to improve solution speed and memory requirements. A host/node algorithm distributes the tasks. The solver parallelizes very well, and scales with the number of processors. Partially parallelized and non-parallelized tasks consume most of the wall clock time in a very fine grain environment. Timing comparisons on a Cray C90 demonstrate that Parallel SPLITFLOW runs 2.4 times faster on 8 processors than its non-parallel counterpart autotasked over 8 processors.
Research in Parallel Algorithms and Software for Computational Aerosciences
NASA Technical Reports Server (NTRS)
Domel, Neal D.
1996-01-01
Phase 1 is complete for the development of a computational fluid dynamics CFD) parallel code with automatic grid generation and adaptation for the Euler analysis of flow over complex geometries. SPLITFLOW, an unstructured Cartesian grid code developed at Lockheed Martin Tactical Aircraft Systems, has been modified for a distributed memory/massively parallel computing environment. The parallel code is operational on an SGI network, Cray J90 and C90 vector machines, SGI Power Challenge, and Cray T3D and IBM SP2 massively parallel machines. Parallel Virtual Machine (PVM) is the message passing protocol for portability to various architectures. A domain decomposition technique was developed which enforces dynamic load balancing to improve solution speed and memory requirements. A host/node algorithm distributes the tasks. The solver parallelizes very well, and scales with the number of processors. Partially parallelized and non-parallelized tasks consume most of the wall clock time in a very fine grain environment. Timing comparisons on a Cray C90 demonstrate that Parallel SPLITFLOW runs 2.4 times faster on 8 processors than its non-parallel counterpart autotasked over 8 processors.
NASA Technical Reports Server (NTRS)
Logan, Terry G.
1994-01-01
The purpose of this study is to investigate the performance of the integral equation computations using numerical source field-panel method in a massively parallel processing (MPP) environment. A comparative study of computational performance of the MPP CM-5 computer and conventional Cray-YMP supercomputer for a three-dimensional flow problem is made. A serial FORTRAN code is converted into a parallel CM-FORTRAN code. Some performance results are obtained on CM-5 with 32, 62, 128 nodes along with those on Cray-YMP with a single processor. The comparison of the performance indicates that the parallel CM-FORTRAN code near or out-performs the equivalent serial FORTRAN code for some cases.
Parallel aeroelastic computations for wing and wing-body configurations
NASA Technical Reports Server (NTRS)
Byun, Chansup
1994-01-01
The objective of this research is to develop computationally efficient methods for solving fluid-structural interaction problems by directly coupling finite difference Euler/Navier-Stokes equations for fluids and finite element dynamics equations for structures on parallel computers. This capability will significantly impact many aerospace projects of national importance such as Advanced Subsonic Civil Transport (ASCT), where the structural stability margin becomes very critical at the transonic region. This research effort will have direct impact on the High Performance Computing and Communication (HPCC) Program of NASA in the area of parallel computing.
Implementation of a 3D mixing layer code on parallel computers
NASA Technical Reports Server (NTRS)
Roe, K.; Thakur, R.; Dang, T.; Bogucz, E.
1995-01-01
This paper summarizes our progress and experience in the development of a Computational-Fluid-Dynamics code on parallel computers to simulate three-dimensional spatially-developing mixing layers. In this initial study, the three-dimensional time-dependent Euler equations are solved using a finite-volume explicit time-marching algorithm. The code was first programmed in Fortran 77 for sequential computers. The code was then converted for use on parallel computers using the conventional message-passing technique, while we have not been able to compile the code with the present version of HPF compilers.
NASA Workshop on Computational Structural Mechanics 1987, part 1
NASA Technical Reports Server (NTRS)
Sykes, Nancy P. (Editor)
1989-01-01
Topics in Computational Structural Mechanics (CSM) are reviewed. CSM parallel structural methods, a transputer finite element solver, architectures for multiprocessor computers, and parallel eigenvalue extraction are among the topics discussed.
Evidence for color fluctuations in hadrons from coherent nuclear diffraction
DOE Office of Scientific and Technical Information (OSTI.GOV)
Frankfurt, L.; Miller, G.A.; Strikman, M.
A QCD-based treatment of projectile size fluctuations is used to compute inelastic diffractive cross sections [sigma][sub diff] for coherent hadron-nuclear processes. We find that fluctuations near the average size give the major contribution to the cross section with [lt] few % contribution from small size configurations. The computed values of [sigma][sub diff] are consistent with the limited available data. The importance of coherent diffraction studies for a wide range of projectiles for high energy Fermilab fixed target experiments is emphasized. The implications of these significant color fluctuations for relativistic heavy ion collisions are discussed.
Parallel computation with molecular-motor-propelled agents in nanofabricated networks.
Nicolau, Dan V; Lard, Mercy; Korten, Till; van Delft, Falco C M J M; Persson, Malin; Bengtsson, Elina; Månsson, Alf; Diez, Stefan; Linke, Heiner; Nicolau, Dan V
2016-03-08
The combinatorial nature of many important mathematical problems, including nondeterministic-polynomial-time (NP)-complete problems, places a severe limitation on the problem size that can be solved with conventional, sequentially operating electronic computers. There have been significant efforts in conceiving parallel-computation approaches in the past, for example: DNA computation, quantum computation, and microfluidics-based computation. However, these approaches have not proven, so far, to be scalable and practical from a fabrication and operational perspective. Here, we report the foundations of an alternative parallel-computation system in which a given combinatorial problem is encoded into a graphical, modular network that is embedded in a nanofabricated planar device. Exploring the network in a parallel fashion using a large number of independent, molecular-motor-propelled agents then solves the mathematical problem. This approach uses orders of magnitude less energy than conventional computers, thus addressing issues related to power consumption and heat dissipation. We provide a proof-of-concept demonstration of such a device by solving, in a parallel fashion, the small instance {2, 5, 9} of the subset sum problem, which is a benchmark NP-complete problem. Finally, we discuss the technical advances necessary to make our system scalable with presently available technology.
NASA Astrophysics Data System (ADS)
Moon, Hongsik
What is the impact of multicore and associated advanced technologies on computational software for science? Most researchers and students have multicore laptops or desktops for their research and they need computing power to run computational software packages. Computing power was initially derived from Central Processing Unit (CPU) clock speed. That changed when increases in clock speed became constrained by power requirements. Chip manufacturers turned to multicore CPU architectures and associated technological advancements to create the CPUs for the future. Most software applications benefited by the increased computing power the same way that increases in clock speed helped applications run faster. However, for Computational ElectroMagnetics (CEM) software developers, this change was not an obvious benefit - it appeared to be a detriment. Developers were challenged to find a way to correctly utilize the advancements in hardware so that their codes could benefit. The solution was parallelization and this dissertation details the investigation to address these challenges. Prior to multicore CPUs, advanced computer technologies were compared with the performance using benchmark software and the metric was FLoting-point Operations Per Seconds (FLOPS) which indicates system performance for scientific applications that make heavy use of floating-point calculations. Is FLOPS an effective metric for parallelized CEM simulation tools on new multicore system? Parallel CEM software needs to be benchmarked not only by FLOPS but also by the performance of other parameters related to type and utilization of the hardware, such as CPU, Random Access Memory (RAM), hard disk, network, etc. The codes need to be optimized for more than just FLOPs and new parameters must be included in benchmarking. In this dissertation, the parallel CEM software named High Order Basis Based Integral Equation Solver (HOBBIES) is introduced. This code was developed to address the needs of the changing computer hardware platforms in order to provide fast, accurate and efficient solutions to large, complex electromagnetic problems. The research in this dissertation proves that the performance of parallel code is intimately related to the configuration of the computer hardware and can be maximized for different hardware platforms. To benchmark and optimize the performance of parallel CEM software, a variety of large, complex projects are created and executed on a variety of computer platforms. The computer platforms used in this research are detailed in this dissertation. The projects run as benchmarks are also described in detail and results are presented. The parameters that affect parallel CEM software on High Performance Computing Clusters (HPCC) are investigated. This research demonstrates methods to maximize the performance of parallel CEM software code.
NASA Astrophysics Data System (ADS)
Iritani, Takumi
2018-03-01
Both direct and HAL QCD methods are currently used to study the hadron interactions in lattice QCD. In the direct method, the eigen-energy of two-particle is measured from the temporal correlation. Due to the contamination of excited states, however, the direct method suffers from the fake eigen-energy problem, which we call the "mirage problem," while the HAL QCD method can extract information from all elastic states by using the spatial correlation. In this work, we further investigate systematic uncertainties of the HAL QCD method such as the quark source operator dependence, the convergence of the derivative expansion of the non-local interaction kernel, and the single baryon saturation, which are found to be well controlled. We also confirm the consistency between the HAL QCD method and the Lüscher's finite volume formula. Based on the HAL QCD potential, we quantitatively confirm that the mirage plateau in the direct method is indeed caused by the contamination of excited states.
A highly efficient multi-core algorithm for clustering extremely large datasets
2010-01-01
Background In recent years, the demand for computational power in computational biology has increased due to rapidly growing data sets from microarray and other high-throughput technologies. This demand is likely to increase. Standard algorithms for analyzing data, such as cluster algorithms, need to be parallelized for fast processing. Unfortunately, most approaches for parallelizing algorithms largely rely on network communication protocols connecting and requiring multiple computers. One answer to this problem is to utilize the intrinsic capabilities in current multi-core hardware to distribute the tasks among the different cores of one computer. Results We introduce a multi-core parallelization of the k-means and k-modes cluster algorithms based on the design principles of transactional memory for clustering gene expression microarray type data and categorial SNP data. Our new shared memory parallel algorithms show to be highly efficient. We demonstrate their computational power and show their utility in cluster stability and sensitivity analysis employing repeated runs with slightly changed parameters. Computation speed of our Java based algorithm was increased by a factor of 10 for large data sets while preserving computational accuracy compared to single-core implementations and a recently published network based parallelization. Conclusions Most desktop computers and even notebooks provide at least dual-core processors. Our multi-core algorithms show that using modern algorithmic concepts, parallelization makes it possible to perform even such laborious tasks as cluster sensitivity and cluster number estimation on the laboratory computer. PMID:20370922
Effective field theories for muonic hydrogen
NASA Astrophysics Data System (ADS)
Peset, Clara
2017-03-01
Experimental measurements of muonic hydrogen bound states have recently started to take place and provide a powerful setting in which to study the properties of QCD. We profit from the power of effective field theories (EFTs) to provide a theoretical framework in which to study muonic hydrogen in a model independent fashion. In particular, we compute expressions for the Lamb shift and the hyperfine splitting. These expressions include the leading logarithmic O(mμα6) terms, as well as the leading {\\cal O}≤ft( {{m_μ }{α ^5}{{m_μ ^2} \\over {Λ {{QCD}}^2}}} \\right) hadronic effects. Most remarkably, our analyses include the determination of the spin-dependent and spin-independent structure functions of the forward virtualphoton Compton tensor of the proton to O(p3) in HBET and including the Delta particle. Using these results we obtain the leading hadronic contributions to the Wilson coeffcients of the lepton-proton four fermion operators in NRQED. The spin-independent coeffcient yields a pure prediction for the two-photon exchange contribution to the muonic hydrogen Lamb shift, which is the main source of uncertainty in our computation. The spindependent coeffcient yields the prediction of the hyperfine splitting. The use of EFTs crucially helps us organizing the computation, in such a way that we can clearly address the parametric accuracy of our result. Furthermore, we review in the context of NRQED all the contributions to the energy shift of O(mμα5, as well as those that scale like mrα6× logarithms.
Support for Debugging Automatically Parallelized Programs
NASA Technical Reports Server (NTRS)
Hood, Robert; Jost, Gabriele; Biegel, Bryan (Technical Monitor)
2001-01-01
This viewgraph presentation provides information on the technical aspects of debugging computer code that has been automatically converted for use in a parallel computing system. Shared memory parallelization and distributed memory parallelization entail separate and distinct challenges for a debugging program. A prototype system has been developed which integrates various tools for the debugging of automatically parallelized programs including the CAPTools Database which provides variable definition information across subroutines as well as array distribution information.
Parallel language constructs for tensor product computations on loosely coupled architectures
NASA Technical Reports Server (NTRS)
Mehrotra, Piyush; Van Rosendale, John
1989-01-01
A set of language primitives designed to allow the specification of parallel numerical algorithms at a higher level is described. The authors focus on tensor product array computations, a simple but important class of numerical algorithms. They consider first the problem of programming one-dimensional kernel routines, such as parallel tridiagonal solvers, and then look at how such parallel kernels can be combined to form parallel tensor product algorithms.
NASA Astrophysics Data System (ADS)
Jang, W.; Engda, T. A.; Neff, J. C.; Herrick, J.
2017-12-01
Many crop models are increasingly used to evaluate crop yields at regional and global scales. However, implementation of these models across large areas using fine-scale grids is limited by computational time requirements. In order to facilitate global gridded crop modeling with various scenarios (i.e., different crop, management schedule, fertilizer, and irrigation) using the Environmental Policy Integrated Climate (EPIC) model, we developed a distributed parallel computing framework in Python. Our local desktop with 14 cores (28 threads) was used to test the distributed parallel computing framework in Iringa, Tanzania which has 406,839 grid cells. High-resolution soil data, SoilGrids (250 x 250 m), and climate data, AgMERRA (0.25 x 0.25 deg) were also used as input data for the gridded EPIC model. The framework includes a master file for parallel computing, input database, input data formatters, EPIC model execution, and output analyzers. Through the master file for parallel computing, the user-defined number of threads of CPU divides the EPIC simulation into jobs. Then, Using EPIC input data formatters, the raw database is formatted for EPIC input data and the formatted data moves into EPIC simulation jobs. Then, 28 EPIC jobs run simultaneously and only interesting results files are parsed and moved into output analyzers. We applied various scenarios with seven different slopes and twenty-four fertilizer ranges. Parallelized input generators create different scenarios as a list for distributed parallel computing. After all simulations are completed, parallelized output analyzers are used to analyze all outputs according to the different scenarios. This saves significant computing time and resources, making it possible to conduct gridded modeling at regional to global scales with high-resolution data. For example, serial processing for the Iringa test case would require 113 hours, while using the framework developed in this study requires only approximately 6 hours, a nearly 95% reduction in computing time.
Nucleon QCD sum rules in the instanton medium
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ryskin, M. G.; Drukarev, E. G., E-mail: drukarev@pnpi.spb.ru; Sadovnikova, V. A.
2015-09-15
We try to find grounds for the standard nucleon QCD sum rules, based on a more detailed description of the QCD vacuum. We calculate the polarization operator of the nucleon current in the instanton medium. The medium (QCD vacuum) is assumed to be a composition of the small-size instantons and some long-wave gluon fluctuations. We solve the corresponding QCD sum rule equations and demonstrate that there is a solution with the value of the nucleon mass close to the physical one if the fraction of the small-size instantons contribution is w{sub s} ≈ 2/3.
Li, Chuan; Petukh, Marharyta; Li, Lin; Alexov, Emil
2013-08-15
Due to the enormous importance of electrostatics in molecular biology, calculating the electrostatic potential and corresponding energies has become a standard computational approach for the study of biomolecules and nano-objects immersed in water and salt phase or other media. However, the electrostatics of large macromolecules and macromolecular complexes, including nano-objects, may not be obtainable via explicit methods and even the standard continuum electrostatics methods may not be applicable due to high computational time and memory requirements. Here, we report further development of the parallelization scheme reported in our previous work (Li, et al., J. Comput. Chem. 2012, 33, 1960) to include parallelization of the molecular surface and energy calculations components of the algorithm. The parallelization scheme utilizes different approaches such as space domain parallelization, algorithmic parallelization, multithreading, and task scheduling, depending on the quantity being calculated. This allows for efficient use of the computing resources of the corresponding computer cluster. The parallelization scheme is implemented in the popular software DelPhi and results in speedup of several folds. As a demonstration of the efficiency and capability of this methodology, the electrostatic potential, and electric field distributions are calculated for the bovine mitochondrial supercomplex illustrating their complex topology, which cannot be obtained by modeling the supercomplex components alone. Copyright © 2013 Wiley Periodicals, Inc.
Redundant binary number representation for an inherently parallel arithmetic on optical computers.
De Biase, G A; Massini, A
1993-02-10
A simple redundant binary number representation suitable for digital-optical computers is presented. By means of this representation it is possible to build an arithmetic with carry-free parallel algebraic sums carried out in constant time and parallel multiplication in log N time. This redundant number representation naturally fits the 2's complement binary number system and permits the construction of inherently parallel arithmetic units that are used in various optical technologies. Some properties of this number representation and several examples of computation are presented.
Backtracking and Re-execution in the Automatic Debugging of Parallelized Programs
NASA Technical Reports Server (NTRS)
Matthews, Gregory; Hood, Robert; Johnson, Stephen; Leggett, Peter; Biegel, Bryan (Technical Monitor)
2002-01-01
In this work we describe a new approach using relative debugging to find differences in computation between a serial program and a parallel version of th it program. We use a combination of re-execution and backtracking in order to find the first difference in computation that may ultimately lead to an incorrect value that the user has indicated. In our prototype implementation we use static analysis information from a parallelization tool in order to perform the backtracking as well as the mapping required between serial and parallel computations.
Flavour symmetry breaking in the kaon parton distribution amplitude
none,
2014-11-01
We compute the kaon's valence-quark (twist-two parton) distribution amplitude (PDA) by projecting its Poincaré-covariant Bethe–Salpeter wave-function onto the light-front. At a scale ζ = 2 GeV, the PDA is a broad, concave and asymmetric function, whose peak is shifted 12–16% away from its position in QCD's conformal limit. These features are a clear expression of SU(3)-flavour-symmetry breaking. They show that the heavier quark in the kaon carries more of the bound-state's momentum than the lighter quark and also that emergent phenomena in QCD modulate the magnitude of flavour-symmetry breaking: it is markedly smaller than one might expect based on themore » difference between light-quark current masses. Our results add to a body of evidence which indicates that at any energy scale accessible with existing or foreseeable facilities, a reliable guide to the interpretation of experiment requires the use of such nonperturbatively broadened PDAs in leading-order, leading-twist formulae for hard exclusive processes instead of the asymptotic PDA associated with QCD's conformal limit. We illustrate this via the ratio of kaon and pion electromagnetic form factors: using our nonperturbative PDAs in the appropriate formulae, F K/F π=1.23 at spacelike-Q 2=17 GeV 2, which compares satisfactorily with the value of 0.92(5) inferred in e +e - annihilation at s=17 GeV 2.« less
Production of heavy Higgs bosons and decay into top quarks at the LHC
NASA Astrophysics Data System (ADS)
Bernreuther, W.; Galler, P.; Mellein, C.; Si, Z.-G.; Uwer, P.
2016-02-01
We investigate the production of heavy, neutral Higgs boson resonances and their decays to top-quark top-antiquark (t t ¯) pairs at the Large Hadron Collider (LHC) at next-to-leading order (NLO) in the strong coupling of quantum chromodynamics (QCD). The NLO corrections to heavy Higgs boson production and the Higgs-QCD interference are calculated in the large mt limit with an effective K-factor rescaling. The nonresonant t t ¯ background is taken into account at NLO QCD including weak-interaction corrections. In order to consistently determine the total decay widths of the heavy Higgs bosons, we consider for definiteness the type-II two-Higgs-doublet extension of the standard model and choose three parameter scenarios that entail two heavy neutral Higgs bosons with masses above the t t ¯ threshold and unsuppressed Yukawa couplings to top quarks. For these three scenarios we compute, for the LHC operating at 13 TeV, the t t ¯ cross section and the distributions of the t t ¯ invariant mass, of the transverse top-quark momentum and rapidity, and of the cosine of the Collins-Soper angle with and without the two heavy Higgs resonances. For selected Mt t ¯ bins we estimate the significances for detecting a heavy Higgs signal in the t t ¯ dileptonic and lepton plus jets decay channels.
Traffic Simulations on Parallel Computers Using Domain Decomposition Techniques
DOT National Transportation Integrated Search
1995-01-01
Large scale simulations of Intelligent Transportation Systems (ITS) can only be acheived by using the computing resources offered by parallel computing architectures. Domain decomposition techniques are proposed which allow the performance of traffic...
A Multi-Level Parallelization Concept for High-Fidelity Multi-Block Solvers
NASA Technical Reports Server (NTRS)
Hatay, Ferhat F.; Jespersen, Dennis C.; Guruswamy, Guru P.; Rizk, Yehia M.; Byun, Chansup; Gee, Ken; VanDalsem, William R. (Technical Monitor)
1997-01-01
The integration of high-fidelity Computational Fluid Dynamics (CFD) analysis tools with the industrial design process benefits greatly from the robust implementations that are transportable across a wide range of computer architectures. In the present work, a hybrid domain-decomposition and parallelization concept was developed and implemented into the widely-used NASA multi-block Computational Fluid Dynamics (CFD) packages implemented in ENSAERO and OVERFLOW. The new parallel solver concept, PENS (Parallel Euler Navier-Stokes Solver), employs both fine and coarse granularity in data partitioning as well as data coalescing to obtain the desired load-balance characteristics on the available computer platforms. This multi-level parallelism implementation itself introduces no changes to the numerical results, hence the original fidelity of the packages are identically preserved. The present implementation uses the Message Passing Interface (MPI) library for interprocessor message passing and memory accessing. By choosing an appropriate combination of the available partitioning and coalescing capabilities only during the execution stage, the PENS solver becomes adaptable to different computer architectures from shared-memory to distributed-memory platforms with varying degrees of parallelism. The PENS implementation on the IBM SP2 distributed memory environment at the NASA Ames Research Center obtains 85 percent scalable parallel performance using fine-grain partitioning of single-block CFD domains using up to 128 wide computational nodes. Multi-block CFD simulations of complete aircraft simulations achieve 75 percent perfect load-balanced executions using data coalescing and the two levels of parallelism. SGI PowerChallenge, SGI Origin 2000, and a cluster of workstations are the other platforms where the robustness of the implementation is tested. The performance behavior on the other computer platforms with a variety of realistic problems will be included as this on-going study progresses.
The Nucleon Axial Form Factor and Staggered Lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Meyer, Aaron Scott
The study of neutrino oscillation physics is a major research goal of the worldwide particle physics program over the upcoming decade. Many new experiments are being built to study the properties of neutrinos and to answer questions about the phenomenon of neutrino oscillation. These experiments need precise theoretical cross sections in order to access fundamental neutrino properties. Neutrino oscillation experiments often use large atomic nuclei as scattering targets, which are challenging for theorists to model. Nuclear models rely on free-nucleon amplitudes as inputs. These amplitudes are constrained by scattering experiments with large nuclear targets that rely on the very samemore » nuclear models. The work in this dissertation is the rst step of a new initiative to isolate and compute elementary amplitudes with theoretical calculations to support the neutrino oscillation experimental program. Here, the eort focuses on computing the axial form factor, which is the largest contributor of systematic error in the primary signal measurement process for neutrino oscillation studies, quasielastic scattering. Two approaches are taken. First, neutrino scattering data on a deuterium target are reanalyzed with a model-independent parametrization of the axial form factor to quantify the present uncertainty in the free-nucleon amplitudes. The uncertainties on the free-nucleon cross section are found to be underestimated by about an order of magnitude compared to the ubiquitous dipole model parametrization. The second approach uses lattice QCD to perform a rst-principles computation of the nucleon axial form factor. The Highly Improved Staggered Quark (HISQ) action is employed for both valence and sea quarks. The results presented in this dissertation are computed at physical pion mass for one lattice spacing. This work presents a computation of the axial form factor at zero momentum transfer, and forms the basis for a computation of the axial form factor momentum dependence with an extrapolation to the continuum limit and a full systematic error budget.« less
Renormalization of Supersymmetric QCD on the Lattice
NASA Astrophysics Data System (ADS)
Costa, Marios; Panagopoulos, Haralambos
2018-03-01
We perform a pilot study of the perturbative renormalization of a Supersymmetric gauge theory with matter fields on the lattice. As a specific example, we consider Supersymmetric N=1 QCD (SQCD). We study the self-energies of all particles which appear in this theory, as well as the renormalization of the coupling constant. To this end we compute, perturbatively to one-loop, the relevant two-point and three-point Green's functions using both dimensional and lattice regularizations. Our lattice formulation involves theWilson discretization for the gluino and quark fields; for gluons we employ the Wilson gauge action; for scalar fields (squarks) we use naive discretization. The gauge group that we consider is SU(Nc), while the number of colors, Nc, the number of flavors, Nf, and the gauge parameter, α, are left unspecified. We obtain analytic expressions for the renormalization factors of the coupling constant (Zg) and of the quark (ZΨ), gluon (Zu), gluino (Zλ), squark (ZA±), and ghost (Zc) fields on the lattice. We also compute the critical values of the gluino, quark and squark masses. Finally, we address the mixing which occurs among squark degrees of freedom beyond tree level: we calculate the corresponding mixing matrix which is necessary in order to disentangle the components of the squark field via an additional finite renormalization.
Supersymmetric QCD on the lattice: An exploratory study
NASA Astrophysics Data System (ADS)
Costa, M.; Panagopoulos, H.
2017-08-01
We perform a pilot study of the perturbative renormalization of a supersymmetric gauge theory with matter fields on the lattice. As a specific example, we consider supersymmetric N =1 QCD (SQCD). We study the self-energies of all particles which appear in this theory, as well as the renormalization of the coupling constant. To this end we compute, perturbatively to one-loop, the relevant two-point and three-point Green's functions using both dimensional and lattice regularizations. Our lattice formulation involves the Wilson discretization for the gluino and quark fields; for gluons we employ the Wilson gauge action; for scalar fields (squarks) we use naïve discretization. The gauge group that we consider is S U (Nc), while the number of colors, Nc, the number of flavors, Nf, and the gauge parameter, α , are left unspecified. We obtain analytic expressions for the renormalization factors of the coupling constant (Zg) and of the quark (Zψ), gluon (Zu), gluino (Zλ), squark (ZA ±), and ghost (Zc) fields on the lattice. We also compute the critical values of the gluino, quark and squark masses. Finally, we address the mixing which occurs among squark degrees of freedom beyond tree level: we calculate the corresponding mixing matrix which is necessary in order to disentangle the components of the squark field via an additional finite renormalization.
Parallel computing on Unix workstation arrays
NASA Astrophysics Data System (ADS)
Reale, F.; Bocchino, F.; Sciortino, S.
1994-12-01
We have tested arrays of general-purpose Unix workstations used as MIMD systems for massive parallel computations. In particular we have solved numerically a demanding test problem with a 2D hydrodynamic code, generally developed to study astrophysical flows, by exucuting it on arrays either of DECstations 5000/200 on Ethernet LAN, or of DECstations 3000/400, equipped with powerful Alpha processors, on FDDI LAN. The code is appropriate for data-domain decomposition, and we have used a library for parallelization previously developed in our Institute, and easily extended to work on Unix workstation arrays by using the PVM software toolset. We have compared the parallel efficiencies obtained on arrays of several processors to those obtained on a dedicated MIMD parallel system, namely a Meiko Computing Surface (CS-1), equipped with Intel i860 processors. We discuss the feasibility of using non-dedicated parallel systems and conclude that the convenience depends essentially on the size of the computational domain as compared to the relative processor power and network bandwidth. We point out that for future perspectives a parallel development of processor and network technology is important, and that the software still offers great opportunities of improvement, especially in terms of latency times in the message-passing protocols. In conditions of significant gain in terms of speedup, such workstation arrays represent a cost-effective approach to massive parallel computations.
Parallelization strategies for continuum-generalized method of moments on the multi-thread systems
NASA Astrophysics Data System (ADS)
Bustamam, A.; Handhika, T.; Ernastuti, Kerami, D.
2017-07-01
Continuum-Generalized Method of Moments (C-GMM) covers the Generalized Method of Moments (GMM) shortfall which is not as efficient as Maximum Likelihood estimator by using the continuum set of moment conditions in a GMM framework. However, this computation would take a very long time since optimizing regularization parameter. Unfortunately, these calculations are processed sequentially whereas in fact all modern computers are now supported by hierarchical memory systems and hyperthreading technology, which allowing for parallel computing. This paper aims to speed up the calculation process of C-GMM by designing a parallel algorithm for C-GMM on the multi-thread systems. First, parallel regions are detected for the original C-GMM algorithm. There are two parallel regions in the original C-GMM algorithm, that are contributed significantly to the reduction of computational time: the outer-loop and the inner-loop. Furthermore, this parallel algorithm will be implemented with standard shared-memory application programming interface, i.e. Open Multi-Processing (OpenMP). The experiment shows that the outer-loop parallelization is the best strategy for any number of observations.
Multirate-based fast parallel algorithms for 2-D DHT-based real-valued discrete Gabor transform.
Tao, Liang; Kwan, Hon Keung
2012-07-01
Novel algorithms for the multirate and fast parallel implementation of the 2-D discrete Hartley transform (DHT)-based real-valued discrete Gabor transform (RDGT) and its inverse transform are presented in this paper. A 2-D multirate-based analysis convolver bank is designed for the 2-D RDGT, and a 2-D multirate-based synthesis convolver bank is designed for the 2-D inverse RDGT. The parallel channels in each of the two convolver banks have a unified structure and can apply the 2-D fast DHT algorithm to speed up their computations. The computational complexity of each parallel channel is low and is independent of the Gabor oversampling rate. All the 2-D RDGT coefficients of an image are computed in parallel during the analysis process and can be reconstructed in parallel during the synthesis process. The computational complexity and time of the proposed parallel algorithms are analyzed and compared with those of the existing fastest algorithms for 2-D discrete Gabor transforms. The results indicate that the proposed algorithms are the fastest, which make them attractive for real-time image processing.
Usami, Masahide; Iwadare, Yoshitaka; Watanabe, Kyota; Ushijima, Hirokage; Kodaira, Masaki; Okada, Takashi; Sasayama, Daimei; Sugiyama, Nobuhiro; Saito, Kazuhiko
2015-07-01
The parent-assessed children-with-difficulties questionnaire (Questionnaire-Children with Difficulties; QCD) is designed to evaluate a child׳s difficulties in functioning during specific periods of the day. This study aimed to use the QCD to evaluate the difficulties in daily functioning experienced by children with depressive disorders. A case-control design was used. The cases comprised 90 junior high school students with depressive disorder, whereas a community sample of 363 junior high school students was enrolled as controls. Behaviors were assessed using the QCD, Depression Self-Rating Scale (DSRS), Tokyo Autistic Behavior Scale (TABS), attention deficit hyperactivity disorder-rating scale (ADHD-RS), and Oppositional Defiant Behavior Inventory (ODBI). We then analyzed the effects of sex and diagnosis on the QCD scores as well as the correlation coefficients between the QCD and the other questionnaires. We included 90 cases (33 boys, 57 girls) with depressive disorders and 363 controls (180 boys, 183 girls). The QCD scores for the children with depressive disorders were significantly lower compared with those from the community sample (P<0.001). The morning, school-time, and night subscores of the QCD were lower for the children with both depressive disorders and truancy problems than for those with depressive disorders alone (P<0.001). Significant correlations were observed between the following: the night QCD subscore and the DSRS scores among boys, the morning QCD subscore and ADHD-RS inattention scores for all groups, and the evening QCD subscore and the TABS score. Parents reported that children with depressive disorders experienced greater difficulties in completing basic daily activities compared with community controls. These difficulties were dependent on sex, symptoms, and the time of day. The use of QCD to assess children with depressive disorders enables clinicians to clarify the time periods at which the children face difficulties. Copyright © 2015 Elsevier B.V. All rights reserved.
Parallel Computational Fluid Dynamics: Current Status and Future Requirements
NASA Technical Reports Server (NTRS)
Simon, Horst D.; VanDalsem, William R.; Dagum, Leonardo; Kutler, Paul (Technical Monitor)
1994-01-01
One or the key objectives of the Applied Research Branch in the Numerical Aerodynamic Simulation (NAS) Systems Division at NASA Allies Research Center is the accelerated introduction of highly parallel machines into a full operational environment. In this report we discuss the performance results obtained from the implementation of some computational fluid dynamics (CFD) applications on the Connection Machine CM-2 and the Intel iPSC/860. We summarize some of the experiences made so far with the parallel testbed machines at the NAS Applied Research Branch. Then we discuss the long term computational requirements for accomplishing some of the grand challenge problems in computational aerosciences. We argue that only massively parallel machines will be able to meet these grand challenge requirements, and we outline the computer science and algorithm research challenges ahead.
NASA Astrophysics Data System (ADS)
Yu, Leiming; Nina-Paravecino, Fanny; Kaeli, David; Fang, Qianqian
2018-01-01
We present a highly scalable Monte Carlo (MC) three-dimensional photon transport simulation platform designed for heterogeneous computing systems. Through the development of a massively parallel MC algorithm using the Open Computing Language framework, this research extends our existing graphics processing unit (GPU)-accelerated MC technique to a highly scalable vendor-independent heterogeneous computing environment, achieving significantly improved performance and software portability. A number of parallel computing techniques are investigated to achieve portable performance over a wide range of computing hardware. Furthermore, multiple thread-level and device-level load-balancing strategies are developed to obtain efficient simulations using multiple central processing units and GPUs.
Crane, Michael; Steinwand, Dan; Beckmann, Tim; Krpan, Greg; Liu, Shu-Guang; Nichols, Erin; Haga, Jim; Maddox, Brian; Bilderback, Chris; Feller, Mark; Homer, George
2001-01-01
The overarching goal of this project is to build a spatially distributed infrastructure for information science research by forming a team of information science researchers and providing them with similar hardware and software tools to perform collaborative research. Four geographically distributed Centers of the U.S. Geological Survey (USGS) are developing their own clusters of low-cost, personal computers into parallel computing environments that provide a costeffective way for the USGS to increase participation in the high-performance computing community. Referred to as Beowulf clusters, these hybrid systems provide the robust computing power required for conducting information science research into parallel computing systems and applications.
Computational Particle Dynamic Simulations on Multicore Processors (CPDMu) Final Report Phase I
DOE Office of Scientific and Technical Information (OSTI.GOV)
Schmalz, Mark S
2011-07-24
Statement of Problem - Department of Energy has many legacy codes for simulation of computational particle dynamics and computational fluid dynamics applications that are designed to run on sequential processors and are not easily parallelized. Emerging high-performance computing architectures employ massively parallel multicore architectures (e.g., graphics processing units) to increase throughput. Parallelization of legacy simulation codes is a high priority, to achieve compatibility, efficiency, accuracy, and extensibility. General Statement of Solution - A legacy simulation application designed for implementation on mainly-sequential processors has been represented as a graph G. Mathematical transformations, applied to G, produce a graph representation {und G}more » for a high-performance architecture. Key computational and data movement kernels of the application were analyzed/optimized for parallel execution using the mapping G {yields} {und G}, which can be performed semi-automatically. This approach is widely applicable to many types of high-performance computing systems, such as graphics processing units or clusters comprised of nodes that contain one or more such units. Phase I Accomplishments - Phase I research decomposed/profiled computational particle dynamics simulation code for rocket fuel combustion into low and high computational cost regions (respectively, mainly sequential and mainly parallel kernels), with analysis of space and time complexity. Using the research team's expertise in algorithm-to-architecture mappings, the high-cost kernels were transformed, parallelized, and implemented on Nvidia Fermi GPUs. Measured speedups (GPU with respect to single-core CPU) were approximately 20-32X for realistic model parameters, without final optimization. Error analysis showed no loss of computational accuracy. Commercial Applications and Other Benefits - The proposed research will constitute a breakthrough in solution of problems related to efficient parallel computation of particle and fluid dynamics simulations. These problems occur throughout DOE, military and commercial sectors: the potential payoff is high. We plan to license or sell the solution to contractors for military and domestic applications such as disaster simulation (aerodynamic and hydrodynamic), Government agencies (hydrological and environmental simulations), and medical applications (e.g., in tomographic image reconstruction). Keywords - High-performance Computing, Graphic Processing Unit, Fluid/Particle Simulation. Summary for Members of Congress - Department of Energy has many simulation codes that must compute faster, to be effective. The Phase I research parallelized particle/fluid simulations for rocket combustion, for high-performance computing systems.« less
NASA Astrophysics Data System (ADS)
Qin, Cheng-Zhi; Zhan, Lijun
2012-06-01
As one of the important tasks in digital terrain analysis, the calculation of flow accumulations from gridded digital elevation models (DEMs) usually involves two steps in a real application: (1) using an iterative DEM preprocessing algorithm to remove the depressions and flat areas commonly contained in real DEMs, and (2) using a recursive flow-direction algorithm to calculate the flow accumulation for every cell in the DEM. Because both algorithms are computationally intensive, quick calculation of the flow accumulations from a DEM (especially for a large area) presents a practical challenge to personal computer (PC) users. In recent years, rapid increases in hardware capacity of the graphics processing units (GPUs) provided in modern PCs have made it possible to meet this challenge in a PC environment. Parallel computing on GPUs using a compute-unified-device-architecture (CUDA) programming model has been explored to speed up the execution of the single-flow-direction algorithm (SFD). However, the parallel implementation on a GPU of the multiple-flow-direction (MFD) algorithm, which generally performs better than the SFD algorithm, has not been reported. Moreover, GPU-based parallelization of the DEM preprocessing step in the flow-accumulation calculations has not been addressed. This paper proposes a parallel approach to calculate flow accumulations (including both iterative DEM preprocessing and a recursive MFD algorithm) on a CUDA-compatible GPU. For the parallelization of an MFD algorithm (MFD-md), two different parallelization strategies using a GPU are explored. The first parallelization strategy, which has been used in the existing parallel SFD algorithm on GPU, has the problem of computing redundancy. Therefore, we designed a parallelization strategy based on graph theory. The application results show that the proposed parallel approach to calculate flow accumulations on a GPU performs much faster than either sequential algorithms or other parallel GPU-based algorithms based on existing parallelization strategies.
Parallelized Stochastic Cutoff Method for Long-Range Interacting Systems
NASA Astrophysics Data System (ADS)
Endo, Eishin; Toga, Yuta; Sasaki, Munetaka
2015-07-01
We present a method of parallelizing the stochastic cutoff (SCO) method, which is a Monte-Carlo method for long-range interacting systems. After interactions are eliminated by the SCO method, we subdivide a lattice into noninteracting interpenetrating sublattices. This subdivision enables us to parallelize the Monte-Carlo calculation in the SCO method. Such subdivision is found by numerically solving the vertex coloring of a graph created by the SCO method. We use an algorithm proposed by Kuhn and Wattenhofer to solve the vertex coloring by parallel computation. This method was applied to a two-dimensional magnetic dipolar system on an L × L square lattice to examine its parallelization efficiency. The result showed that, in the case of L = 2304, the speed of computation increased about 102 times by parallel computation with 288 processors.
Identifying failure in a tree network of a parallel computer
Archer, Charles J.; Pinnow, Kurt W.; Wallenfelt, Brian P.
2010-08-24
Methods, parallel computers, and products are provided for identifying failure in a tree network of a parallel computer. The parallel computer includes one or more processing sets including an I/O node and a plurality of compute nodes. For each processing set embodiments include selecting a set of test compute nodes, the test compute nodes being a subset of the compute nodes of the processing set; measuring the performance of the I/O node of the processing set; measuring the performance of the selected set of test compute nodes; calculating a current test value in dependence upon the measured performance of the I/O node of the processing set, the measured performance of the set of test compute nodes, and a predetermined value for I/O node performance; and comparing the current test value with a predetermined tree performance threshold. If the current test value is below the predetermined tree performance threshold, embodiments include selecting another set of test compute nodes. If the current test value is not below the predetermined tree performance threshold, embodiments include selecting from the test compute nodes one or more potential problem nodes and testing individually potential problem nodes and links to potential problem nodes.
Design of on-board parallel computer on nano-satellite
NASA Astrophysics Data System (ADS)
You, Zheng; Tian, Hexiang; Yu, Shijie; Meng, Li
2007-11-01
This paper provides one scheme of the on-board parallel computer system designed for the Nano-satellite. Based on the development request that the Nano-satellite should have a small volume, low weight, low power cost, and intelligence, this scheme gets rid of the traditional one-computer system and dual-computer system with endeavor to improve the dependability, capability and intelligence simultaneously. According to the method of integration design, it employs the parallel computer system with shared memory as the main structure, connects the telemetric system, attitude control system, and the payload system by the intelligent bus, designs the management which can deal with the static tasks and dynamic task-scheduling, protect and recover the on-site status and so forth in light of the parallel algorithms, and establishes the fault diagnosis, restoration and system restructure mechanism. It accomplishes an on-board parallel computer system with high dependability, capability and intelligence, a flexible management on hardware resources, an excellent software system, and a high ability in extension, which satisfies with the conception and the tendency of the integration electronic design sufficiently.
NASA Astrophysics Data System (ADS)
Neff, John A.
1989-12-01
Experiments originating from Gestalt psychology have shown that representing information in a symbolic form provides a more effective means to understanding. Computer scientists have been struggling for the last two decades to determine how best to create, manipulate, and store collections of symbolic structures. In the past, much of this struggling led to software innovations because that was the path of least resistance. For example, the development of heuristics for organizing the searching through knowledge bases was much less expensive than building massively parallel machines that could search in parallel. That is now beginning to change with the emergence of parallel architectures which are showing the potential for handling symbolic structures. This paper will review the relationships between symbolic computing and parallel computing architectures, and will identify opportunities for optics to significantly impact the performance of such computing machines. Although neural networks are an exciting subset of massively parallel computing structures, this paper will not touch on this area since it is receiving a great deal of attention in the literature. That is, the concepts presented herein do not consider the distributed representation of knowledge.
NASA Technical Reports Server (NTRS)
Crockett, Thomas W.
1995-01-01
This article provides a broad introduction to the subject of parallel rendering, encompassing both hardware and software systems. The focus is on the underlying concepts and the issues which arise in the design of parallel rendering algorithms and systems. We examine the different types of parallelism and how they can be applied in rendering applications. Concepts from parallel computing, such as data decomposition, task granularity, scalability, and load balancing, are considered in relation to the rendering problem. We also explore concepts from computer graphics, such as coherence and projection, which have a significant impact on the structure of parallel rendering algorithms. Our survey covers a number of practical considerations as well, including the choice of architectural platform, communication and memory requirements, and the problem of image assembly and display. We illustrate the discussion with numerous examples from the parallel rendering literature, representing most of the principal rendering methods currently used in computer graphics.
Enhancing PC Cluster-Based Parallel Branch-and-Bound Algorithms for the Graph Coloring Problem
NASA Astrophysics Data System (ADS)
Taoka, Satoshi; Takafuji, Daisuke; Watanabe, Toshimasa
A branch-and-bound algorithm (BB for short) is the most general technique to deal with various combinatorial optimization problems. Even if it is used, computation time is likely to increase exponentially. So we consider its parallelization to reduce it. It has been reported that the computation time of a parallel BB heavily depends upon node-variable selection strategies. And, in case of a parallel BB, it is also necessary to prevent increase in communication time. So, it is important to pay attention to how many and what kind of nodes are to be transferred (called sending-node selection strategy). In this paper, for the graph coloring problem, we propose some sending-node selection strategies for a parallel BB algorithm by adopting MPI for parallelization and experimentally evaluate how these strategies affect computation time of a parallel BB on a PC cluster network.
Bazavov, A; Ding, H-T; Hegde, P; Kaczmarek, O; Karsch, F; Laermann, E; Maezawa, Y; Mukherjee, Swagato; Ohno, H; Petreczky, P; Schmidt, C; Sharma, S; Soeldner, W; Wagner, M
2014-08-15
We compare lattice QCD results for appropriate combinations of net strangeness fluctuations and their correlations with net baryon number fluctuations with predictions from two hadron resonance gas (HRG) models having different strange hadron content. The conventionally used HRG model based on experimentally established strange hadrons fails to describe the lattice QCD results in the hadronic phase close to the QCD crossover. Supplementing the conventional HRG with additional, experimentally uncharted strange hadrons predicted by quark model calculations and observed in lattice QCD spectrum calculations leads to good descriptions of strange hadron thermodynamics below the QCD crossover. We show that the thermodynamic presence of these additional states gets imprinted in the yields of the ground-state strange hadrons leading to a systematic 5-8 MeV decrease of the chemical freeze-out temperatures of ground-state strange baryons.
Biocellion: accelerating computer simulation of multicellular biological system models.
Kang, Seunghwa; Kahan, Simon; McDermott, Jason; Flann, Nicholas; Shmulevich, Ilya
2014-11-01
Biological system behaviors are often the outcome of complex interactions among a large number of cells and their biotic and abiotic environment. Computational biologists attempt to understand, predict and manipulate biological system behavior through mathematical modeling and computer simulation. Discrete agent-based modeling (in combination with high-resolution grids to model the extracellular environment) is a popular approach for building biological system models. However, the computational complexity of this approach forces computational biologists to resort to coarser resolution approaches to simulate large biological systems. High-performance parallel computers have the potential to address the computing challenge, but writing efficient software for parallel computers is difficult and time-consuming. We have developed Biocellion, a high-performance software framework, to solve this computing challenge using parallel computers. To support a wide range of multicellular biological system models, Biocellion asks users to provide their model specifics by filling the function body of pre-defined model routines. Using Biocellion, modelers without parallel computing expertise can efficiently exploit parallel computers with less effort than writing sequential programs from scratch. We simulate cell sorting, microbial patterning and a bacterial system in soil aggregate as case studies. Biocellion runs on x86 compatible systems with the 64 bit Linux operating system and is freely available for academic use. Visit http://biocellion.com for additional information. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Modelling parallel programs and multiprocessor architectures with AXE
NASA Technical Reports Server (NTRS)
Yan, Jerry C.; Fineman, Charles E.
1991-01-01
AXE, An Experimental Environment for Parallel Systems, was designed to model and simulate for parallel systems at the process level. It provides an integrated environment for specifying computation models, multiprocessor architectures, data collection, and performance visualization. AXE is being used at NASA-Ames for developing resource management strategies, parallel problem formulation, multiprocessor architectures, and operating system issues related to the High Performance Computing and Communications Program. AXE's simple, structured user-interface enables the user to model parallel programs and machines precisely and efficiently. Its quick turn-around time keeps the user interested and productive. AXE models multicomputers. The user may easily modify various architectural parameters including the number of sites, connection topologies, and overhead for operating system activities. Parallel computations in AXE are represented as collections of autonomous computing objects known as players. Their use and behavior is described. Performance data of the multiprocessor model can be observed on a color screen. These include CPU and message routing bottlenecks, and the dynamic status of the software.
Efficient parallel resolution of the simplified transport equations in mixed-dual formulation
NASA Astrophysics Data System (ADS)
Barrault, M.; Lathuilière, B.; Ramet, P.; Roman, J.
2011-03-01
A reactivity computation consists of computing the highest eigenvalue of a generalized eigenvalue problem, for which an inverse power algorithm is commonly used. Very fine modelizations are difficult to treat for our sequential solver, based on the simplified transport equations, in terms of memory consumption and computational time. A first implementation of a Lagrangian based domain decomposition method brings to a poor parallel efficiency because of an increase in the power iterations [1]. In order to obtain a high parallel efficiency, we improve the parallelization scheme by changing the location of the loop over the subdomains in the overall algorithm and by benefiting from the characteristics of the Raviart-Thomas finite element. The new parallel algorithm still allows us to locally adapt the numerical scheme (mesh, finite element order). However, it can be significantly optimized for the matching grid case. The good behavior of the new parallelization scheme is demonstrated for the matching grid case on several hundreds of nodes for computations based on a pin-by-pin discretization.
Performing an allreduce operation on a plurality of compute nodes of a parallel computer
Faraj, Ahmad [Rochester, MN
2012-04-17
Methods, apparatus, and products are disclosed for performing an allreduce operation on a plurality of compute nodes of a parallel computer. Each compute node includes at least two processing cores. Each processing core has contribution data for the allreduce operation. Performing an allreduce operation on a plurality of compute nodes of a parallel computer includes: establishing one or more logical rings among the compute nodes, each logical ring including at least one processing core from each compute node; performing, for each logical ring, a global allreduce operation using the contribution data for the processing cores included in that logical ring, yielding a global allreduce result for each processing core included in that logical ring; and performing, for each compute node, a local allreduce operation using the global allreduce results for each processing core on that compute node.
Hybrid massively parallel fast sweeping method for static Hamilton-Jacobi equations
NASA Astrophysics Data System (ADS)
Detrixhe, Miles; Gibou, Frédéric
2016-10-01
The fast sweeping method is a popular algorithm for solving a variety of static Hamilton-Jacobi equations. Fast sweeping algorithms for parallel computing have been developed, but are severely limited. In this work, we present a multilevel, hybrid parallel algorithm that combines the desirable traits of two distinct parallel methods. The fine and coarse grained components of the algorithm take advantage of heterogeneous computer architecture common in high performance computing facilities. We present the algorithm and demonstrate its effectiveness on a set of example problems including optimal control, dynamic games, and seismic wave propagation. We give results for convergence, parallel scaling, and show state-of-the-art speedup values for the fast sweeping method.
Turbomachinery CFD on parallel computers
NASA Technical Reports Server (NTRS)
Blech, Richard A.; Milner, Edward J.; Quealy, Angela; Townsend, Scott E.
1992-01-01
The role of multistage turbomachinery simulation in the development of propulsion system models is discussed. Particularly, the need for simulations with higher fidelity and faster turnaround time is highlighted. It is shown how such fast simulations can be used in engineering-oriented environments. The use of parallel processing to achieve the required turnaround times is discussed. Current work by several researchers in this area is summarized. Parallel turbomachinery CFD research at the NASA Lewis Research Center is then highlighted. These efforts are focused on implementing the average-passage turbomachinery model on MIMD, distributed memory parallel computers. Performance results are given for inviscid, single blade row and viscous, multistage applications on several parallel computers, including networked workstations.
NASA Astrophysics Data System (ADS)
Newman, Gregory A.
2014-01-01
Many geoscientific applications exploit electrostatic and electromagnetic fields to interrogate and map subsurface electrical resistivity—an important geophysical attribute for characterizing mineral, energy, and water resources. In complex three-dimensional geologies, where many of these resources remain to be found, resistivity mapping requires large-scale modeling and imaging capabilities, as well as the ability to treat significant data volumes, which can easily overwhelm single-core and modest multicore computing hardware. To treat such problems requires large-scale parallel computational resources, necessary for reducing the time to solution to a time frame acceptable to the exploration process. The recognition that significant parallel computing processes must be brought to bear on these problems gives rise to choices that must be made in parallel computing hardware and software. In this review, some of these choices are presented, along with the resulting trade-offs. We also discuss future trends in high-performance computing and the anticipated impact on electromagnetic (EM) geophysics. Topics discussed in this review article include a survey of parallel computing platforms, graphics processing units to multicore CPUs with a fast interconnect, along with effective parallel solvers and associated solver libraries effective for inductive EM modeling and imaging.
Toward an automated parallel computing environment for geosciences
NASA Astrophysics Data System (ADS)
Zhang, Huai; Liu, Mian; Shi, Yaolin; Yuen, David A.; Yan, Zhenzhen; Liang, Guoping
2007-08-01
Software for geodynamic modeling has not kept up with the fast growing computing hardware and network resources. In the past decade supercomputing power has become available to most researchers in the form of affordable Beowulf clusters and other parallel computer platforms. However, to take full advantage of such computing power requires developing parallel algorithms and associated software, a task that is often too daunting for geoscience modelers whose main expertise is in geosciences. We introduce here an automated parallel computing environment built on open-source algorithms and libraries. Users interact with this computing environment by specifying the partial differential equations, solvers, and model-specific properties using an English-like modeling language in the input files. The system then automatically generates the finite element codes that can be run on distributed or shared memory parallel machines. This system is dynamic and flexible, allowing users to address different problems in geosciences. It is capable of providing web-based services, enabling users to generate source codes online. This unique feature will facilitate high-performance computing to be integrated with distributed data grids in the emerging cyber-infrastructures for geosciences. In this paper we discuss the principles of this automated modeling environment and provide examples to demonstrate its versatility.
Critical end point in the presence of a chiral chemical potential
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cui, Z. -F.; Cloët, I. C.; Lu, Y.
A class of Polyakov-loop-modified Nambu-Jona-Lasinio models has been used to support a conjecture that numerical simulations of lattice-regularized QCD defined with a chiral chemical potential can provide information about the existence and location of a critical end point in the QCD phase diagram drawn in the plane spanned by baryon chemical potential and temperature. That conjecture is challenged by conflicts between the model results and analyses of the same problem using simulations of lattice-regularized QCD (lQCD) and well-constrained Dyson-Schwinger equation (DSE) studies. We find the conflict is resolved in favor of the lQCD and DSE predictions when both a physicallymore » motivated regularization is employed to suppress the contribution of high-momentum quark modes in the definition of the effective potential connected with the Polyakov-loop-modified Nambu-Jona-Lasinio models and the four-fermion coupling in those models does not react strongly to changes in the mean field that is assumed to mock-up Polyakov-loop dynamics. With the lQCD and DSE predictions thus confirmed, it seems unlikely that simulations of lQCD with mu(5) > 0 can shed any light on a critical end point in the regular QCD phase diagram.« less
Computer architecture evaluation for structural dynamics computations: Project summary
NASA Technical Reports Server (NTRS)
Standley, Hilda M.
1989-01-01
The intent of the proposed effort is the examination of the impact of the elements of parallel architectures on the performance realized in a parallel computation. To this end, three major projects are developed: a language for the expression of high level parallelism, a statistical technique for the synthesis of multicomputer interconnection networks based upon performance prediction, and a queueing model for the analysis of shared memory hierarchies.
Multi-threading: A new dimension to massively parallel scientific computation
NASA Astrophysics Data System (ADS)
Nielsen, Ida M. B.; Janssen, Curtis L.
2000-06-01
Multi-threading is becoming widely available for Unix-like operating systems, and the application of multi-threading opens new ways for performing parallel computations with greater efficiency. We here briefly discuss the principles of multi-threading and illustrate the application of multi-threading for a massively parallel direct four-index transformation of electron repulsion integrals. Finally, other potential applications of multi-threading in scientific computing are outlined.
2008-02-09
Campbell, S. Ogata, and F. Shimojo, “ Multimillion atom simulations of nanosystems on parallel computers,” in Proceedings of the International...nanomesas: multimillion -atom molecular dynamics simulations on parallel computers,” J. Appl. Phys. 94, 6762 (2003). 21. P. Vashishta, R. K. Kalia...and A. Nakano, “ Multimillion atom molecular dynamics simulations of nanoparticles on parallel computers,” Journal of Nanoparticle Research 5, 119-135
Using parallel computing for the display and simulation of the space debris environment
NASA Astrophysics Data System (ADS)
Möckel, M.; Wiedemann, C.; Flegel, S.; Gelhaus, J.; Vörsmann, P.; Klinkrad, H.; Krag, H.
2011-07-01
Parallelism is becoming the leading paradigm in today's computer architectures. In order to take full advantage of this development, new algorithms have to be specifically designed for parallel execution while many old ones have to be upgraded accordingly. One field in which parallel computing has been firmly established for many years is computer graphics. Calculating and displaying three-dimensional computer generated imagery in real time requires complex numerical operations to be performed at high speed on a large number of objects. Since most of these objects can be processed independently, parallel computing is applicable in this field. Modern graphics processing units (GPUs) have become capable of performing millions of matrix and vector operations per second on multiple objects simultaneously. As a side project, a software tool is currently being developed at the Institute of Aerospace Systems that provides an animated, three-dimensional visualization of both actual and simulated space debris objects. Due to the nature of these objects it is possible to process them individually and independently from each other. Therefore, an analytical orbit propagation algorithm has been implemented to run on a GPU. By taking advantage of all its processing power a huge performance increase, compared to its CPU-based counterpart, could be achieved. For several years efforts have been made to harness this computing power for applications other than computer graphics. Software tools for the simulation of space debris are among those that could profit from embracing parallelism. With recently emerged software development tools such as OpenCL it is possible to transfer the new algorithms used in the visualization outside the field of computer graphics and implement them, for example, into the space debris simulation environment. This way they can make use of parallel hardware such as GPUs and Multi-Core-CPUs for faster computation. In this paper the visualization software will be introduced, including a comparison between the serial and the parallel method of orbit propagation. Ways of how to use the benefits of the latter method for space debris simulation will be discussed. An introduction to OpenCL will be given as well as an exemplary algorithm from the field of space debris simulation.
Using parallel computing for the display and simulation of the space debris environment
NASA Astrophysics Data System (ADS)
Moeckel, Marek; Wiedemann, Carsten; Flegel, Sven Kevin; Gelhaus, Johannes; Klinkrad, Heiner; Krag, Holger; Voersmann, Peter
Parallelism is becoming the leading paradigm in today's computer architectures. In order to take full advantage of this development, new algorithms have to be specifically designed for parallel execution while many old ones have to be upgraded accordingly. One field in which parallel computing has been firmly established for many years is computer graphics. Calculating and displaying three-dimensional computer generated imagery in real time requires complex numerical operations to be performed at high speed on a large number of objects. Since most of these objects can be processed independently, parallel computing is applicable in this field. Modern graphics processing units (GPUs) have become capable of performing millions of matrix and vector operations per second on multiple objects simultaneously. As a side project, a software tool is currently being developed at the Institute of Aerospace Systems that provides an animated, three-dimensional visualization of both actual and simulated space debris objects. Due to the nature of these objects it is possible to process them individually and independently from each other. Therefore, an analytical orbit propagation algorithm has been implemented to run on a GPU. By taking advantage of all its processing power a huge performance increase, compared to its CPU-based counterpart, could be achieved. For several years efforts have been made to harness this computing power for applications other than computer graphics. Software tools for the simulation of space debris are among those that could profit from embracing parallelism. With recently emerged software development tools such as OpenCL it is possible to transfer the new algorithms used in the visualization outside the field of computer graphics and implement them, for example, into the space debris simulation environment. This way they can make use of parallel hardware such as GPUs and Multi-Core-CPUs for faster computation. In this paper the visualization software will be introduced, including a comparison between the serial and the parallel method of orbit propagation. Ways of how to use the benefits of the latter method for space debris simulation will be discussed. An introduction of OpenCL will be given as well as an exemplary algorithm from the field of space debris simulation.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ma, Yan-Qing; Qiu, Jian-Wei
Following our previous proposal, we construct a class of good "lattice cross sections" (LCSs), from which we can study the partonic structure of hadrons from ab initio lattice QCD calculations. These good LCSs, on the one hand, can be calculated directly in lattice QCD, and on the other hand, can be factorized into parton distribution functions (PDFs) with calculable coefficients, in the same way as QCD factorization for factorizable hadronic cross sections. PDFs could be extracted from QCD global analysis of the lattice QCD generated data of LCSs. In conclusion, we also show that the proposed functions for lattice QCDmore » calculation of PDFs in the literature are special cases of these good LCSs.« less
The quark condensate in multi-flavour QCD – planar equivalence confronting lattice simulations
Armoni, Adi; Shifman, Mikhail; Shore, Graham; ...
2015-02-01
Planar equivalence between the large N limits of N=1 Super Yang–Mills (SYM) theory and a variant of QCD with fermions in the antisymmetric representation is a powerful tool to obtain analytic non-perturbative results in QCD itself. In particular, it allows the quark condensate for N=3 QCD with quarks in the fundamental representation to be inferred from exact calculations of the gluino condensate in N=1 SYM. In this paper, we review and refine our earlier predictions for the quark condensate in QCD with a general number nf of flavours and confront these with lattice results.
Quantum information, cognition, and music.
Dalla Chiara, Maria L; Giuntini, Roberto; Leporini, Roberto; Negri, Eleonora; Sergioli, Giuseppe
2015-01-01
Parallelism represents an essential aspect of human mind/brain activities. One can recognize some common features between psychological parallelism and the characteristic parallel structures that arise in quantum theory and in quantum computation. The article is devoted to a discussion of the following questions: a comparison between classical probabilistic Turing machines and quantum Turing machines.possible applications of the quantum computational semantics to cognitive problems.parallelism in music.
Quantum information, cognition, and music
Dalla Chiara, Maria L.; Giuntini, Roberto; Leporini, Roberto; Negri, Eleonora; Sergioli, Giuseppe
2015-01-01
Parallelism represents an essential aspect of human mind/brain activities. One can recognize some common features between psychological parallelism and the characteristic parallel structures that arise in quantum theory and in quantum computation. The article is devoted to a discussion of the following questions: a comparison between classical probabilistic Turing machines and quantum Turing machines.possible applications of the quantum computational semantics to cognitive problems.parallelism in music. PMID:26539139
QCD tests in $$p\\bar{p}$$ collisions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Huth, John E.; Mangano, Michelangelo L.
1993-02-01
We review the status of QCD tests in high energy p-pbar collisions. Contents: i) Introduction ii) QCD in Hadronic Collisions iii) Jet Production iv) Heavy Flavour Production v) W and Z Production vi) Direct Photons.
Johnson, Timothy C.; Versteeg, Roelof J.; Ward, Andy; Day-Lewis, Frederick D.; Revil, André
2010-01-01
Electrical geophysical methods have found wide use in the growing discipline of hydrogeophysics for characterizing the electrical properties of the subsurface and for monitoring subsurface processes in terms of the spatiotemporal changes in subsurface conductivity, chargeability, and source currents they govern. Presently, multichannel and multielectrode data collections systems can collect large data sets in relatively short periods of time. Practitioners, however, often are unable to fully utilize these large data sets and the information they contain because of standard desktop-computer processing limitations. These limitations can be addressed by utilizing the storage and processing capabilities of parallel computing environments. We have developed a parallel distributed-memory forward and inverse modeling algorithm for analyzing resistivity and time-domain induced polar-ization (IP) data. The primary components of the parallel computations include distributed computation of the pole solutions in forward mode, distributed storage and computation of the Jacobian matrix in inverse mode, and parallel execution of the inverse equation solver. We have tested the corresponding parallel code in three efforts: (1) resistivity characterization of the Hanford 300 Area Integrated Field Research Challenge site in Hanford, Washington, U.S.A., (2) resistivity characterization of a volcanic island in the southern Tyrrhenian Sea in Italy, and (3) resistivity and IP monitoring of biostimulation at a Superfund site in Brandywine, Maryland, U.S.A. Inverse analysis of each of these data sets would be limited or impossible in a standard serial computing environment, which underscores the need for parallel high-performance computing to fully utilize the potential of electrical geophysical methods in hydrogeophysical applications.
Topics in QCD at Nonzero Temperature and Density
NASA Astrophysics Data System (ADS)
Pangeni, Kamal
Understanding the behavior of matter at ultra-high density such as neutron stars require the knowledge of ground state properties of Quantum chromodynamics (QCD) at finite chemical potential. However, this task has turned out to be very difficult because of two main reasons: 1) QCD may still be strongly coupled at those regimes making perturbative calculations unreliable and 2) QCD at finite density suffers from the sign problem that makes the use of lattice simulation problematic and it even affects phenomenological models. In the first part of this thesis, we show that the sign problem in analytical calculations of finite density models can be solved by considering the CK-symmetric, where C is charge conjugation and K is complex conjugation, complex saddle points of the effective action. We then explore the properties and consequences of such complex saddle points at non-zero temperature and density. Due to CK symmetry, the mass matrix eigenvalues in these models are not always real but can be complex, which results in damped oscillation of the density-density correlation function, a new feature of finite density models. To address the generality of such behavior, we next consider a lattice model of QCD with static quarks at strong-coupling. Computation of the mass spectrum confirms the existence of complex eigenvalues in much of temperature-chemical potential plane. This provides an independent confirmation of our results obtained using phenomenological models of QCD. The existence of regions in parameter space where density-density correlation function exhibit damped oscillation is one of the hallmarks of typical liquid-gas system. The formalism developed to tackle the sign problem in QCD models actually gives a simple understanding for the existence of such behavior in liquid-gas system. To this end, we develop a generic field theoretic model for the treatment of liquid-gas phase transition. An effective field theory at finite density derived from a fundamental four dimensional field theory turns out to be complex but CK symmetric. The existence of CK symmetry results in complex mass eigenvalues, which in turn leads to damped oscillatory behavior of the density-density correlation function. In the last part of this thesis, we study the effect of large amplitude density oscillations on the transport properties of superfluid nuclear matter. In nuclear matter at neutron-star densities and temperature, Cooper pairing leads to the formations of a gap in the nucleon excitation spectra resulting in exponentially strong Boltzmann suppression of many transport coefficients. Previous calculations have shown evidence that density oscillations of sufficiently large amplitude can overcome this suppression for flavor-changing beta processes via the mechanism of "gap-bridging". We address the simplifications made in that initial work, and show that gap bridging can counteract Boltzmann suppression of neutrino emissivity for the realistic case of modified Urca processes in matter with 3 P2 neutron pairing.
Strangeness S =-1 hyperon-nucleon interactions: Chiral effective field theory versus lattice QCD
NASA Astrophysics Data System (ADS)
Song, Jing; Li, Kai-Wen; Geng, Li-Sheng
2018-06-01
Hyperon-nucleon interactions serve as basic inputs to studies of hypernuclear physics and dense (neutron) stars. Unfortunately, a precise understanding of these important quantities has lagged far behind that of the nucleon-nucleon interaction due to lack of high-precision experimental data. Historically, hyperon-nucleon interactions are either formulated in quark models or meson exchange models. In recent years, lattice QCD simulations and chiral effective field theory approaches start to offer new insights from first principles. In the present work, we contrast the state-of-the-art lattice QCD simulations with the latest chiral hyperon-nucleon forces and show that the leading order relativistic chiral results can already describe the lattice QCD data reasonably well. Given the fact that the lattice QCD simulations are performed with pion masses ranging from the (almost) physical point to 700 MeV, such studies provide a useful check on both the chiral effective field theory approaches as well as lattice QCD simulations. Nevertheless more precise lattice QCD simulations are eagerly needed to refine our understanding of hyperon-nucleon interactions.
Exclusive QCD processes, quark-hadron duality, and the transition to perturbative QCD
NASA Astrophysics Data System (ADS)
Corianò, Claudio; Li, Hsiang-nan; Savkli, Cetin
1998-07-01
Experiments at CEBAF will scan the intermediate-energy region of the QCD dynamics for the nucleon form factors and for Compton Scattering. These experiments will definitely clarify the role of resummed perturbation theory and of quark-hadron duality (QCD sum rules) in this regime. With this perspective in mind, we review the factorization theorem of perturbative QCD for exclusive processes at intermediate energy scales, which embodies the transverse degrees of freedom of a parton and the Sudakov resummation of the corresponding large logarithms. We concentrate on the pion and proton electromagnetic form factors and on pion Compton scattering. New ingredients, such as the evolution of the pion wave function and the complete two-loop expression of the Sudakov factor, are included. The sensitivity of our predictions to the infrared cutoff for the Sudakov evolution is discussed. We also elaborate on QCD sum rule methods for Compton Scattering, which provide an alternative description of this process. We show that, by comparing the local duality analysis to resummed perturbation theory, it is possible to describe the transition of exclusive processes to perturbative QCD.
Extension of the HAL QCD approach to inelastic and multi-particle scatterings in lattice QCD
NASA Astrophysics Data System (ADS)
Aoki, S.
We extend the HAL QCD approach, with which potentials between two hadrons can be obtained in QCD at energy below inelastic thresholds, to inelastic and multi-particle scatterings. We first derive asymptotic behaviors of the Nambu-Bethe-Salpeter (NBS) wave function at large space separations for systems with more than 2 particles, in terms of the one-shell $T$-matrix consrainted by the unitarity of quantum field theories. We show that its asymptotic behavior contains phase shifts and mixing angles of $n$ particle scatterings. This property is one of the essential ingredients of the HAL QCD scheme to define "potential" from the NBS wave function in quantum field theories such as QCD. We next construct energy independent but non-local potentials above inelastic thresholds, in terms of these NBS wave functions. We demonstrate an existence of energy-independent coupled channel potentials with a non-relativistic approximation, where momenta of all particles are small compared with their own masses. Combining these two results, we can employ the HAL QCD approach also to investigate inelastic and multi-particle scatterings.
Locating hardware faults in a data communications network of a parallel computer
Archer, Charles J.; Megerian, Mark G.; Ratterman, Joseph D.; Smith, Brian E.
2010-01-12
Hardware faults location in a data communications network of a parallel computer. Such a parallel computer includes a plurality of compute nodes and a data communications network that couples the compute nodes for data communications and organizes the compute node as a tree. Locating hardware faults includes identifying a next compute node as a parent node and a root of a parent test tree, identifying for each child compute node of the parent node a child test tree having the child compute node as root, running a same test suite on the parent test tree and each child test tree, and identifying the parent compute node as having a defective link connected from the parent compute node to a child compute node if the test suite fails on the parent test tree and succeeds on all the child test trees.
Parallel Architectures and Parallel Algorithms for Integrated Vision Systems. Ph.D. Thesis
NASA Technical Reports Server (NTRS)
Choudhary, Alok Nidhi
1989-01-01
Computer vision is regarded as one of the most complex and computationally intensive problems. An integrated vision system (IVS) is a system that uses vision algorithms from all levels of processing to perform for a high level application (e.g., object recognition). An IVS normally involves algorithms from low level, intermediate level, and high level vision. Designing parallel architectures for vision systems is of tremendous interest to researchers. Several issues are addressed in parallel architectures and parallel algorithms for integrated vision systems.
Parallel implementation of geometrical shock dynamics for two dimensional converging shock waves
NASA Astrophysics Data System (ADS)
Qiu, Shi; Liu, Kuang; Eliasson, Veronica
2016-10-01
Geometrical shock dynamics (GSD) theory is an appealing method to predict the shock motion in the sense that it is more computationally efficient than solving the traditional Euler equations, especially for converging shock waves. However, to solve and optimize large scale configurations, the main bottleneck is the computational cost. Among the existing numerical GSD schemes, there is only one that has been implemented on parallel computers, with the purpose to analyze detonation waves. To extend the computational advantage of the GSD theory to more general applications such as converging shock waves, a numerical implementation using a spatial decomposition method has been coupled with a front tracking approach on parallel computers. In addition, an efficient tridiagonal system solver for massively parallel computers has been applied to resolve the most expensive function in this implementation, resulting in an efficiency of 0.93 while using 32 HPCC cores. Moreover, symmetric boundary conditions have been developed to further reduce the computational cost, achieving a speedup of 19.26 for a 12-sided polygonal converging shock.
NASA Technical Reports Server (NTRS)
Saini, Subhash; Frumkin, Michael; Hribar, Michelle; Jin, Hao-Qiang; Waheed, Abdul; Yan, Jerry
1998-01-01
Porting applications to new high performance parallel and distributed computing platforms is a challenging task. Since writing parallel code by hand is extremely time consuming and costly, porting codes would ideally be automated by using some parallelization tools and compilers. In this paper, we compare the performance of the hand written NAB Parallel Benchmarks against three parallel versions generated with the help of tools and compilers: 1) CAPTools: an interactive computer aided parallelization too] that generates message passing code, 2) the Portland Group's HPF compiler and 3) using compiler directives with the native FORTAN77 compiler on the SGI Origin2000.
A transient FETI methodology for large-scale parallel implicit computations in structural mechanics
NASA Technical Reports Server (NTRS)
Farhat, Charbel; Crivelli, Luis; Roux, Francois-Xavier
1992-01-01
Explicit codes are often used to simulate the nonlinear dynamics of large-scale structural systems, even for low frequency response, because the storage and CPU requirements entailed by the repeated factorizations traditionally found in implicit codes rapidly overwhelm the available computing resources. With the advent of parallel processing, this trend is accelerating because explicit schemes are also easier to parallelize than implicit ones. However, the time step restriction imposed by the Courant stability condition on all explicit schemes cannot yet -- and perhaps will never -- be offset by the speed of parallel hardware. Therefore, it is essential to develop efficient and robust alternatives to direct methods that are also amenable to massively parallel processing because implicit codes using unconditionally stable time-integration algorithms are computationally more efficient when simulating low-frequency dynamics. Here we present a domain decomposition method for implicit schemes that requires significantly less storage than factorization algorithms, that is several times faster than other popular direct and iterative methods, that can be easily implemented on both shared and local memory parallel processors, and that is both computationally and communication-wise efficient. The proposed transient domain decomposition method is an extension of the method of Finite Element Tearing and Interconnecting (FETI) developed by Farhat and Roux for the solution of static problems. Serial and parallel performance results on the CRAY Y-MP/8 and the iPSC-860/128 systems are reported and analyzed for realistic structural dynamics problems. These results establish the superiority of the FETI method over both the serial/parallel conjugate gradient algorithm with diagonal scaling and the serial/parallel direct method, and contrast the computational power of the iPSC-860/128 parallel processor with that of the CRAY Y-MP/8 system.
Massively parallel sparse matrix function calculations with NTPoly
NASA Astrophysics Data System (ADS)
Dawson, William; Nakajima, Takahito
2018-04-01
We present NTPoly, a massively parallel library for computing the functions of sparse, symmetric matrices. The theory of matrix functions is a well developed framework with a wide range of applications including differential equations, graph theory, and electronic structure calculations. One particularly important application area is diagonalization free methods in quantum chemistry. When the input and output of the matrix function are sparse, methods based on polynomial expansions can be used to compute matrix functions in linear time. We present a library based on these methods that can compute a variety of matrix functions. Distributed memory parallelization is based on a communication avoiding sparse matrix multiplication algorithm. OpenMP task parallellization is utilized to implement hybrid parallelization. We describe NTPoly's interface and show how it can be integrated with programs written in many different programming languages. We demonstrate the merits of NTPoly by performing large scale calculations on the K computer.
NASA Technical Reports Server (NTRS)
Nguyen, D. T.; Watson, Willie R. (Technical Monitor)
2005-01-01
The overall objectives of this research work are to formulate and validate efficient parallel algorithms, and to efficiently design/implement computer software for solving large-scale acoustic problems, arised from the unified frameworks of the finite element procedures. The adopted parallel Finite Element (FE) Domain Decomposition (DD) procedures should fully take advantages of multiple processing capabilities offered by most modern high performance computing platforms for efficient parallel computation. To achieve this objective. the formulation needs to integrate efficient sparse (and dense) assembly techniques, hybrid (or mixed) direct and iterative equation solvers, proper pre-conditioned strategies, unrolling strategies, and effective processors' communicating schemes. Finally, the numerical performance of the developed parallel finite element procedures will be evaluated by solving series of structural, and acoustic (symmetrical and un-symmetrical) problems (in different computing platforms). Comparisons with existing "commercialized" and/or "public domain" software are also included, whenever possible.
Parallel, Asynchronous Executive (PAX): System concepts, facilities, and architecture
NASA Technical Reports Server (NTRS)
Jones, W. H.
1983-01-01
The Parallel, Asynchronous Executive (PAX) is a software operating system simulation that allows many computers to work on a single problem at the same time. PAX is currently implemented on a UNIVAC 1100/42 computer system. Independent UNIVAC runstreams are used to simulate independent computers. Data are shared among independent UNIVAC runstreams through shared mass-storage files. PAX has achieved the following: (1) applied several computing processes simultaneously to a single, logically unified problem; (2) resolved most parallel processor conflicts by careful work assignment; (3) resolved by means of worker requests to PAX all conflicts not resolved by work assignment; (4) provided fault isolation and recovery mechanisms to meet the problems of an actual parallel, asynchronous processing machine. Additionally, one real-life problem has been constructed for the PAX environment. This is CASPER, a collection of aerodynamic and structural dynamic problem simulation routines. CASPER is not discussed in this report except to provide examples of parallel-processing techniques.
Applications of Parallel Computation in Micro-Mechanics and Finite Element Method
NASA Technical Reports Server (NTRS)
Tan, Hui-Qian
1996-01-01
This project discusses the application of parallel computations related with respect to material analyses. Briefly speaking, we analyze some kind of material by elements computations. We call an element a cell here. A cell is divided into a number of subelements called subcells and all subcells in a cell have the identical structure. The detailed structure will be given later in this paper. It is obvious that the problem is "well-structured". SIMD machine would be a better choice. In this paper we try to look into the potentials of SIMD machine in dealing with finite element computation by developing appropriate algorithms on MasPar, a SIMD parallel machine. In section 2, the architecture of MasPar will be discussed. A brief review of the parallel programming language MPL also is given in that section. In section 3, some general parallel algorithms which might be useful to the project will be proposed. And, combining with the algorithms, some features of MPL will be discussed in more detail. In section 4, the computational structure of cell/subcell model will be given. The idea of designing the parallel algorithm for the model will be demonstrated. Finally in section 5, a summary will be given.
Optics Program Modified for Multithreaded Parallel Computing
NASA Technical Reports Server (NTRS)
Lou, John; Bedding, Dave; Basinger, Scott
2006-01-01
A powerful high-performance computer program for simulating and analyzing adaptive and controlled optical systems has been developed by modifying the serial version of the Modeling and Analysis for Controlled Optical Systems (MACOS) program to impart capabilities for multithreaded parallel processing on computing systems ranging from supercomputers down to Symmetric Multiprocessing (SMP) personal computers. The modifications included the incorporation of OpenMP, a portable and widely supported application interface software, that can be used to explicitly add multithreaded parallelism to an application program under a shared-memory programming model. OpenMP was applied to parallelize ray-tracing calculations, one of the major computing components in MACOS. Multithreading is also used in the diffraction propagation of light in MACOS based on pthreads [POSIX Thread, (where "POSIX" signifies a portable operating system for UNIX)]. In tests of the parallelized version of MACOS, the speedup in ray-tracing calculations was found to be linear, or proportional to the number of processors, while the speedup in diffraction calculations ranged from 50 to 60 percent, depending on the type and number of processors. The parallelized version of MACOS is portable, and, to the user, its interface is basically the same as that of the original serial version of MACOS.
NASA Technical Reports Server (NTRS)
Hockney, George; Lee, Seungwon
2008-01-01
A computer program known as PyPele, originally written as a Pythonlanguage extension module of a C++ language program, has been rewritten in pure Python language. The original version of PyPele dispatches and coordinates parallel-processing tasks on cluster computers and provides a conceptual framework for spacecraft-mission- design and -analysis software tools to run in an embarrassingly parallel mode. The original version of PyPele uses SSH (Secure Shell a set of standards and an associated network protocol for establishing a secure channel between a local and a remote computer) to coordinate parallel processing. Instead of SSH, the present Python version of PyPele uses Message Passing Interface (MPI) [an unofficial de-facto standard language-independent application programming interface for message- passing on a parallel computer] while keeping the same user interface. The use of MPI instead of SSH and the preservation of the original PyPele user interface make it possible for parallel application programs written previously for the original version of PyPele to run on MPI-based cluster computers. As a result, engineers using the previously written application programs can take advantage of embarrassing parallelism without need to rewrite those programs.
Massively parallel information processing systems for space applications
NASA Technical Reports Server (NTRS)
Schaefer, D. H.
1979-01-01
NASA is developing massively parallel systems for ultra high speed processing of digital image data collected by satellite borne instrumentation. Such systems contain thousands of processing elements. Work is underway on the design and fabrication of the 'Massively Parallel Processor', a ground computer containing 16,384 processing elements arranged in a 128 x 128 array. This computer uses existing technology. Advanced work includes the development of semiconductor chips containing thousands of feedthrough paths. Massively parallel image analog to digital conversion technology is also being developed. The goal is to provide compact computers suitable for real-time onboard processing of images.
n-body simulations using message passing parallel computers.
NASA Astrophysics Data System (ADS)
Grama, A. Y.; Kumar, V.; Sameh, A.
The authors present new parallel formulations of the Barnes-Hut method for n-body simulations on message passing computers. These parallel formulations partition the domain efficiently incurring minimal communication overhead. This is in contrast to existing schemes that are based on sorting a large number of keys or on the use of global data structures. The new formulations are augmented by alternate communication strategies which serve to minimize communication overhead. The impact of these communication strategies is experimentally studied. The authors report on experimental results obtained from an astrophysical simulation on an nCUBE2 parallel computer.
A design methodology for portable software on parallel computers
NASA Technical Reports Server (NTRS)
Nicol, David M.; Miller, Keith W.; Chrisman, Dan A.
1993-01-01
This final report for research that was supported by grant number NAG-1-995 documents our progress in addressing two difficulties in parallel programming. The first difficulty is developing software that will execute quickly on a parallel computer. The second difficulty is transporting software between dissimilar parallel computers. In general, we expect that more hardware-specific information will be included in software designs for parallel computers than in designs for sequential computers. This inclusion is an instance of portability being sacrificed for high performance. New parallel computers are being introduced frequently. Trying to keep one's software on the current high performance hardware, a software developer almost continually faces yet another expensive software transportation. The problem of the proposed research is to create a design methodology that helps designers to more precisely control both portability and hardware-specific programming details. The proposed research emphasizes programming for scientific applications. We completed our study of the parallelizability of a subsystem of the NASA Earth Radiation Budget Experiment (ERBE) data processing system. This work is summarized in section two. A more detailed description is provided in Appendix A ('Programming Practices to Support Eventual Parallelism'). Mr. Chrisman, a graduate student, wrote and successfully defended a Ph.D. dissertation proposal which describes our research associated with the issues of software portability and high performance. The list of research tasks are specified in the proposal. The proposal 'A Design Methodology for Portable Software on Parallel Computers' is summarized in section three and is provided in its entirety in Appendix B. We are currently studying a proposed subsystem of the NASA Clouds and the Earth's Radiant Energy System (CERES) data processing system. This software is the proof-of-concept for the Ph.D. dissertation. We have implemented and measured the performance of a portion of this subsystem on the Intel iPSC/2 parallel computer. These results are provided in section four. Our future work is summarized in section five, our acknowledgements are stated in section six, and references for published papers associated with NAG-1-995 are provided in section seven.
Convergence issues in domain decomposition parallel computation of hovering rotor
NASA Astrophysics Data System (ADS)
Xiao, Zhongyun; Liu, Gang; Mou, Bin; Jiang, Xiong
2018-05-01
Implicit LU-SGS time integration algorithm has been widely used in parallel computation in spite of its lack of information from adjacent domains. When applied to parallel computation of hovering rotor flows in a rotating frame, it brings about convergence issues. To remedy the problem, three LU factorization-based implicit schemes (consisting of LU-SGS, DP-LUR and HLU-SGS) are investigated comparatively. A test case of pure grid rotation is designed to verify these algorithms, which show that LU-SGS algorithm introduces errors on boundary cells. When partition boundaries are circumferential, errors arise in proportion to grid speed, accumulating along with the rotation, and leading to computational failure in the end. Meanwhile, DP-LUR and HLU-SGS methods show good convergence owing to boundary treatment which are desirable in domain decomposition parallel computations.
Voss, Andreas; Leonhart, Rainer; Stahl, Christoph
2007-11-01
Psychological research is based in large parts on response latencies, which are often registered by keypresses on a standard computer keyboard. Recording response latencies with a standard keyboard is problematic because keypresses are buffered within the keyboard hardware before they are signaled to the computer, adding error variance to the recorded latencies. This can be circumvented by using external response pads connected to the computer's parallel port. In this article, we describe how to build inexpensive, reliable, and easy-to-use response pads with six keys from two standard computer mice that can be connected to the PC's parallel port. We also address the problem of recording data from the parallel port with different software packages under Microsoft's Windows XP.
NASA Astrophysics Data System (ADS)
Zimovets, Artem; Matviychuk, Alexander; Ushakov, Vladimir
2016-12-01
The paper presents two different approaches to reduce the time of computer calculation of reachability sets. First of these two approaches use different data structures for storing the reachability sets in the computer memory for calculation in single-threaded mode. Second approach is based on using parallel algorithms with reference to the data structures from the first approach. Within the framework of this paper parallel algorithm of approximate reachability set calculation on computer with SMP-architecture is proposed. The results of numerical modelling are presented in the form of tables which demonstrate high efficiency of parallel computing technology and also show how computing time depends on the used data structure.
Performance analysis of parallel branch and bound search with the hypercube architecture
NASA Technical Reports Server (NTRS)
Mraz, Richard T.
1987-01-01
With the availability of commercial parallel computers, researchers are examining new classes of problems which might benefit from parallel computing. This paper presents results of an investigation of the class of search intensive problems. The specific problem discussed is the Least-Cost Branch and Bound search method of deadline job scheduling. The object-oriented design methodology was used to map the problem into a parallel solution. While the initial design was good for a prototype, the best performance resulted from fine-tuning the algorithm for a specific computer. The experiments analyze the computation time, the speed up over a VAX 11/785, and the load balance of the problem when using loosely coupled multiprocessor system based on the hypercube architecture.
Dynamic modeling of parallel robots for computed-torque control implementation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Codourey, A.
1998-12-01
In recent years, increased interest in parallel robots has been observed. Their control with modern theory, such as the computed-torque method, has, however, been restrained, essentially due to the difficulty in establishing a simple dynamic model that can be calculated in real time. In this paper, a simple method based on the virtual work principle is proposed for modeling parallel robots. The mass matrix of the robot, needed for decoupling control strategies, does not explicitly appear in the formulation; however, it can be computed separately, based on kinetic energy considerations. The method is applied to the DELTA parallel robot, leadingmore » to a very efficient model that has been implemented in a real-time computed-torque control algorithm.« less
Architecture-Adaptive Computing Environment: A Tool for Teaching Parallel Programming
NASA Technical Reports Server (NTRS)
Dorband, John E.; Aburdene, Maurice F.
2002-01-01
Recently, networked and cluster computation have become very popular. This paper is an introduction to a new C based parallel language for architecture-adaptive programming, aCe C. The primary purpose of aCe (Architecture-adaptive Computing Environment) is to encourage programmers to implement applications on parallel architectures by providing them the assurance that future architectures will be able to run their applications with a minimum of modification. A secondary purpose is to encourage computer architects to develop new types of architectures by providing an easily implemented software development environment and a library of test applications. This new language should be an ideal tool to teach parallel programming. In this paper, we will focus on some fundamental features of aCe C.
Portability and Cross-Platform Performance of an MPI-Based Parallel Polygon Renderer
NASA Technical Reports Server (NTRS)
Crockett, Thomas W.
1999-01-01
Visualizing the results of computations performed on large-scale parallel computers is a challenging problem, due to the size of the datasets involved. One approach is to perform the visualization and graphics operations in place, exploiting the available parallelism to obtain the necessary rendering performance. Over the past several years, we have been developing algorithms and software to support visualization applications on NASA's parallel supercomputers. Our results have been incorporated into a parallel polygon rendering system called PGL. PGL was initially developed on tightly-coupled distributed-memory message-passing systems, including Intel's iPSC/860 and Paragon, and IBM's SP2. Over the past year, we have ported it to a variety of additional platforms, including the HP Exemplar, SGI Origin2OOO, Cray T3E, and clusters of Sun workstations. In implementing PGL, we have had two primary goals: cross-platform portability and high performance. Portability is important because (1) our manpower resources are limited, making it difficult to develop and maintain multiple versions of the code, and (2) NASA's complement of parallel computing platforms is diverse and subject to frequent change. Performance is important in delivering adequate rendering rates for complex scenes and ensuring that parallel computing resources are used effectively. Unfortunately, these two goals are often at odds. In this paper we report on our experiences with portability and performance of the PGL polygon renderer across a range of parallel computing platforms.
A compositional reservoir simulator on distributed memory parallel computers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rame, M.; Delshad, M.
1995-12-31
This paper presents the application of distributed memory parallel computes to field scale reservoir simulations using a parallel version of UTCHEM, The University of Texas Chemical Flooding Simulator. The model is a general purpose highly vectorized chemical compositional simulator that can simulate a wide range of displacement processes at both field and laboratory scales. The original simulator was modified to run on both distributed memory parallel machines (Intel iPSC/960 and Delta, Connection Machine 5, Kendall Square 1 and 2, and CRAY T3D) and a cluster of workstations. A domain decomposition approach has been taken towards parallelization of the code. Amore » portion of the discrete reservoir model is assigned to each processor by a set-up routine that attempts a data layout as even as possible from the load-balance standpoint. Each of these subdomains is extended so that data can be shared between adjacent processors for stencil computation. The added routines that make parallel execution possible are written in a modular fashion that makes the porting to new parallel platforms straight forward. Results of the distributed memory computing performance of Parallel simulator are presented for field scale applications such as tracer flood and polymer flood. A comparison of the wall-clock times for same problems on a vector supercomputer is also presented.« less
Hybrid massively parallel fast sweeping method for static Hamilton–Jacobi equations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Detrixhe, Miles, E-mail: mdetrixhe@engineering.ucsb.edu; University of California Santa Barbara, Santa Barbara, CA, 93106; Gibou, Frédéric, E-mail: fgibou@engineering.ucsb.edu
The fast sweeping method is a popular algorithm for solving a variety of static Hamilton–Jacobi equations. Fast sweeping algorithms for parallel computing have been developed, but are severely limited. In this work, we present a multilevel, hybrid parallel algorithm that combines the desirable traits of two distinct parallel methods. The fine and coarse grained components of the algorithm take advantage of heterogeneous computer architecture common in high performance computing facilities. We present the algorithm and demonstrate its effectiveness on a set of example problems including optimal control, dynamic games, and seismic wave propagation. We give results for convergence, parallel scaling,more » and show state-of-the-art speedup values for the fast sweeping method.« less
Hybrid parallel computing architecture for multiview phase shifting
NASA Astrophysics Data System (ADS)
Zhong, Kai; Li, Zhongwei; Zhou, Xiaohui; Shi, Yusheng; Wang, Congjun
2014-11-01
The multiview phase-shifting method shows its powerful capability in achieving high resolution three-dimensional (3-D) shape measurement. Unfortunately, this ability results in very high computation costs and 3-D computations have to be processed offline. To realize real-time 3-D shape measurement, a hybrid parallel computing architecture is proposed for multiview phase shifting. In this architecture, the central processing unit can co-operate with the graphic processing unit (GPU) to achieve hybrid parallel computing. The high computation cost procedures, including lens distortion rectification, phase computation, correspondence, and 3-D reconstruction, are implemented in GPU, and a three-layer kernel function model is designed to simultaneously realize coarse-grained and fine-grained paralleling computing. Experimental results verify that the developed system can perform 50 fps (frame per second) real-time 3-D measurement with 260 K 3-D points per frame. A speedup of up to 180 times is obtained for the performance of the proposed technique using a NVIDIA GT560Ti graphics card rather than a sequential C in a 3.4 GHZ Inter Core i7 3770.
Resonant conversions of QCD axions into hidden axions and suppressed isocurvature perturbations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kitajima, Naoya; Takahashi, Fuminobu, E-mail: kitajima@tuhep.phys.tohoku.ac.jp, E-mail: fumi@tuhep.phys.tohoku.ac.jp
2015-01-01
We study in detail MSW-like resonant conversions of QCD axions into hidden axions, including cases where the adiabaticity condition is only marginally satisfied, and where anharmonic effects are non-negligible. When the resonant conversion is efficient, the QCD axion abundance is suppressed by the hidden and QCD axion mass ratio. We find that, when the resonant conversion is incomplete due to a weak violation of the adiabaticity, the CDM isocurvature perturbations can be significantly suppressed, while non-Gaussianity of the isocurvature perturbations generically remain unsuppressed. The isocurvature bounds on the inflation scale can therefore be relaxed by the partial resonant conversion ofmore » the QCD axions into hidden axions.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dine, Michael; Draper, Patrick; Stephenson-Haskins, Laurel
Here, we study the large N θ dependence and the η' potential in supersymmetric QCD with small soft SUSY-breaking terms. Known exact results in SUSY QCD are found to reflect a variety of expectations from large N perturbation theory, including the presence of branches and the behavior of theories with matter (both with N f << N and N f ~ N ). But, there are also striking departures from ordinary QCD and the conventional large N description: instanton effects, when under control, are not exponentially suppressed at large N , and branched structure in supersymmetric QCD is always associatedmore » with approximate discrete symmetries. We suggest that these differences motivate further study of large N QCD on the lattice.« less
Update on ɛK with lattice QCD inputs
NASA Astrophysics Data System (ADS)
Jang, Yong-Chull; Lee, Weonjong; Lee, Sunkyu; Leem, Jaehoon
2018-03-01
We report updated results for ɛK, the indirect CP violation parameter in neutral kaons, which is evaluated directly from the standard model with lattice QCD inputs. We use lattice QCD inputs to fix B\\hatk,|Vcb|,ξ0,ξ2,|Vus|, and mc(mc). Since Lattice 2016, the UTfit group has updated the Wolfenstein parameters in the angle-only-fit method, and the HFLAV group has also updated |Vcb|. Our results show that the evaluation of ɛK with exclusive |Vcb| (lattice QCD inputs) has 4.0σ tension with the experimental value, while that with inclusive |Vcb| (heavy quark expansion based on OPE and QCD sum rules) shows no tension.
Scheme Variations of the QCD Coupling and Hadronic τ Decays
NASA Astrophysics Data System (ADS)
Boito, Diogo; Jamin, Matthias; Miravitllas, Ramon
2016-10-01
The quantum chromodynamics (QCD) coupling αs is not a physical observable of the theory, since it depends on conventions related to the renormalization procedure. We introduce a definition of the QCD coupling, denoted by α^s, whose running is explicitly renormalization scheme invariant. The scheme dependence of the new coupling α^s is parametrized by a single parameter C , related to transformations of the QCD scale Λ . It is demonstrated that appropriate choices of C can lead to substantial improvements in the perturbative prediction of physical observables. As phenomenological applications, we study e+e- scattering and decays of the τ lepton into hadrons, both being governed by the QCD Adler function.
Spectroscopic parameters and decays of the resonance Z_b(10610)
NASA Astrophysics Data System (ADS)
Agaev, S. S.; Azizi, K.; Sundu, H.
2017-12-01
The resonance Z_b(10610) is investigated as the diquark-antidiquark Z_b=[bu][\\overline{bd}] state with spin-parity JP=1+. The mass and current coupling of the resonance Z_b(10610) are evaluated using QCD two-point sum rule and taking into account the vacuum condensates up to ten dimensions. We study the vertices Z_bΥ (nS)π (n=1,2,3) by applying the QCD light-cone sum rule to compute the corresponding strong couplings g_{Z_bΥ (nS)π } and widths of the decays Z_b → Υ (nS)π . We explore also the vertices Z_b hb(mP)π (m=1,2) and calculate the couplings g_{Z_b hb(mP)π } and the widths of the decay channels Z_b → hb(mP)π . To this end, we calculate the mass and decay constants of the h_b(1P) and h_b(2P) mesons. The results obtained are compared with experimental data of the Belle Collaboration.
Neutron and proton electric dipole moments from N f=2+1 domain-wall fermion lattice QCD
Shintani, Eigo; Blum, Thomas; Izubuchi, Taku; ...
2016-05-05
We present a lattice calculation of the neutron and proton electric dipole moments (EDM’s) with N f = 2 + 1 flavors of domain-wall fermions. The neutron and proton EDM form factors are extracted from three-point functions at the next-to-leading order in the θ vacuum of QCD. In this computation, we use pion masses 330 and 420 MeV and 2.7 fm 3 lattices with Iwasaki gauge action and a 170 MeV pion and 4.6 fm 3 lattice with I-DSDR gauge action, all generated by the RBC and UKQCD collaborations. The all-mode-averaging technique enables an efficient, high statistics calculation; however themore » statistical errors on our results are still relatively large, so we investigate a new direction to reduce them, reweighting with the local topological charge density which appears promising. Furthermore, we discuss the chiral behavior and finite size effects of the EDM’s in the context of baryon chiral perturbation theory.« less
NASA Astrophysics Data System (ADS)
Bergner, Georg; Piemonte, Stefano
2018-04-01
Non-Abelian gauge theories with fermions transforming in the adjoint representation of the gauge group (AdjQCD) are a fundamental ingredient of many models that describe the physics beyond the Standard Model. Two relevant examples are N =1 supersymmetric Yang-Mills (SYM) theory and minimal walking technicolor, which are gauge theories coupled to one adjoint Majorana and two adjoint Dirac fermions, respectively. While confinement is a property of N =1 SYM, minimal walking technicolor is expected to be infrared conformal. We study the propagators of ghost and gluon fields in the Landau gauge to compute the running coupling in the MiniMom scheme. We analyze several different ensembles of lattice Monte Carlo simulations for the SU(2) adjoint QCD with Nf=1 /2 ,1 ,3 /2 , and 2 Dirac fermions. We show how the running of the coupling changes as the number of interacting fermions is increased towards the conformal window.
Complete NLO corrections to W+W+ scattering and its irreducible background at the LHC
NASA Astrophysics Data System (ADS)
Biedermann, Benedikt; Denner, Ansgar; Pellen, Mathieu
2017-10-01
The process pp → μ +ν μ e+νejj receives several contributions of different orders in the strong and electroweak coupling constants. Using appropriate event selections, this process is dominated by vector-boson scattering (VBS) and has recently been measured at the LHC. It is thus of prime importance to estimate precisely each contribution. In this article we compute for the first time the full NLO QCD and electroweak corrections to VBS and its irreducible background processes with realistic experimental cuts. We do not rely on approximations but use complete amplitudes involving two different orders at tree level and three different orders at one-loop level. Since we take into account all interferences, at NLO level the corrections to the VBS process and to the QCD-induced irreducible background process contribute at the same orders. Hence the two processes cannot be unambiguously distinguished, and all contributions to the μ +ν μ e+νejj final state should be preferably measured together.
Baikov, P A; Chetyrkin, K G; Kühn, J H
2006-01-13
We compute, for the first time, the absorptive part of the massless correlator of two quark scalar currents in five loops. As physical applications, we consider the [symbol: see text](alpha(s)4) corrections to the decay rate of the standard model Higgs boson into quarks, as well as the constraints on the strange quark mass following from QCD sum rules.
NASA Astrophysics Data System (ADS)
Valasek, Lukas; Glasa, Jan
2017-12-01
Current fire simulation systems are capable to utilize advantages of high-performance computer (HPC) platforms available and to model fires efficiently in parallel. In this paper, efficiency of a corridor fire simulation on a HPC computer cluster is discussed. The parallel MPI version of Fire Dynamics Simulator is used for testing efficiency of selected strategies of allocation of computational resources of the cluster using a greater number of computational cores. Simulation results indicate that if the number of cores used is not equal to a multiple of the total number of cluster node cores there are allocation strategies which provide more efficient calculations.
Scattering processes and resonances from lattice QCD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Briceno, Raul A.; Dudek, Jozef J.; Young, Ross D.
The vast majority of hadrons observed in nature are not stable under the strong interaction; rather they are resonances whose existence is deduced from enhancements in the energy dependence of scattering amplitudes. The study of hadron resonances offers a window into the workings of quantum chromodynamics (QCD) in the low-energy nonperturbative region, and in addition many probes of the limits of the electroweak sector of the standard model consider processes which feature hadron resonances. From a theoretical standpoint, this is a challenging field: the same dynamics that binds quarks and gluons into hadron resonances also controls their decay into lightermore » hadrons, so a complete approach to QCD is required. Presently, lattice QCD is the only available tool that provides the required nonperturbative evaluation of hadron observables. This paper reviews progress in the study of few-hadron reactions in which resonances and bound states appear using lattice QCD techniques. The leading approach is described that takes advantage of the periodic finite spatial volume used in lattice QCD calculations to extract scattering amplitudes from the discrete spectrum of QCD eigenstates in a box. An explanation is given of how from explicit lattice QCD calculations one can rigorously garner information about a variety of resonance properties, including their masses, widths, decay couplings, and form factors. Finally, the challenges which currently limit the field are discussed along with the steps being taken to resolve them.« less
Scattering processes and resonances from lattice QCD
NASA Astrophysics Data System (ADS)
Briceño, Raúl A.; Dudek, Jozef J.; Young, Ross D.
2018-04-01
The vast majority of hadrons observed in nature are not stable under the strong interaction; rather they are resonances whose existence is deduced from enhancements in the energy dependence of scattering amplitudes. The study of hadron resonances offers a window into the workings of quantum chromodynamics (QCD) in the low-energy nonperturbative region, and in addition many probes of the limits of the electroweak sector of the standard model consider processes which feature hadron resonances. From a theoretical standpoint, this is a challenging field: the same dynamics that binds quarks and gluons into hadron resonances also controls their decay into lighter hadrons, so a complete approach to QCD is required. Presently, lattice QCD is the only available tool that provides the required nonperturbative evaluation of hadron observables. This article reviews progress in the study of few-hadron reactions in which resonances and bound states appear using lattice QCD techniques. The leading approach is described that takes advantage of the periodic finite spatial volume used in lattice QCD calculations to extract scattering amplitudes from the discrete spectrum of QCD eigenstates in a box. An explanation is given of how from explicit lattice QCD calculations one can rigorously garner information about a variety of resonance properties, including their masses, widths, decay couplings, and form factors. The challenges which currently limit the field are discussed along with the steps being taken to resolve them.
Scattering processes and resonances from lattice QCD
Briceno, Raul A.; Dudek, Jozef J.; Young, Ross D.
2018-04-18
The vast majority of hadrons observed in nature are not stable under the strong interaction; rather they are resonances whose existence is deduced from enhancements in the energy dependence of scattering amplitudes. The study of hadron resonances offers a window into the workings of quantum chromodynamics (QCD) in the low-energy nonperturbative region, and in addition many probes of the limits of the electroweak sector of the standard model consider processes which feature hadron resonances. From a theoretical standpoint, this is a challenging field: the same dynamics that binds quarks and gluons into hadron resonances also controls their decay into lightermore » hadrons, so a complete approach to QCD is required. Presently, lattice QCD is the only available tool that provides the required nonperturbative evaluation of hadron observables. This paper reviews progress in the study of few-hadron reactions in which resonances and bound states appear using lattice QCD techniques. The leading approach is described that takes advantage of the periodic finite spatial volume used in lattice QCD calculations to extract scattering amplitudes from the discrete spectrum of QCD eigenstates in a box. An explanation is given of how from explicit lattice QCD calculations one can rigorously garner information about a variety of resonance properties, including their masses, widths, decay couplings, and form factors. Finally, the challenges which currently limit the field are discussed along with the steps being taken to resolve them.« less
A Hybrid MPI/OpenMP Approach for Parallel Groundwater Model Calibration on Multicore Computers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tang, Guoping; D'Azevedo, Ed F; Zhang, Fan
2010-01-01
Groundwater model calibration is becoming increasingly computationally time intensive. We describe a hybrid MPI/OpenMP approach to exploit two levels of parallelism in software and hardware to reduce calibration time on multicore computers with minimal parallelization effort. At first, HydroGeoChem 5.0 (HGC5) is parallelized using OpenMP for a uranium transport model with over a hundred species involving nearly a hundred reactions, and a field scale coupled flow and transport model. In the first application, a single parallelizable loop is identified to consume over 97% of the total computational time. With a few lines of OpenMP compiler directives inserted into the code,more » the computational time reduces about ten times on a compute node with 16 cores. The performance is further improved by selectively parallelizing a few more loops. For the field scale application, parallelizable loops in 15 of the 174 subroutines in HGC5 are identified to take more than 99% of the execution time. By adding the preconditioned conjugate gradient solver and BICGSTAB, and using a coloring scheme to separate the elements, nodes, and boundary sides, the subroutines for finite element assembly, soil property update, and boundary condition application are parallelized, resulting in a speedup of about 10 on a 16-core compute node. The Levenberg-Marquardt (LM) algorithm is added into HGC5 with the Jacobian calculation and lambda search parallelized using MPI. With this hybrid approach, compute nodes at the number of adjustable parameters (when the forward difference is used for Jacobian approximation), or twice that number (if the center difference is used), are used to reduce the calibration time from days and weeks to a few hours for the two applications. This approach can be extended to global optimization scheme and Monte Carol analysis where thousands of compute nodes can be efficiently utilized.« less
Variable-Complexity Multidisciplinary Optimization on Parallel Computers
NASA Technical Reports Server (NTRS)
Grossman, Bernard; Mason, William H.; Watson, Layne T.; Haftka, Raphael T.
1998-01-01
This report covers work conducted under grant NAG1-1562 for the NASA High Performance Computing and Communications Program (HPCCP) from December 7, 1993, to December 31, 1997. The objective of the research was to develop new multidisciplinary design optimization (MDO) techniques which exploit parallel computing to reduce the computational burden of aircraft MDO. The design of the High-Speed Civil Transport (HSCT) air-craft was selected as a test case to demonstrate the utility of our MDO methods. The three major tasks of this research grant included: development of parallel multipoint approximation methods for the aerodynamic design of the HSCT, use of parallel multipoint approximation methods for structural optimization of the HSCT, mathematical and algorithmic development including support in the integration of parallel computation for items (1) and (2). These tasks have been accomplished with the development of a response surface methodology that incorporates multi-fidelity models. For the aerodynamic design we were able to optimize with up to 20 design variables using hundreds of expensive Euler analyses together with thousands of inexpensive linear theory simulations. We have thereby demonstrated the application of CFD to a large aerodynamic design problem. For the predicting structural weight we were able to combine hundreds of structural optimizations of refined finite element models with thousands of optimizations based on coarse models. Computations have been carried out on the Intel Paragon with up to 128 nodes. The parallel computation allowed us to perform combined aerodynamic-structural optimization using state of the art models of a complex aircraft configurations.
Line-plane broadcasting in a data communications network of a parallel computer
Archer, Charles J.; Berg, Jeremy E.; Blocksome, Michael A.; Smith, Brian E.
2010-06-08
Methods, apparatus, and products are disclosed for line-plane broadcasting in a data communications network of a parallel computer, the parallel computer comprising a plurality of compute nodes connected together through the network, the network optimized for point to point data communications and characterized by at least a first dimension, a second dimension, and a third dimension, that include: initiating, by a broadcasting compute node, a broadcast operation, including sending a message to all of the compute nodes along an axis of the first dimension for the network; sending, by each compute node along the axis of the first dimension, the message to all of the compute nodes along an axis of the second dimension for the network; and sending, by each compute node along the axis of the second dimension, the message to all of the compute nodes along an axis of the third dimension for the network.
Line-plane broadcasting in a data communications network of a parallel computer
Archer, Charles J.; Berg, Jeremy E.; Blocksome, Michael A.; Smith, Brian E.
2010-11-23
Methods, apparatus, and products are disclosed for line-plane broadcasting in a data communications network of a parallel computer, the parallel computer comprising a plurality of compute nodes connected together through the network, the network optimized for point to point data communications and characterized by at least a first dimension, a second dimension, and a third dimension, that include: initiating, by a broadcasting compute node, a broadcast operation, including sending a message to all of the compute nodes along an axis of the first dimension for the network; sending, by each compute node along the axis of the first dimension, the message to all of the compute nodes along an axis of the second dimension for the network; and sending, by each compute node along the axis of the second dimension, the message to all of the compute nodes along an axis of the third dimension for the network.
NLO cross sections in 4 dimensions without DREG
NASA Astrophysics Data System (ADS)
Hernández-Pinto, R. J.; Driencourt-Mangin, F.; Rodrigo, G.; Sborlini, G. F. R.
2016-10-01
In this review, we present a new method for computing physical cross sections at NLO accuracy in QCD without using the standard Dimensional Regularisation. The algorithm is based on the Loop-Tree Duality theorem, which allow us to obtain loop integrals as a sum of phase-space integrals; in this way, transforming loop integrals into phase-space integrals, we propose a method to merge virtual and real contributions in order to find observables at NLO in d = 4 space-time dimensions. In addition, the strategy described is used for computing the γ* → qq̅(g) process. A more detailed discussion related on this topic can be found in Ref [1].