Violante-Carvalho, Nelson
2005-12-01
Synthetic Aperture Radar (SAR) onboard satellites is the only source of directional wave spectra with continuous and global coverage. Millions of SAR Wave Mode (SWM) imagettes have been acquired since the launch in the early 1990's of the first European Remote Sensing Satellite ERS-1 and its successors ERS-2 and ENVISAT, which has opened up many possibilities specially for wave data assimilation purposes. The main aim of data assimilation is to improve the forecasting introducing available observations into the modeling procedures in order to minimize the differences between model estimates and measurements. However there are limitations in the retrieval of the directional spectrum from SAR images due to nonlinearities in the mapping mechanism. The Max-Planck Institut (MPI) scheme, the first proposed and most widely used algorithm to retrieve directional wave spectra from SAR images, is employed to compare significant wave heights retrieved from ERS-1 SAR against buoy measurements and against the WAM wave model. It is shown that for periods shorter than 12 seconds the WAM model performs better than the MPI, despite the fact that the model is used as first guess to the MPI method, that is the retrieval is deteriorating the first guess. For periods longer than 12 seconds, the part of the spectrum that is directly measured by SAR, the performance of the MPI scheme is at least as good as the WAM model.
Institute for Defense Analysis. Annual Report 1995.
1995-01-01
staff have been involved in the community-wide development of MPI as well as in its application to specific NSA problems. 35 Parallel Groebner ...Basis Code — Symbolic Computing on Parallel Machines The Groebner basis method is a set of algorithms for reformulating very complex algebraic expres
A Log-Scaling Fault Tolerant Agreement Algorithm for a Fault Tolerant MPI
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hursey, Joshua J; Naughton, III, Thomas J; Vallee, Geoffroy R
The lack of fault tolerance is becoming a limiting factor for application scalability in HPC systems. The MPI does not provide standardized fault tolerance interfaces and semantics. The MPI Forum's Fault Tolerance Working Group is proposing a collective fault tolerant agreement algorithm for the next MPI standard. Such algorithms play a central role in many fault tolerant applications. This paper combines a log-scaling two-phase commit agreement algorithm with a reduction operation to provide the necessary functionality for the new collective without any additional messages. Error handling mechanisms are described that preserve the fault tolerance properties while maintaining overall scalability.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Amestoy, Patrick R.; Duff, Iain S.; L'Excellent, Jean-Yves
2001-10-10
We examine the mechanics of the send and receive mechanism of MPI and in particular how we can implement message passing in a robust way so that our performance is not significantly affected by changes to the MPI system. This leads us to using the Isend/Irecv protocol which will entail sometimes significant algorithmic changes. We discuss this within the context of two different algorithms for sparse Gaussian elimination that we have parallelized. One is a multifrontal solver called MUMPS, the other is a supernodal solver called SuperLU. Both algorithms are difficult to parallelize on distributed memory machines. Our initial strategiesmore » were based on simple MPI point-to-point communication primitives. With such approaches, the parallel performance of both codes are very sensitive to the MPI implementation, the way MPI internal buffers are used in particular. We then modified our codes to use more sophisticated nonblocking versions of MPI communication. This significantly improved the performance robustness (independent of the MPI buffering mechanism) and scalability, but at the cost of increased code complexity.« less
What does fault tolerant Deep Learning need from MPI?
DOE Office of Scientific and Technical Information (OSTI.GOV)
Amatya, Vinay C.; Vishnu, Abhinav; Siegel, Charles M.
Deep Learning (DL) algorithms have become the {\\em de facto} Machine Learning (ML) algorithm for large scale data analysis. DL algorithms are computationally expensive -- even distributed DL implementations which use MPI require days of training (model learning) time on commonly studied datasets. Long running DL applications become susceptible to faults -- requiring development of a fault tolerant system infrastructure, in addition to fault tolerant DL algorithms. This raises an important question: {\\em What is needed from MPI for designing fault tolerant DL implementations?} In this paper, we address this problem for permanent faults. We motivate the need for amore » fault tolerant MPI specification by an in-depth consideration of recent innovations in DL algorithms and their properties, which drive the need for specific fault tolerance features. We present an in-depth discussion on the suitability of different parallelism types (model, data and hybrid); a need (or lack thereof) for check-pointing of any critical data structures; and most importantly, consideration for several fault tolerance proposals (user-level fault mitigation (ULFM), Reinit) in MPI and their applicability to fault tolerant DL implementations. We leverage a distributed memory implementation of Caffe, currently available under the Machine Learning Toolkit for Extreme Scale (MaTEx). We implement our approaches by extending MaTEx-Caffe for using ULFM-based implementation. Our evaluation using the ImageNet dataset and AlexNet neural network topology demonstrates the effectiveness of the proposed fault tolerant DL implementation using OpenMPI based ULFM.« less
Accelerating k-NN Algorithm with Hybrid MPI and OpenSHMEM
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lin, Jian; Hamidouche, Khaled; Zheng, Jie
2015-08-05
Machine Learning algorithms are benefiting from the continuous improvement of programming models, including MPI, MapReduce and PGAS. k-Nearest Neighbors (k-NN) algorithm is a widely used machine learning algorithm, applied to supervised learning tasks such as classification. Several parallel implementations of k-NN have been proposed in the literature and practice. However, on high-performance computing systems with high-speed interconnects, it is important to further accelerate existing designs of the k-NN algorithm through taking advantage of scalable programming models. To improve the performance of k-NN on large-scale environment with InfiniBand network, this paper proposes several alternative hybrid MPI+OpenSHMEM designs and performs a systemicmore » evaluation and analysis on typical workloads. The hybrid designs leverage the one-sided memory access to better overlap communication with computation than the existing pure MPI design, and propose better schemes for efficient buffer management. The implementation based on k-NN program from MaTEx with MVAPICH2-X (Unified MPI+PGAS Communication Runtime over InfiniBand) shows up to 9.0% time reduction for training KDD Cup 2010 workload over 512 cores, and 27.6% time reduction for small workload with balanced communication and computation. Experiments of running with varied number of cores show that our design can maintain good scalability.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Katti, Amogh; Di Fatta, Giuseppe; Naughton III, Thomas J
Future extreme-scale high-performance computing systems will be required to work under frequent component failures. The MPI Forum's User Level Failure Mitigation proposal has introduced an operation, MPI_Comm_shrink, to synchronize the alive processes on the list of failed processes, so that applications can continue to execute even in the presence of failures by adopting algorithm-based fault tolerance techniques. This MPI_Comm_shrink operation requires a fault tolerant failure detection and consensus algorithm. This paper presents and compares two novel failure detection and consensus algorithms. The proposed algorithms are based on Gossip protocols and are inherently fault-tolerant and scalable. The proposed algorithms were implementedmore » and tested using the Extreme-scale Simulator. The results show that in both algorithms the number of Gossip cycles to achieve global consensus scales logarithmically with system size. The second algorithm also shows better scalability in terms of memory and network bandwidth usage and a perfect synchronization in achieving global consensus.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Katti, Amogh; Di Fatta, Giuseppe; Naughton, Thomas
Future extreme-scale high-performance computing systems will be required to work under frequent component failures. The MPI Forum s User Level Failure Mitigation proposal has introduced an operation, MPI Comm shrink, to synchronize the alive processes on the list of failed processes, so that applications can continue to execute even in the presence of failures by adopting algorithm-based fault tolerance techniques. This MPI Comm shrink operation requires a failure detection and consensus algorithm. This paper presents three novel failure detection and consensus algorithms using Gossiping. The proposed algorithms were implemented and tested using the Extreme-scale Simulator. The results show that inmore » all algorithms the number of Gossip cycles to achieve global consensus scales logarithmically with system size. The second algorithm also shows better scalability in terms of memory and network bandwidth usage and a perfect synchronization in achieving global consensus. The third approach is a three-phase distributed failure detection and consensus algorithm and provides consistency guarantees even in very large and extreme-scale systems while at the same time being memory and bandwidth efficient.« less
Accelerating free breathing myocardial perfusion MRI using multi coil radial k-t SLR
Lingala, Sajan Goud; DiBella, Edward; Adluru, Ganesh; McGann, Christopher; Jacob, Mathews
2013-01-01
The clinical utility of myocardial perfusion MR imaging (MPI) is often restricted by the inability of current acquisition schemes to simultaneously achieve high spatio-temporal resolution, good volume coverage, and high signal to noise ratio. Moreover, many subjects often find it difficult to hold their breath for sufficiently long durations making it difficult to obtain reliable MPI data. Accelerated acquisition of free breathing MPI data can overcome some of these challenges. Recently, an algorithm termed as k − t SLR has been proposed to accelerate dynamic MRI by exploiting sparsity and low rank properties of dynamic MRI data. The main focus of this paper is to further improve k − t SLR and demonstrate its utility in considerably accelerating free breathing MPI. We extend its previous implementation to account for multi-coil radial MPI acquisitions. We perform k − t sampling experiments to compare different radial trajectories and determine the best sampling pattern. We also introduce a novel augmented Lagrangian framework to considerably improve the algorithm's convergence rate. The proposed algorithm is validated using free breathing rest and stress radial perfusion data sets from two normal subjects and one patient with ischemia. k − t SLR was observed to provide faithful reconstructions at high acceleration levels with minimal artifacts compared to existing MPI acceleration schemes such as spatio-temporal constrained reconstruction (STCR) and k − t SPARSE/SENSE. PMID:24077063
Epidemic failure detection and consensus for extreme parallelism
Katti, Amogh; Di Fatta, Giuseppe; Naughton, Thomas; ...
2017-02-01
Future extreme-scale high-performance computing systems will be required to work under frequent component failures. The MPI Forum s User Level Failure Mitigation proposal has introduced an operation, MPI Comm shrink, to synchronize the alive processes on the list of failed processes, so that applications can continue to execute even in the presence of failures by adopting algorithm-based fault tolerance techniques. This MPI Comm shrink operation requires a failure detection and consensus algorithm. This paper presents three novel failure detection and consensus algorithms using Gossiping. The proposed algorithms were implemented and tested using the Extreme-scale Simulator. The results show that inmore » all algorithms the number of Gossip cycles to achieve global consensus scales logarithmically with system size. The second algorithm also shows better scalability in terms of memory and network bandwidth usage and a perfect synchronization in achieving global consensus. The third approach is a three-phase distributed failure detection and consensus algorithm and provides consistency guarantees even in very large and extreme-scale systems while at the same time being memory and bandwidth efficient.« less
Conflict Detection Algorithm to Minimize Locking for MPI-IO Atomicity
NASA Astrophysics Data System (ADS)
Sehrish, Saba; Wang, Jun; Thakur, Rajeev
Many scientific applications require high-performance concurrent I/O accesses to a file by multiple processes. Those applications rely indirectly on atomic I/O capabilities in order to perform updates to structured datasets, such as those stored in HDF5 format files. Current support for atomicity in MPI-IO is provided by locking around the operations, imposing lock overhead in all situations, even though in many cases these operations are non-overlapping in the file. We propose to isolate non-overlapping accesses from overlapping ones in independent I/O cases, allowing the non-overlapping ones to proceed without imposing lock overhead. To enable this, we have implemented an efficient conflict detection algorithm in MPI-IO using MPI file views and datatypes. We show that our conflict detection scheme incurs minimal overhead on I/O operations, making it an effective mechanism for avoiding locks when they are not needed.
Calibration free beam hardening correction for cardiac CT perfusion imaging
NASA Astrophysics Data System (ADS)
Levi, Jacob; Fahmi, Rachid; Eck, Brendan L.; Fares, Anas; Wu, Hao; Vembar, Mani; Dhanantwari, Amar; Bezerra, Hiram G.; Wilson, David L.
2016-03-01
Myocardial perfusion imaging using CT (MPI-CT) and coronary CTA have the potential to make CT an ideal noninvasive gate-keeper for invasive coronary angiography. However, beam hardening artifacts (BHA) prevent accurate blood flow calculation in MPI-CT. BH Correction (BHC) methods require either energy-sensitive CT, not widely available, or typically a calibration-based method. We developed a calibration-free, automatic BHC (ABHC) method suitable for MPI-CT. The algorithm works with any BHC method and iteratively determines model parameters using proposed BHA-specific cost function. In this work, we use the polynomial BHC extended to three materials. The image is segmented into soft tissue, bone, and iodine images, based on mean HU and temporal enhancement. Forward projections of bone and iodine images are obtained, and in each iteration polynomial correction is applied. Corrections are then back projected and combined to obtain the current iteration's BHC image. This process is iterated until cost is minimized. We evaluate the algorithm on simulated and physical phantom images and on preclinical MPI-CT data. The scans were obtained on a prototype spectral detector CT (SDCT) scanner (Philips Healthcare). Mono-energetic reconstructed images were used as the reference. In the simulated phantom, BH streak artifacts were reduced from 12+/-2HU to 1+/-1HU and cupping was reduced by 81%. Similarly, in physical phantom, BH streak artifacts were reduced from 48+/-6HU to 1+/-5HU and cupping was reduced by 86%. In preclinical MPI-CT images, BHA was reduced from 28+/-6 HU to less than 4+/-4HU at peak enhancement. Results suggest that the algorithm can be used to reduce BHA in conventional CT and improve MPI-CT accuracy.
An MPI + $X$ implementation of contact global search using Kokkos
Hansen, Glen A.; Xavier, Patrick G.; Mish, Sam P.; ...
2015-10-05
This paper describes an approach that seeks to parallelize the spatial search associated with computational contact mechanics. In contact mechanics, the purpose of the spatial search is to find “nearest neighbors,” which is the prelude to an imprinting search that resolves the interactions between the external surfaces of contacting bodies. In particular, we are interested in the contact global search portion of the spatial search associated with this operation on domain-decomposition-based meshes. Specifically, we describe an implementation that combines standard domain-decomposition-based MPI-parallel spatial search with thread-level parallelism (MPI-X) available on advanced computer architectures (those with GPU coprocessors). Our goal ismore » to demonstrate the efficacy of the MPI-X paradigm in the overall contact search. Standard MPI-parallel implementations typically use a domain decomposition of the external surfaces of bodies within the domain in an attempt to efficiently distribute computational work. This decomposition may or may not be the same as the volume decomposition associated with the host physics. The parallel contact global search phase is then employed to find and distribute surface entities (nodes and faces) that are needed to compute contact constraints between entities owned by different MPI ranks without further inter-rank communication. Key steps of the contact global search include computing bounding boxes, building surface entity (node and face) search trees and finding and distributing entities required to complete on-rank (local) spatial searches. To enable source-code portability and performance across a variety of different computer architectures, we implemented the algorithm using the Kokkos hardware abstraction library. While we targeted development towards machines with a GPU accelerator per MPI rank, we also report performance results for OpenMP with a conventional multi-core compute node per rank. Results here demonstrate a 47 % decrease in the time spent within the global search algorithm, comparing the reference ACME algorithm with the GPU implementation, on an 18M face problem using four MPI ranks. As a result, while further work remains to maximize performance on the GPU, this result illustrates the potential of the proposed implementation.« less
Accelerating free breathing myocardial perfusion MRI using multi coil radial k - t SLR
NASA Astrophysics Data System (ADS)
Goud Lingala, Sajan; DiBella, Edward; Adluru, Ganesh; McGann, Christopher; Jacob, Mathews
2013-10-01
The clinical utility of myocardial perfusion MR imaging (MPI) is often restricted by the inability of current acquisition schemes to simultaneously achieve high spatio-temporal resolution, good volume coverage, and high signal to noise ratio. Moreover, many subjects often find it difficult to hold their breath for sufficiently long durations making it difficult to obtain reliable MPI data. Accelerated acquisition of free breathing MPI data can overcome some of these challenges. Recently, an algorithm termed as k - t SLR has been proposed to accelerate dynamic MRI by exploiting sparsity and low rank properties of dynamic MRI data. The main focus of this paper is to further improve k - t SLR and demonstrate its utility in considerably accelerating free breathing MPI. We extend its previous implementation to account for multi-coil radial MPI acquisitions. We perform k - t sampling experiments to compare different radial trajectories and determine the best sampling pattern. We also introduce a novel augmented Lagrangian framework to considerably improve the algorithm’s convergence rate. The proposed algorithm is validated using free breathing rest and stress radial perfusion data sets from two normal subjects and one patient with ischemia. k - t SLR was observed to provide faithful reconstructions at high acceleration levels with minimal artifacts compared to existing MPI acceleration schemes such as spatio-temporal constrained reconstruction and k - t SPARSE/SENSE.
Geometry planning and image registration in magnetic particle imaging using bimodal fiducial markers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Werner, F., E-mail: f.werner@uke.de; Hofmann, M.; Them, K.
Purpose: Magnetic particle imaging (MPI) is a quantitative imaging modality that allows the distribution of superparamagnetic nanoparticles to be visualized. Compared to other imaging techniques like x-ray radiography, computed tomography (CT), and magnetic resonance imaging (MRI), MPI only provides a signal from the administered tracer, but no additional morphological information, which complicates geometry planning and the interpretation of MP images. The purpose of the authors’ study was to develop bimodal fiducial markers that can be visualized by MPI and MRI in order to create MP–MR fusion images. Methods: A certain arrangement of three bimodal fiducial markers was developed and usedmore » in a combined MRI/MPI phantom and also during in vivo experiments in order to investigate its suitability for geometry planning and image fusion. An algorithm for automated marker extraction in both MR and MP images and rigid registration was established. Results: The developed bimodal fiducial markers can be visualized by MRI and MPI and allow for geometry planning as well as automated registration and fusion of MR–MP images. Conclusions: To date, exact positioning of the object to be imaged within the field of view (FOV) and the assignment of reconstructed MPI signals to corresponding morphological regions has been difficult. The developed bimodal fiducial markers and the automated image registration algorithm help to overcome these difficulties.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shamis, Pavel; Graham, Richard L; Gorentla Venkata, Manjunath
The scalability and performance of collective communication operations limit the scalability and performance of many scientific applications. This paper presents two new blocking and nonblocking Broadcast algorithms for communicators with arbitrary communication topology, and studies their performance. These algorithms benefit from increased concurrency and a reduced memory footprint, making them suitable for use on large-scale systems. Measuring small, medium, and large data Broadcasts on a Cray-XT5, using 24,576 MPI processes, the Cheetah algorithms outperform the native MPI on that system by 51%, 69%, and 9%, respectively, at the same process count. These results demonstrate an algorithmic approach to the implementationmore » of the important class of collective communications, which is high performing, scalable, and also uses resources in a scalable manner.« less
NASA Astrophysics Data System (ADS)
Nguyen, An Hung; Guillemette, Thomas; Lambert, Andrew J.; Pickering, Mark R.; Garratt, Matthew A.
2017-09-01
Image registration is a fundamental image processing technique. It is used to spatially align two or more images that have been captured at different times, from different sensors, or from different viewpoints. There have been many algorithms proposed for this task. The most common of these being the well-known Lucas-Kanade (LK) and Horn-Schunck approaches. However, the main limitation of these approaches is the computational complexity required to implement the large number of iterations necessary for successful alignment of the images. Previously, a multi-pass image interpolation algorithm (MP-I2A) was developed to considerably reduce the number of iterations required for successful registration compared with the LK algorithm. This paper develops a kernel-warping algorithm (KWA), a modified version of the MP-I2A, which requires fewer iterations to successfully register two images and less memory space for the field-programmable gate array (FPGA) implementation than the MP-I2A. These reductions increase feasibility of the implementation of the proposed algorithm on FPGAs with very limited memory space and other hardware resources. A two-FPGA system rather than single FPGA system is successfully developed to implement the KWA in order to compensate insufficiency of hardware resources supported by one FPGA, and increase parallel processing ability and scalability of the system.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Murase, Kenya, E-mail: murase@sahs.med.osaka-u.ac.jp; Song, Ruixiao; Hiratsuka, Samu
We investigated the feasibility of visualizing blood coagulation using a system for magnetic particle imaging (MPI). A magnetic field-free line is generated using two opposing neodymium magnets and transverse images are reconstructed from the third-harmonic signals received by a gradiometer coil, using the maximum likelihood-expectation maximization algorithm. Our MPI system was used to image the blood coagulation induced by adding CaCl{sub 2} to whole sheep blood mixed with magnetic nanoparticles (MNPs). The “MPI value” was defined as the pixel value of the transverse image reconstructed from the third-harmonic signals. MPI values were significantly smaller for coagulated blood samples than thosemore » without coagulation. We confirmed the rationale of these results by calculating the third-harmonic signals for the measured viscosities of samples, with an assumption that the magnetization and particle size distribution of MNPs obey the Langevin equation and log-normal distribution, respectively. We concluded that MPI can be useful for visualizing blood coagulation.« less
Facilitating Co-Design for Extreme-Scale Systems Through Lightweight Simulation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Engelmann, Christian; Lauer, Frank
This work focuses on tools for investigating algorithm performance at extreme scale with millions of concurrent threads and for evaluating the impact of future architecture choices to facilitate the co-design of high-performance computing (HPC) architectures and applications. The approach focuses on lightweight simulation of extreme-scale HPC systems with the needed amount of accuracy. The prototype presented in this paper is able to provide this capability using a parallel discrete event simulation (PDES), such that a Message Passing Interface (MPI) application can be executed at extreme scale, and its performance properties can be evaluated. The results of an initial prototype aremore » encouraging as a simple 'hello world' MPI program could be scaled up to 1,048,576 virtual MPI processes on a four-node cluster, and the performance properties of two MPI programs could be evaluated at up to 16,384 virtual MPI processes on the same system.« less
A Convex Formulation for Magnetic Particle Imaging X-Space Reconstruction.
Konkle, Justin J; Goodwill, Patrick W; Hensley, Daniel W; Orendorff, Ryan D; Lustig, Michael; Conolly, Steven M
2015-01-01
Magnetic Particle Imaging (mpi) is an emerging imaging modality with exceptional promise for clinical applications in rapid angiography, cell therapy tracking, cancer imaging, and inflammation imaging. Recent publications have demonstrated quantitative mpi across rat sized fields of view with x-space reconstruction methods. Critical to any medical imaging technology is the reliability and accuracy of image reconstruction. Because the average value of the mpi signal is lost during direct-feedthrough signal filtering, mpi reconstruction algorithms must recover this zero-frequency value. Prior x-space mpi recovery techniques were limited to 1d approaches which could introduce artifacts when reconstructing a 3d image. In this paper, we formulate x-space reconstruction as a 3d convex optimization problem and apply robust a priori knowledge of image smoothness and non-negativity to reduce non-physical banding and haze artifacts. We conclude with a discussion of the powerful extensibility of the presented formulation for future applications.
Advances in Patch-Based Adaptive Mesh Refinement Scalability
Gunney, Brian T.N.; Anderson, Robert W.
2015-12-18
Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extensionmore » of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.« less
Advances in Patch-Based Adaptive Mesh Refinement Scalability
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gunney, Brian T.N.; Anderson, Robert W.
Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extensionmore » of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.« less
MPI-FAUN: An MPI-Based Framework for Alternating-Updating Nonnegative Matrix Factorization
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kannan, Ramakrishnan; Ballard, Grey; Park, Haesun
Non-negative matrix factorization (NMF) is the problem of determining two non-negative low rank factors W and H, for the given input matrix A, such that A≈WH. NMF is a useful tool for many applications in different domains such as topic modeling in text mining, background separation in video analysis, and community detection in social networks. Despite its popularity in the data mining community, there is a lack of efficient parallel algorithms to solve the problem for big data sets. The main contribution of this work is a new, high-performance parallel computational framework for a broad class of NMF algorithms thatmore » iteratively solves alternating non-negative least squares (NLS) subproblems for W and H. It maintains the data and factor matrices in memory (distributed across processors), uses MPI for interprocessor communication, and, in the dense case, provably minimizes communication costs (under mild assumptions). The framework is flexible and able to leverage a variety of NMF and NLS algorithms, including Multiplicative Update, Hierarchical Alternating Least Squares, and Block Principal Pivoting. Our implementation allows us to benchmark and compare different algorithms on massive dense and sparse data matrices of size that spans from few hundreds of millions to billions. We demonstrate the scalability of our algorithm and compare it with baseline implementations, showing significant performance improvements. The code and the datasets used for conducting the experiments are available online.« less
MPI-FAUN: An MPI-Based Framework for Alternating-Updating Nonnegative Matrix Factorization
Kannan, Ramakrishnan; Ballard, Grey; Park, Haesun
2017-10-30
Non-negative matrix factorization (NMF) is the problem of determining two non-negative low rank factors W and H, for the given input matrix A, such that A≈WH. NMF is a useful tool for many applications in different domains such as topic modeling in text mining, background separation in video analysis, and community detection in social networks. Despite its popularity in the data mining community, there is a lack of efficient parallel algorithms to solve the problem for big data sets. The main contribution of this work is a new, high-performance parallel computational framework for a broad class of NMF algorithms thatmore » iteratively solves alternating non-negative least squares (NLS) subproblems for W and H. It maintains the data and factor matrices in memory (distributed across processors), uses MPI for interprocessor communication, and, in the dense case, provably minimizes communication costs (under mild assumptions). The framework is flexible and able to leverage a variety of NMF and NLS algorithms, including Multiplicative Update, Hierarchical Alternating Least Squares, and Block Principal Pivoting. Our implementation allows us to benchmark and compare different algorithms on massive dense and sparse data matrices of size that spans from few hundreds of millions to billions. We demonstrate the scalability of our algorithm and compare it with baseline implementations, showing significant performance improvements. The code and the datasets used for conducting the experiments are available online.« less
Psychometric evaluation of the Spanish version of the MPI-SCI.
Soler, M D; Cruz-Almeida, Y; Saurí, J; Widerström-Noga, E G
2013-07-01
Postal surveys. To confirm the factor structure of the Spanish version of the MPI-SCI (MPI-SCI-S, Multidimensional Pain Inventory in the SCI population) and to test its internal consistency and construct validity in a Spanish population. Guttmann Institute, Barcelona, Spain. The MPI-SCI-S along with Spanish measures of pain intensity (Numerical Rating Scale), pain interference (Brief Pain Inventory), functional independence (Functional Independence Measure), depression (Beck Depression Inventory), locus of control (Multidimensional health Locus of Control), support (Functional Social Support Questionnaire (Duke-UNC)), psychological well-being (Psychological Global Well-Being Index) and demographic/injury characteristics were assessed in persons with spinal cord injury (SCI) and chronic pain (n=126). Confirmatory factor analysis suggested an adequate factor structure for the MPI-SCI-S. The internal consistency of the MPI-SCI-S subscales ranged from acceptable (r=0.66, Life Control) to excellent (r=0.94, Life Interference). All MPI-SCI-S subscales showed adequate construct validity, with the exception of the Negative and Solicitous Responses subscales. The Spanish version of the MPI-SCI is adequate for evaluating chronic pain impact following SCI in a Spanish-speaking population. Future studies should include additional measures of pain-related support in the Spanish-speaking SCI population.
High-Performance Data Analysis Tools for Sun-Earth Connection Missions
NASA Technical Reports Server (NTRS)
Messmer, Peter
2011-01-01
The data analysis tool of choice for many Sun-Earth Connection missions is the Interactive Data Language (IDL) by ITT VIS. The increasing amount of data produced by these missions and the increasing complexity of image processing algorithms requires access to higher computing power. Parallel computing is a cost-effective way to increase the speed of computation, but algorithms oftentimes have to be modified to take advantage of parallel systems. Enhancing IDL to work on clusters gives scientists access to increased performance in a familiar programming environment. The goal of this project was to enable IDL applications to benefit from both computing clusters as well as graphics processing units (GPUs) for accelerating data analysis tasks. The tool suite developed in this project enables scientists now to solve demanding data analysis problems in IDL that previously required specialized software, and it allows them to be solved orders of magnitude faster than on conventional PCs. The tool suite consists of three components: (1) TaskDL, a software tool that simplifies the creation and management of task farms, collections of tasks that can be processed independently and require only small amounts of data communication; (2) mpiDL, a tool that allows IDL developers to use the Message Passing Interface (MPI) inside IDL for problems that require large amounts of data to be exchanged among multiple processors; and (3) GPULib, a tool that simplifies the use of GPUs as mathematical coprocessors from within IDL. mpiDL is unique in its support for the full MPI standard and its support of a broad range of MPI implementations. GPULib is unique in enabling users to take advantage of an inexpensive piece of hardware, possibly already installed in their computer, and achieve orders of magnitude faster execution time for numerically complex algorithms. TaskDL enables the simple setup and management of task farms on compute clusters. The products developed in this project have the potential to interact, so one can build a cluster of PCs, each equipped with a GPU, and use mpiDL to communicate between the nodes and GPULib to accelerate the computations on each node.
Streaming data analytics via message passing with application to graph algorithms
Plimpton, Steven J.; Shead, Tim
2014-05-06
The need to process streaming data, which arrives continuously at high-volume in real-time, arises in a variety of contexts including data produced by experiments, collections of environmental or network sensors, and running simulations. Streaming data can also be formulated as queries or transactions which operate on a large dynamic data store, e.g. a distributed database. We describe a lightweight, portable framework named PHISH which enables a set of independent processes to compute on a stream of data in a distributed-memory parallel manner. Datums are routed between processes in patterns defined by the application. PHISH can run on top of eithermore » message-passing via MPI or sockets via ZMQ. The former means streaming computations can be run on any parallel machine which supports MPI; the latter allows them to run on a heterogeneous, geographically dispersed network of machines. We illustrate how PHISH can support streaming MapReduce operations, and describe streaming versions of three algorithms for large, sparse graph analytics: triangle enumeration, subgraph isomorphism matching, and connected component finding. Lastly, we also provide benchmark timings for MPI versus socket performance of several kernel operations useful in streaming algorithms.« less
Accelerating DNA analysis applications on GPU clusters
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tumeo, Antonino; Villa, Oreste
DNA analysis is an emerging application of high performance bioinformatic. Modern sequencing machinery are able to provide, in few hours, large input streams of data which needs to be matched against exponentially growing databases known fragments. The ability to recognize these patterns effectively and fastly may allow extending the scale and the reach of the investigations performed by biology scientists. Aho-Corasick is an exact, multiple pattern matching algorithm often at the base of this application. High performance systems are a promising platform to accelerate this algorithm, which is computationally intensive but also inherently parallel. Nowadays, high performance systems also includemore » heterogeneous processing elements, such as Graphic Processing Units (GPUs), to further accelerate parallel algorithms. Unfortunately, the Aho-Corasick algorithm exhibits large performance variabilities, depending on the size of the input streams, on the number of patterns to search and on the number of matches, and poses significant challenges on current high performance software and hardware implementations. An adequate mapping of the algorithm on the target architecture, coping with the limit of the underlining hardware, is required to reach the desired high throughputs. Load balancing also plays a crucial role when considering the limited bandwidth among the nodes of these systems. In this paper we present an efficient implementation of the Aho-Corasick algorithm for high performance clusters accelerated with GPUs. We discuss how we partitioned and adapted the algorithm to fit the Tesla C1060 GPU and then present a MPI based implementation for a heterogeneous high performance cluster. We compare this implementation to MPI and MPI with pthreads based implementations for a homogeneous cluster of x86 processors, discussing the stability vs. the performance and the scaling of the solutions, taking into consideration aspects such as the bandwidth among the different nodes.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gorentla Venkata, Manjunath; Shamis, Pavel; Graham, Richard L
2013-01-01
Many scientific simulations, using the Message Passing Interface (MPI) programming model, are sensitive to the performance and scalability of reduction collective operations such as MPI Allreduce and MPI Reduce. These operations are the most widely used abstractions to perform mathematical operations over all processes that are part of the simulation. In this work, we propose a hierarchical design to implement the reduction operations on multicore systems. This design aims to improve the efficiency of reductions by 1) tailoring the algorithms and customizing the implementations for various communication mechanisms in the system 2) providing the ability to configure the depth ofmore » hierarchy to match the system architecture, and 3) providing the ability to independently progress each of this hierarchy. Using this design, we implement MPI Allreduce and MPI Reduce operations (and its nonblocking variants MPI Iallreduce and MPI Ireduce) for all message sizes, and evaluate on multiple architectures including InfiniBand and Cray XT5. We leverage and enhance our existing infrastructure, Cheetah, which is a framework for implementing hierarchical collective operations to implement these reductions. The experimental results show that the Cheetah reduction operations outperform the production-grade MPI implementations such as Open MPI default, Cray MPI, and MVAPICH2, demonstrating its efficiency, flexibility and portability. On Infini- Band systems, with a microbenchmark, a 512-process Cheetah nonblocking Allreduce and Reduce achieves a speedup of 23x and 10x, respectively, compared to the default Open MPI reductions. The blocking variants of the reduction operations also show similar performance benefits. A 512-process nonblocking Cheetah Allreduce achieves a speedup of 3x, compared to the default MVAPICH2 Allreduce implementation. On a Cray XT5 system, a 6144-process Cheetah Allreduce outperforms the Cray MPI by 145%. The evaluation with an application kernel, Conjugate Gradient solver, shows that the Cheetah reductions speeds up total time to solution by 195%, demonstrating the potential benefits for scientific simulations.« less
Development of mpi_EPIC model for global agroecosystem modeling
Kang, Shujiang; Wang, Dali; Jeff A. Nichols; ...
2014-12-31
Models that address policy-maker concerns about multi-scale effects of food and bioenergy production systems are computationally demanding. We integrated the message passing interface algorithm into the process-based EPIC model to accelerate computation of ecosystem effects. Simulation performance was further enhanced by applying the Vampir framework. When this enhanced mpi_EPIC model was tested, total execution time for a global 30-year simulation of a switchgrass cropping system was shortened to less than 0.5 hours on a supercomputer. The results illustrate that mpi_EPIC using parallel design can balance simulation workloads and facilitate large-scale, high-resolution analysis of agricultural production systems, management alternatives and environmentalmore » effects.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tessier, Francois; Vishwanath, Venkatram
2017-11-28
Reading and writing data efficiently from different tiers of storage is necessary for most scientific simulations to achieve good performance at scale. Many software solutions have been developed to decrease the I/O bottleneck. One wellknown strategy, in the context of collective I/O operations, is the two-phase I/O scheme. This strategy consists of selecting a subset of processes to aggregate contiguous pieces of data before performing reads/writes. In our previous work, we implemented the two-phase I/O scheme with a MPI-based topology-aware algorithm. Our algorithm showed very good performance at scale compared to the standard I/O libraries such as POSIX I/O andmore » MPI I/O. However, the algorithm had several limitations hindering a satisfying reproducibility of our experiments. In this paper, we extend our work by 1) identifying the obstacles we face to reproduce our experiments and 2) discovering solutions that reduce the unpredictability of our results.« less
Eigensolver for a Sparse, Large Hermitian Matrix
NASA Technical Reports Server (NTRS)
Tisdale, E. Robert; Oyafuso, Fabiano; Klimeck, Gerhard; Brown, R. Chris
2003-01-01
A parallel-processing computer program finds a few eigenvalues in a sparse Hermitian matrix that contains as many as 100 million diagonal elements. This program finds the eigenvalues faster, using less memory, than do other, comparable eigensolver programs. This program implements a Lanczos algorithm in the American National Standards Institute/ International Organization for Standardization (ANSI/ISO) C computing language, using the Message Passing Interface (MPI) standard to complement an eigensolver in PARPACK. [PARPACK (Parallel Arnoldi Package) is an extension, to parallel-processing computer architectures, of ARPACK (Arnoldi Package), which is a collection of Fortran 77 subroutines that solve large-scale eigenvalue problems.] The eigensolver runs on Beowulf clusters of computers at the Jet Propulsion Laboratory (JPL).
Suplatov, Dmitry; Popova, Nina; Zhumatiy, Sergey; Voevodin, Vladimir; Švedas, Vytas
2016-04-01
Rapid expansion of online resources providing access to genomic, structural, and functional information associated with biological macromolecules opens an opportunity to gain a deeper understanding of the mechanisms of biological processes due to systematic analysis of large datasets. This, however, requires novel strategies to optimally utilize computer processing power. Some methods in bioinformatics and molecular modeling require extensive computational resources. Other algorithms have fast implementations which take at most several hours to analyze a common input on a modern desktop station, however, due to multiple invocations for a large number of subtasks the full task requires a significant computing power. Therefore, an efficient computational solution to large-scale biological problems requires both a wise parallel implementation of resource-hungry methods as well as a smart workflow to manage multiple invocations of relatively fast algorithms. In this work, a new computer software mpiWrapper has been developed to accommodate non-parallel implementations of scientific algorithms within the parallel supercomputing environment. The Message Passing Interface has been implemented to exchange information between nodes. Two specialized threads - one for task management and communication, and another for subtask execution - are invoked on each processing unit to avoid deadlock while using blocking calls to MPI. The mpiWrapper can be used to launch all conventional Linux applications without the need to modify their original source codes and supports resubmission of subtasks on node failure. We show that this approach can be used to process huge amounts of biological data efficiently by running non-parallel programs in parallel mode on a supercomputer. The C++ source code and documentation are available from http://biokinet.belozersky.msu.ru/mpiWrapper .
DOE Office of Scientific and Technical Information (OSTI.GOV)
Orton, Elizabeth J., E-mail: eorton@physics.carleton.ca; Kemp, Robert A. de; Glenn Wells, R.
2014-10-15
Purpose: Myocardial perfusion imaging (MPI) is used for diagnosis and prognosis of coronary artery disease. When MPI studies are performed with positron emission tomography (PET) and the radioactive tracer rubidium-82 chloride ({sup 82}Rb), a small but non-negligible fraction of studies (∼10%) suffer from extracardiac interference: high levels of tracer uptake in structures adjacent to the heart which mask the true cardiac tracer uptake. At present, there are no clinically available options for automated detection or correction of this problem. This work presents an algorithm that detects and classifies the severity of extracardiac interference in {sup 82}Rb PET MPI images andmore » reports the accuracy and failure rate of the method. Methods: A set of 200 {sup 82}Rb PET MPI images were reviewed by a trained nuclear cardiologist and interference severity reported on a four-class scale, from absent to severe. An automated algorithm was developed that compares uptake at the external border of the myocardium to three thresholds, separating the four interference severity classes. A minimum area of interference was required, and the search region was limited to that facing the stomach wall and spleen. Maximizing concordance (Cohen’s Kappa) and minimizing failure rate for the set of 200 clinician-read images were used to find the optimal population-based constants defining search limit and minimum area parameters and the thresholds for the algorithm. Tenfold stratified cross-validation was used to find optimal thresholds and report accuracy measures (sensitivity, specificity, and Kappa). Results: The algorithm was capable of detecting interference with a mean [95% confidence interval] sensitivity/specificity/Kappa of 0.97 [0.94, 1.00]/0.82 [0.66, 0.98]/0.79 [0.65, 0.92], and a failure rate of 1.0% ± 0.2%. The four-class overall Kappa was 0.72 [0.64, 0.81]. Separation of mild versus moderate-or-greater interference was performed with good accuracy (sensitivity/specificity/Kappa = 0.92 [0.86, 0.99]/0.86 [0.71, 1.00]/0.78 [0.64, 0.92]), while separation of moderate versus severe interference severity classes showed reduced sensitivity/Kappa but little change in specificity (sensitivity/specificity/Kappa = 0.83 [0.77, 0.88]/0.82 [0.77, 0.88]/0.65 [0.60, 0.70]). Specificity was greater than sensitivity for all interference classes. Algorithm execution time was <1 min. Conclusions: The algorithm produced here has a low failure rate and high accuracy for detection of extracardiac interference in {sup 82}Rb PET MPI scans. It provides a fast, reliable, automated method for assessing severity of extracardiac interference.« less
An MPI-based MoSST core dynamics model
NASA Astrophysics Data System (ADS)
Jiang, Weiyuan; Kuang, Weijia
2008-09-01
Distributed systems are among the main cost-effective and expandable platforms for high-end scientific computing. Therefore scalable numerical models are important for effective use of such systems. In this paper, we present an MPI-based numerical core dynamics model for simulation of geodynamo and planetary dynamos, and for simulation of core-mantle interactions. The model is developed based on MPI libraries. Two algorithms are used for node-node communication: a "master-slave" architecture and a "divide-and-conquer" architecture. The former is easy to implement but not scalable in communication. The latter is scalable in both computation and communication. The model scalability is tested on Linux PC clusters with up to 128 nodes. This model is also benchmarked with a published numerical dynamo model solution.
Zhan, X.
2005-01-01
A parallel Fortran-MPI (Message Passing Interface) software for numerical inversion of the Laplace transform based on a Fourier series method is developed to meet the need of solving intensive computational problems involving oscillatory water level's response to hydraulic tests in a groundwater environment. The software is a parallel version of ACM (The Association for Computing Machinery) Transactions on Mathematical Software (TOMS) Algorithm 796. Running 38 test examples indicated that implementation of MPI techniques with distributed memory architecture speedups the processing and improves the efficiency. Applications to oscillatory water levels in a well during aquifer tests are presented to illustrate how this package can be applied to solve complicated environmental problems involved in differential and integral equations. The package is free and is easy to use for people with little or no previous experience in using MPI but who wish to get off to a quick start in parallel computing. ?? 2004 Elsevier Ltd. All rights reserved.
Woo, Minjeong; Wood, Connor; Kwon, Doyoon; Park, Kyu-Ho Paul; Fejer, György; Delorme, Vincent
2018-01-01
Lung alveolar macrophages (AMs) are in the first line of immune defense against respiratory pathogens and play key roles in the pathogenesis of Mycobacterium tuberculosis ( Mtb ) in humans. Nevertheless, AMs are available only in limited amounts for in vitro studies, which hamper the detailed molecular understanding of host- Mtb interactions in these macrophages. The recent establishment of the self-renewing and primary Max Planck Institute (MPI) cells, functionally very close to lung AMs, opens unique opportunities for in vitro studies of host-pathogen interactions in respiratory diseases. Here, we investigated the suitability of MPI cells as a host cell system for Mtb infection. Bacterial, cellular, and innate immune features of MPI cells infected with Mtb were characterized. Live bacteria were readily internalized and efficiently replicated in MPI cells, similarly to primary murine macrophages and other cell lines. MPI cells were also suitable for the determination of anti-tuberculosis (TB) drug activity. The primary innate immune response of MPI cells to live Mtb showed significantly higher and earlier induction of the pro-inflammatory cytokines TNFα, interleukin 6 (IL-6), IL-1α, and IL-1β, as compared to stimulation with heat-killed (HK) bacteria. MPI cells previously showed a lack of induction of the anti-inflammatory cytokine IL-10 to a wide range of stimuli, including HK Mtb . By contrast, we show here that live Mtb is able to induce significant amounts of IL-10 in MPI cells. Autophagy experiments using light chain 3B immunostaining, as well as LysoTracker labeling of acidic vacuoles, demonstrated that MPI cells efficiently control killed Mtb by elimination through phagolysosomes. MPI cells were also able to accumulate lipid droplets in their cytoplasm following exposure to lipoproteins. Collectively, this study establishes the MPI cells as a relevant, versatile host cell model for TB research, allowing a deeper understanding of AMs functions in this pathology.
Practical Formal Verification of MPI and Thread Programs
NASA Astrophysics Data System (ADS)
Gopalakrishnan, Ganesh; Kirby, Robert M.
Large-scale simulation codes in science and engineering are written using the Message Passing Interface (MPI). Shared memory threads are widely used directly, or to implement higher level programming abstractions. Traditional debugging methods for MPI or thread programs are incapable of providing useful formal guarantees about coverage. They get bogged down in the sheer number of interleavings (schedules), often missing shallow bugs. In this tutorial we will introduce two practical formal verification tools: ISP (for MPI C programs) and Inspect (for Pthread C programs). Unlike other formal verification tools, ISP and Inspect run directly on user source codes (much like a debugger). They pursue only the relevant set of process interleavings, using our own customized Dynamic Partial Order Reduction algorithms. For a given test harness, DPOR allows these tools to guarantee the absence of deadlocks, instrumented MPI object leaks and communication races (using ISP), and shared memory races (using Inspect). ISP and Inspect have been used to verify large pieces of code: in excess of 10,000 lines of MPI/C for ISP in under 5 seconds, and about 5,000 lines of Pthread/C code in a few hours (and much faster with the use of a cluster or by exploiting special cases such as symmetry) for Inspect. We will also demonstrate the Microsoft Visual Studio and Eclipse Parallel Tools Platform integrations of ISP (these will be available on the LiveCD).
Military Interoperable Digital Hospital Testbed
2012-07-01
Subject outcome measures include blood pressure, waist circumference, weight, body mass index (BMI), body fat, HDL cholesterol, triglycerides , glucose ...master patient index (MPI), 625 duplicate chest x-rays and CT scans of the head between sending and receiving institution (taken within 0-7 days) were... Index (MPI) software. The report included chest x-rays (CPT 71010 and 71020) and CT scans of the head (CPT 70450) for the stated time periods. The
Distributed memory parallel Markov random fields using graph partitioning
DOE Office of Scientific and Technical Information (OSTI.GOV)
Heinemann, C.; Perciano, T.; Ushizima, D.
Markov random fields (MRF) based algorithms have attracted a large amount of interest in image analysis due to their ability to exploit contextual information about data. Image data generated by experimental facilities, though, continues to grow larger and more complex, making it more difficult to analyze in a reasonable amount of time. Applying image processing algorithms to large datasets requires alternative approaches to circumvent performance problems. Aiming to provide scientists with a new tool to recover valuable information from such datasets, we developed a general purpose distributed memory parallel MRF-based image analysis framework (MPI-PMRF). MPI-PMRF overcomes performance and memory limitationsmore » by distributing data and computations across processors. The proposed approach was successfully tested with synthetic and experimental datasets. Additionally, the performance of the MPI-PMRF framework is analyzed through a detailed scalability study. We show that a performance increase is obtained while maintaining an accuracy of the segmentation results higher than 98%. The contributions of this paper are: (a) development of a distributed memory MRF framework; (b) measurement of the performance increase of the proposed approach; (c) verification of segmentation accuracy in both synthetic and experimental, real-world datasets« less
Large Scale Frequent Pattern Mining using MPI One-Sided Model
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vishnu, Abhinav; Agarwal, Khushbu
In this paper, we propose a work-stealing runtime --- Library for Work Stealing LibWS --- using MPI one-sided model for designing scalable FP-Growth --- {\\em de facto} frequent pattern mining algorithm --- on large scale systems. LibWS provides locality efficient and highly scalable work-stealing techniques for load balancing on a variety of data distributions. We also propose a novel communication algorithm for FP-growth data exchange phase, which reduces the communication complexity from state-of-the-art O(p) to O(f + p/f) for p processes and f frequent attributed-ids. FP-Growth is implemented using LibWS and evaluated on several work distributions and support counts. Anmore » experimental evaluation of the FP-Growth on LibWS using 4096 processes on an InfiniBand Cluster demonstrates excellent efficiency for several work distributions (87\\% efficiency for Power-law and 91% for Poisson). The proposed distributed FP-Tree merging algorithm provides 38x communication speedup on 4096 cores.« less
TADtool: visual parameter identification for TAD-calling algorithms.
Kruse, Kai; Hug, Clemens B; Hernández-Rodríguez, Benjamín; Vaquerizas, Juan M
2016-10-15
Eukaryotic genomes are hierarchically organized into topologically associating domains (TADs). The computational identification of these domains and their associated properties critically depends on the choice of suitable parameters of TAD-calling algorithms. To reduce the element of trial-and-error in parameter selection, we have developed TADtool: an interactive plot to find robust TAD-calling parameters with immediate visual feedback. TADtool allows the direct export of TADs called with a chosen set of parameters for two of the most common TAD calling algorithms: directionality and insulation index. It can be used as an intuitive, standalone application or as a Python package for maximum flexibility. TADtool is available as a Python package from GitHub (https://github.com/vaquerizaslab/tadtool) or can be installed directly via PyPI, the Python package index (tadtool). kai.kruse@mpi-muenster.mpg.de, jmv@mpi-muenster.mpg.deSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.
Normal Databases for the Relative Quantification of Myocardial Perfusion
Rubeaux, Mathieu; Xu, Yuan; Germano, Guido; Berman, Daniel S.; Slomka, Piotr J.
2016-01-01
Purpose of review Myocardial perfusion imaging (MPI) with SPECT is performed clinically worldwide to detect and monitor coronary artery disease (CAD). MPI allows an objective quantification of myocardial perfusion at stress and rest. This established technique relies on normal databases to compare patient scans against reference normal limits. In this review, we aim to introduce the process of MPI quantification with normal databases and describe the associated perfusion quantitative measures that are used. Recent findings New equipment and new software reconstruction algorithms have been introduced which require the development of new normal limits. The appearance and regional count variations of normal MPI scan may differ between these new scanners and standard Anger cameras. Therefore, these new systems may require the determination of new normal limits to achieve optimal accuracy in relative myocardial perfusion quantification. Accurate diagnostic and prognostic results rivaling those obtained by expert readers can be obtained by this widely used technique. Summary Throughout this review, we emphasize the importance of the different normal databases and the need for specific databases relative to distinct imaging procedures. use of appropriate normal limits allows optimal quantification of MPI by taking into account subtle image differences due to the hardware and software used, and the population studied. PMID:28138354
Magnetic particle imaging: from proof of principle to preclinical applications
NASA Astrophysics Data System (ADS)
Knopp, T.; Gdaniec, N.; Möddel, M.
2017-07-01
Tomographic imaging has become a mandatory tool for the diagnosis of a majority of diseases in clinical routine. Since each method has its pros and cons, a variety of them is regularly used in clinics to satisfy all application needs. Magnetic particle imaging (MPI) is a relatively new tomographic imaging technique that images magnetic nanoparticles with a high spatiotemporal resolution in a quantitative way, and in turn is highly suited for vascular and targeted imaging. MPI was introduced in 2005 and now enters the preclinical research phase, where medical researchers get access to this new technology and exploit its potential under physiological conditions. Within this paper, we review the development of MPI since its introduction in 2005. Besides an in-depth description of the basic principles, we provide detailed discussions on imaging sequences, reconstruction algorithms, scanner instrumentation and potential medical applications.
Parallel processing optimization strategy based on MapReduce model in cloud storage environment
NASA Astrophysics Data System (ADS)
Cui, Jianming; Liu, Jiayi; Li, Qiuyan
2017-05-01
Currently, a large number of documents in the cloud storage process employed the way of packaging after receiving all the packets. From the local transmitter this stored procedure to the server, packing and unpacking will consume a lot of time, and the transmission efficiency is low as well. A new parallel processing algorithm is proposed to optimize the transmission mode. According to the operation machine graphs model work, using MPI technology parallel execution Mapper and Reducer mechanism. It is good to use MPI technology to implement Mapper and Reducer parallel mechanism. After the simulation experiment of Hadoop cloud computing platform, this algorithm can not only accelerate the file transfer rate, but also shorten the waiting time of the Reducer mechanism. It will break through traditional sequential transmission constraints and reduce the storage coupling to improve the transmission efficiency.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gorentla Venkata, Manjunath; Graham, Richard L; Ladd, Joshua S
This paper describes the design and implementation of InfiniBand (IB) CORE-Direct based blocking and nonblocking broadcast operations within the Cheetah collective operation framework. It describes a novel approach that fully ofFLoads collective operations and employs only user-supplied buffers. For a 64 rank communicator, the latency of CORE-Direct based hierarchical algorithm is better than production-grade Message Passing Interface (MPI) implementations, 150% better than the default Open MPI algorithm and 115% better than the shared memory optimized MVAPICH implementation for a one kilobyte (KB) message, and for eight mega-bytes (MB) it is 48% and 64% better, respectively. Flat-topology broadcast achieves 99.9% overlapmore » in a polling based communication-computation test, and 95.1% overlap for a wait based test, compared with 92.4% and 17.0%, respectively, for a similar Central Processing Unit (CPU) based implementation.« less
Collignon, Barbara; Schulz, Roland; Smith, Jeremy C; Baudry, Jerome
2011-04-30
A message passing interface (MPI)-based implementation (Autodock4.lga.MPI) of the grid-based docking program Autodock4 has been developed to allow simultaneous and independent docking of multiple compounds on up to thousands of central processing units (CPUs) using the Lamarkian genetic algorithm. The MPI version reads a single binary file containing precalculated grids that represent the protein-ligand interactions, i.e., van der Waals, electrostatic, and desolvation potentials, and needs only two input parameter files for the entire docking run. In comparison, the serial version of Autodock4 reads ASCII grid files and requires one parameter file per compound. The modifications performed result in significantly reduced input/output activity compared with the serial version. Autodock4.lga.MPI scales up to 8192 CPUs with a maximal overhead of 16.3%, of which two thirds is due to input/output operations and one third originates from MPI operations. The optimal docking strategy, which minimizes docking CPU time without lowering the quality of the database enrichments, comprises the docking of ligands preordered from the most to the least flexible and the assignment of the number of energy evaluations as a function of the number of rotatable bounds. In 24 h, on 8192 high-performance computing CPUs, the present MPI version would allow docking to a rigid protein of about 300K small flexible compounds or 11 million rigid compounds.
Cheetah: A Framework for Scalable Hierarchical Collective Operations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Graham, Richard L; Gorentla Venkata, Manjunath; Ladd, Joshua S
2011-01-01
Collective communication operations, used by many scientific applications, tend to limit overall parallel application performance and scalability. Computer systems are becoming more heterogeneous with increasing node and core-per-node counts. Also, a growing number of data-access mechanisms, of varying characteristics, are supported within a single computer system. We describe a new hierarchical collective communication framework that takes advantage of hardware-specific data-access mechanisms. It is flexible, with run-time hierarchy specification, and sharing of collective communication primitives between collective algorithms. Data buffers are shared between levels in the hierarchy reducing collective communication management overhead. We have implemented several versions of the Message Passingmore » Interface (MPI) collective operations, MPI Barrier() and MPI Bcast(), and run experiments using up to 49, 152 processes on a Cray XT5, and a small InfiniBand based cluster. At 49, 152 processes our barrier implementation outperforms the optimized native implementation by 75%. 32 Byte and one Mega-Byte broadcasts outperform it by 62% and 11%, respectively, with better scalability characteristics. Improvements relative to the default Open MPI implementation are much larger.« less
Shrimankar, D D; Sathe, S R
2016-01-01
Sequence alignment is an important tool for describing the relationships between DNA sequences. Many sequence alignment algorithms exist, differing in efficiency, in their models of the sequences, and in the relationship between sequences. The focus of this study is to obtain an optimal alignment between two sequences of biological data, particularly DNA sequences. The algorithm is discussed with particular emphasis on time, speedup, and efficiency optimizations. Parallel programming presents a number of critical challenges to application developers. Today's supercomputer often consists of clusters of SMP nodes. Programming paradigms such as OpenMP and MPI are used to write parallel codes for such architectures. However, the OpenMP programs cannot be scaled for more than a single SMP node. However, programs written in MPI can have more than single SMP nodes. But such a programming paradigm has an overhead of internode communication. In this work, we explore the tradeoffs between using OpenMP and MPI. We demonstrate that the communication overhead incurs significantly even in OpenMP loop execution and increases with the number of cores participating. We also demonstrate a communication model to approximate the overhead from communication in OpenMP loops. Our results are astonishing and interesting to a large variety of input data files. We have developed our own load balancing and cache optimization technique for message passing model. Our experimental results show that our own developed techniques give optimum performance of our parallel algorithm for various sizes of input parameter, such as sequence size and tile size, on a wide variety of multicore architectures.
Shrimankar, D. D.; Sathe, S. R.
2016-01-01
Sequence alignment is an important tool for describing the relationships between DNA sequences. Many sequence alignment algorithms exist, differing in efficiency, in their models of the sequences, and in the relationship between sequences. The focus of this study is to obtain an optimal alignment between two sequences of biological data, particularly DNA sequences. The algorithm is discussed with particular emphasis on time, speedup, and efficiency optimizations. Parallel programming presents a number of critical challenges to application developers. Today’s supercomputer often consists of clusters of SMP nodes. Programming paradigms such as OpenMP and MPI are used to write parallel codes for such architectures. However, the OpenMP programs cannot be scaled for more than a single SMP node. However, programs written in MPI can have more than single SMP nodes. But such a programming paradigm has an overhead of internode communication. In this work, we explore the tradeoffs between using OpenMP and MPI. We demonstrate that the communication overhead incurs significantly even in OpenMP loop execution and increases with the number of cores participating. We also demonstrate a communication model to approximate the overhead from communication in OpenMP loops. Our results are astonishing and interesting to a large variety of input data files. We have developed our own load balancing and cache optimization technique for message passing model. Our experimental results show that our own developed techniques give optimum performance of our parallel algorithm for various sizes of input parameter, such as sequence size and tile size, on a wide variety of multicore architectures. PMID:27932868
Ishihara, Masaru; Onoguchi, Masahisa; Taniguchi, Yasuyo; Shibutani, Takayuki
2017-12-01
The aim of this study was to clarify the differences in thallium-201-chloride (thallium-201) myocardial perfusion imaging (MPI) scans evaluated by conventional anger-type single-photon emission computed tomography (conventional SPECT) versus cadmium-zinc-telluride SPECT (CZT SPECT) imaging in normal databases for different ethnic groups. MPI scans from 81 consecutive Japanese patients were examined using conventional SPECT and CZT SPECT and analyzed with the pre-installed quantitative perfusion SPECT (QPS) software. We compared the summed stress score (SSS), summed rest score (SRS), and summed difference score (SDS) for the two SPECT devices. For a normal MPI reference, we usually use Japanese databases for MPI created by the Japanese Society of Nuclear Medicine, which can be used with conventional SPECT but not with CZT SPECT. In this study, we used new Japanese normal databases constructed in our institution to compare conventional and CZT SPECT. Compared with conventional SPECT, CZT SPECT showed lower SSS (p < 0.001), SRS (p = 0.001), and SDS (p = 0.189) using the pre-installed SPECT database. In contrast, CZT SPECT showed no significant difference from conventional SPECT in QPS analysis using the normal databases from our institution. Myocardial perfusion analyses by CZT SPECT should be evaluated using normal databases based on the ethnic group being evaluated.
[A utopian episode - Carl Friedrich von Weizsäcker in the networks of the Max-Planck Society].
Kant, Horst; Renn, Jürgen
2014-01-01
Carl Friedrich von Weizsäcker was a key figure in the history of the Max Planck Society (MPS). This essay contextualises his work with the development of the MPS, highlighting the institutional and personal networks upon which it was based. Some of the stations addressed in the following are his role in the German Uranium Project, in preparing the Mainau Declaration, the Göttingen Manifesto, and the Memorandum of Tübingen as well as his involvement in the foundation of the Max Planck Institute (MPI) for Human Development and his own MPI for the Research of Living Conditions in the Modern World located in Starnberg. The relationship between Weizsäcker and Hellmut Becker, long-time friend and founding director of the MPI for Human Development, will be of particular interest. Another issue broached here is the connection between natural science and the humanities in Weizsäcker's work, and subsequently the relation between these two science cultures in the MPS. Finally, we look at the challenges Weizsäcker's work could present to the MPS today.
A new deadlock resolution protocol and message matching algorithm for the extreme-scale simulator
Engelmann, Christian; Naughton, III, Thomas J.
2016-03-22
Investigating the performance of parallel applications at scale on future high-performance computing (HPC) architectures and the performance impact of different HPC architecture choices is an important component of HPC hardware/software co-design. The Extreme-scale Simulator (xSim) is a simulation toolkit for investigating the performance of parallel applications at scale. xSim scales to millions of simulated Message Passing Interface (MPI) processes. The overhead introduced by a simulation tool is an important performance and productivity aspect. This paper documents two improvements to xSim: (1)~a new deadlock resolution protocol to reduce the parallel discrete event simulation overhead and (2)~a new simulated MPI message matchingmore » algorithm to reduce the oversubscription management overhead. The results clearly show a significant performance improvement. The simulation overhead for running the NAS Parallel Benchmark suite was reduced from 102% to 0% for the embarrassingly parallel (EP) benchmark and from 1,020% to 238% for the conjugate gradient (CG) benchmark. xSim offers a highly accurate simulation mode for better tracking of injected MPI process failures. Furthermore, with highly accurate simulation, the overhead was reduced from 3,332% to 204% for EP and from 37,511% to 13,808% for CG.« less
A Locality-Based Threading Algorithm for the Configuration-Interaction Method
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shan, Hongzhang; Williams, Samuel; Johnson, Calvin
The Configuration Interaction (CI) method has been widely used to solve the non-relativistic many-body Schrodinger equation. One great challenge to implementing it efficiently on manycore architectures is its immense memory and data movement requirements. To address this issue, within each node, we exploit a hybrid MPI+OpenMP programming model in lieu of the traditional flat MPI programming model. Here in this paper, we develop optimizations that partition the workloads among OpenMP threads based on data locality,-which is essential in ensuring applications with complex data access patterns scale well on manycore architectures. The new algorithm scales to 256 threadson the 64-core Intelmore » Knights Landing (KNL) manycore processor and 24 threads on dual-socket Ivy Bridge (Xeon) nodes. Compared with the original implementation, the performance has been improved by up to 7× on theKnights Landing processor and 3× on the dual-socket Ivy Bridge node.« less
A Locality-Based Threading Algorithm for the Configuration-Interaction Method
Shan, Hongzhang; Williams, Samuel; Johnson, Calvin; ...
2017-07-03
The Configuration Interaction (CI) method has been widely used to solve the non-relativistic many-body Schrodinger equation. One great challenge to implementing it efficiently on manycore architectures is its immense memory and data movement requirements. To address this issue, within each node, we exploit a hybrid MPI+OpenMP programming model in lieu of the traditional flat MPI programming model. Here in this paper, we develop optimizations that partition the workloads among OpenMP threads based on data locality,-which is essential in ensuring applications with complex data access patterns scale well on manycore architectures. The new algorithm scales to 256 threadson the 64-core Intelmore » Knights Landing (KNL) manycore processor and 24 threads on dual-socket Ivy Bridge (Xeon) nodes. Compared with the original implementation, the performance has been improved by up to 7× on theKnights Landing processor and 3× on the dual-socket Ivy Bridge node.« less
Multilevel Parallelization of AutoDock 4.2.
Norgan, Andrew P; Coffman, Paul K; Kocher, Jean-Pierre A; Katzmann, David J; Sosa, Carlos P
2011-04-28
Virtual (computational) screening is an increasingly important tool for drug discovery. AutoDock is a popular open-source application for performing molecular docking, the prediction of ligand-receptor interactions. AutoDock is a serial application, though several previous efforts have parallelized various aspects of the program. In this paper, we report on a multi-level parallelization of AutoDock 4.2 (mpAD4). Using MPI and OpenMP, AutoDock 4.2 was parallelized for use on MPI-enabled systems and to multithread the execution of individual docking jobs. In addition, code was implemented to reduce input/output (I/O) traffic by reusing grid maps at each node from docking to docking. Performance of mpAD4 was examined on two multiprocessor computers. Using MPI with OpenMP multithreading, mpAD4 scales with near linearity on the multiprocessor systems tested. In situations where I/O is limiting, reuse of grid maps reduces both system I/O and overall screening time. Multithreading of AutoDock's Lamarkian Genetic Algorithm with OpenMP increases the speed of execution of individual docking jobs, and when combined with MPI parallelization can significantly reduce the execution time of virtual screens. This work is significant in that mpAD4 speeds the execution of certain molecular docking workloads and allows the user to optimize the degree of system-level (MPI) and node-level (OpenMP) parallelization to best fit both workloads and computational resources.
Message Passing vs. Shared Address Space on a Cluster of SMPs
NASA Technical Reports Server (NTRS)
Shan, Hongzhang; Singh, Jaswinder Pal; Oliker, Leonid; Biswas, Rupak
2000-01-01
The convergence of scalable computer architectures using clusters of PCs (or PC-SMPs) with commodity networking has become an attractive platform for high end scientific computing. Currently, message-passing and shared address space (SAS) are the two leading programming paradigms for these systems. Message-passing has been standardized with MPI, and is the most common and mature programming approach. However message-passing code development can be extremely difficult, especially for irregular structured computations. SAS offers substantial ease of programming, but may suffer from performance limitations due to poor spatial locality, and high protocol overhead. In this paper, we compare the performance of and programming effort, required for six applications under both programming models on a 32 CPU PC-SMP cluster. Our application suite consists of codes that typically do not exhibit high efficiency under shared memory programming. due to their high communication to computation ratios and complex communication patterns. Results indicate that SAS can achieve about half the parallel efficiency of MPI for most of our applications: however, on certain classes of problems SAS performance is competitive with MPI. We also present new algorithms for improving the PC cluster performance of MPI collective operations.
Magnetic particle imaging for in vivo blood flow velocity measurements in mice
NASA Astrophysics Data System (ADS)
Kaul, Michael G.; Salamon, Johannes; Knopp, Tobias; Ittrich, Harald; Adam, Gerhard; Weller, Horst; Jung, Caroline
2018-03-01
Magnetic particle imaging (MPI) is a new imaging technology. It is a potential candidate to be used for angiographic purposes, to study perfusion and cell migration. The aim of this work was to measure velocities of the flowing blood in the inferior vena cava of mice, using MPI, and to evaluate it in comparison with magnetic resonance imaging (MRI). A phantom mimicking the flow within the inferior vena cava with velocities of up to 21 cm s‑1 was used for the evaluation of the applied analysis techniques. Time–density and distance–density analyses for bolus tracking were performed to calculate flow velocities. These findings were compared with the calibrated velocities set by a flow pump, and it can be concluded that velocities of up to 21 cm s‑1 can be measured by MPI. A time–density analysis using an arrival time estimation algorithm showed the best agreement with the preset velocities. In vivo measurements were performed in healthy FVB mice (n = 10). MRI experiments were performed using phase contrast (PC) for velocity mapping. For MPI measurements, a standardized injection of a superparamagnetic iron oxide tracer was applied. In vivo MPI data were evaluated by a time–density analysis and compared to PC MRI. A Bland–Altman analysis revealed good agreement between the in vivo velocities acquired by MRI of 4.0 ± 1.5 cm s‑1 and those measured by MPI of 4.8 ± 1.1 cm s‑1. Magnetic particle imaging is a new tool with which to measure and quantify flow velocities. It is fast, radiation-free, and produces 3D images. It therefore offers the potential for vascular imaging.
Einstein, Andrew J.; Weiner, Shepard D.; Bernheim, Adam; Kulon, Michal; Bokhari, Sabahat; Johnson, Lynne L.; Moses, Jeffrey W.; Balter, Stephen
2013-01-01
Context Myocardial perfusion imaging (MPI) is the single medical test with the highest radiation burden to the US population. While many patients undergoing MPI receive repeat MPI testing, or additional procedures involving ionizing radiation, no data are available characterizing their total longitudinal radiation burden and relating radiation burden with reasons for testing. Objective To characterize procedure counts, cumulative estimated effective doses of radiation, and clinical indications, for patients undergoing MPI. Design, Setting, Patients Retrospective cohort study evaluating, for 1097 consecutive patients undergoing index MPI during the first 100 days of 2006 at Columbia University Medical Center, all preceding medical imaging procedures involving ionizing radiation undergone beginning October 1988, and all subsequent procedures through June 2008, at that center. Main Outcome Measures Cumulative estimated effective dose of radiation, number of procedures involving radiation, and indications for testing. Results Patients underwent a median (interquartile range, mean) of 15 (6–32, 23.9) procedures involving radiation exposure; 4 (2–8, 6.5) were high-dose (≥3 mSv, i.e. one year's background radiation), including 1 (1–2, 1.8) MPI studies per patient. 31% of patients received cumulative estimated effective dose from all medical sources >100mSv. Multiple MPIs were performed in 39% of patients, for whom cumulative estimated effective dose was 121 (81–189, 149) mSv. Men and whites had higher cumulative estimated effective doses, and there was a trend towards men being more likely to undergo multiple MPIs than women (40.8% vs. 36.6%, Odds ratio 1.29, 95% confidence interval 0.98–1.69). Over 80% of initial and 90% of repeat MPI exams were performed in patients with known cardiac disease or symptoms consistent with it. Conclusion In this institution, multiple testing with MPI was very common, and in many patients associated with very high cumulative estimated doses of radiation. PMID:21078807
Automatic Thread-Level Parallelization in the Chombo AMR Library
DOE Office of Scientific and Technical Information (OSTI.GOV)
Christen, Matthias; Keen, Noel; Ligocki, Terry
2011-05-26
The increasing on-chip parallelism has some substantial implications for HPC applications. Currently, hybrid programming models (typically MPI+OpenMP) are employed for mapping software to the hardware in order to leverage the hardware?s architectural features. In this paper, we present an approach that automatically introduces thread level parallelism into Chombo, a parallel adaptive mesh refinement framework for finite difference type PDE solvers. In Chombo, core algorithms are specified in the ChomboFortran, a macro language extension to F77 that is part of the Chombo framework. This domain-specific language forms an already used target language for an automatic migration of the large number ofmore » existing algorithms into a hybrid MPI+OpenMP implementation. It also provides access to the auto-tuning methodology that enables tuning certain aspects of an algorithm to hardware characteristics. Performance measurements are presented for a few of the most relevant kernels with respect to a specific application benchmark using this technique as well as benchmark results for the entire application. The kernel benchmarks show that, using auto-tuning, up to a factor of 11 in performance was gained with 4 threads with respect to the serial reference implementation.« less
Assimilating soil moisture into an Earth System Model
NASA Astrophysics Data System (ADS)
Stacke, Tobias; Hagemann, Stefan
2017-04-01
Several modelling studies reported potential impacts of soil moisture anomalies on regional climate. In particular for short prediction periods, perturbations of the soil moisture state may result in significant alteration of surface temperature in the following season. However, it is not clear yet whether or not soil moisture anomalies affect climate also on larger temporal and spatial scales. In an earlier study, we showed that soil moisture anomalies can persist for several seasons in the deeper soil layers of a land surface model. Additionally, those anomalies can influence root zone moisture, in particular during explicitly dry or wet periods. Thus, one prerequisite for predictability, namely the existence of long term memory, is evident for simulated soil moisture and might be exploited to improve climate predictions. The second prerequisite is the sensitivity of the climate system to soil moisture. In order to investigate this sensitivity for decadal simulations, we implemented a soil moisture assimilation scheme into the Max-Planck Institute for Meteorology's Earth System Model (MPI-ESM). The assimilation scheme is based on a simple nudging algorithm and updates the surface soil moisture state once per day. In our experiments, the MPI-ESM is used which includes model components for the interactive simulation of atmosphere, land and ocean. Artificial assimilation data is created from a control simulation to nudge the MPI-ESM towards predominantly dry and wet states. First analyses are focused on the impact of the assimilation on land surface variables and reveal distinct differences in the long-term mean values between wet and dry state simulations. Precipitation, evapotranspiration and runoff are larger in the wet state compared to the dry state, resulting in an increased moisture transport from the land to atmosphere and ocean. Consequently, surface temperatures are lower in the wet state simulations by more than one Kelvin. In terms of spatial pattern, the largest differences between both simulations are seen for continental areas, while regions with a maritime climate are least sensitive to soil moisture assimilation.
NASA Astrophysics Data System (ADS)
Hatzaki, M.; Flocas, H. A.; Kouroutzoglou, J.; Keay, K.; Simmonds, I.; Giannakopoulos, C. A.; Brikolas, V.
2011-12-01
A number of studies suggest that cyclone activity over both hemispheres has changed over the second half of the 20th century. The assessment of the future changes of the cyclonic activity as imposed by global warming conditions is very important since these cyclones can be associated with extreme precipitation conditions, severe storms and floods. This is more important for the Mediterranean that has been found to be more vulnerable to climate change. The main objective of the current study is to better understand and assess future changes in the main characteristics of Mediterranean cyclones, including temporal and spatial variations of frequency of cyclonic tracks, and dynamic and kinematic parameters, such as intensity, size, propagation velocity, as well as trend analysis. For this purpose, the MPI-HH regional coupled climate model of the Max Planck Institute for Meteorology is employed consisting of the REgional atmosphere MOdel (REMO), the Max-Planck-Institute for Meteorology ocean model (MPI-OM) and the Hydrological Discharge Model (HD Model). A 25 km resolution domain is established on a rotated latitude-longitude coordinate system, while the physical parameterizations are taken from the global climate model ECHAM-4. These model data became available through the EU Project CIRCE which aims to perform, for the first time, climate change projections with a realistic representation of the Mediterranean Sea. The model results for the present climate are evaluated against ERA-40 Reanalysis (available through ECMWF), for the period 1962-2001. The identification and tracking of cyclones is performed with the aid of the Melbourne University algorithm (MS algorithm), according to the Lagrangian perspective. MS algorithm characterizes a cyclone only if a vorticity maximum could be connected with a local pressure minimum. According to the results, a decrease of the storm number and a tendency towards deeper cyclones is expected in the future, in general agreement with the results of previous studies. However, new findings reveal with respect to the dynamic/kinematic characteristics of the cyclonic tracks. The model experiments verify that considerable inter-monthly variations of track density occur in the Mediterranean region. The study of the kinematic and dynamic parameters of the cyclonic tracks according to their origin domain show that the vast majority originate within the examined area itself. ACKNOWLEDGMENTS: M. Hatzaki would like to thank the Greek State Scholarships Foundation for financial support through the program of postdoctoral research. The support of EU-FP6 project CIRCE Integrated Project-Climate Change and Impact Research: the Mediterranean Environment (http://www.circeproject.eu) for climate model data provision is also greatly acknowledged.
Addison, Daniel; Singh, Vinita; Okyere-Asante, K; Okafor, Henry
2014-01-01
Patients presenting with chest pain and evidence of functional ischemia by myocardial perfusion imaging (MPI), but lacking commensurate angiographic disease pose a diagnostic and therapeutic dilemma. They are often dismissed as having 'false-positive MPI'. Moreover, a majority of the available long-term outcome data for it has been derived from homogenous female populations. In this study, we sought to evaluate the long-term outcomes of this presentation in a multiethnic male-predominant cohort. We retrospectively identified 47 patients who presented to our institution between 2002 and 2005 with chest pain and evidence of ischemia on MPI, but with no significant angiographic disease on subsequent cardiac catheterization (cases). The occurrence of adverse cardiovascular outcomes (chest pain, congestive heart failure, acute myocardial infarction and stroke) post-index coronary angiogram was tracked. Similar data was collected for 37 patients who also presented with chest pain, but normal MPI over the same period (controls). Overall average follow-up was over 22 months. Fifty-three percent (26/47) of the cases had one or more of the adverse outcomes as compared with 22% (8/37) of controls (P < 0.01). Of these, 13 (50.0%) and 3 (37.5%) were males, respectively. Ischemia on MPI is predictive of long-term adverse cardiovascular outcomes despite normal ('false-negative') coronary angiography. This appears to be gender-neutral.
DeCicco, Anthony E; Sokil, Alexis B; Marhefka, Gregary D; Reist, Kirk; Hansen, Christopher L
2015-04-01
Obesity is not only associated with an increased risk of coronary artery disease, but also decreases the accuracy of many diagnostic modalities pertinent to this disease. Advances in myocardial perfusion imaging (MPI) have mitigated somewhat the effects of obesity, although the feasibility of MPI in the super-obese (defined as a BMI > 50) is currently untested. We undertook this study to assess the practicality of MPI in the super-obese using a multi-headed solid-state gamma camera with attenuation correction. We retrospectively identified consecutive super-obese patients referred for MPI at our institution. The images were interpreted by 3 blinded, experienced readers and graded for quality and diagnosis, and subjectively evaluated the contribution of attenuation correction. Clinical follow-up was obtained from review of medical records. 72 consecutive super-obese patients were included. Their BMI ranged from 50 to 67 (55.7 ± 5.1). Stress image quality was considered good or excellent in 45 (63%), satisfactory in 24 (33%), poor in 3 (4%), and uninterpretable in 0 patients. Rest images were considered good or excellent in 34 (49%), satisfactory in 23 (33%), poor in 13 (19%), and uninterpretable in 0 patients. Attenuation correction changed the interpretation in 34 (47%) of studies. MPI is feasible and provides acceptable image quality for super-obese patients, although it may be camera and protocol dependent.
NASA Astrophysics Data System (ADS)
Somavarapu, Dhathri H.
This thesis proposes a new parallel computing genetic algorithm framework for designing fuel-optimal trajectories for interplanetary spacecraft missions. The framework can capture the deep search space of the problem with the use of a fixed chromosome structure and hidden-genes concept, can explore the diverse set of candidate solutions with the use of the adaptive and twin-space crowding techniques and, can execute on any high-performance computing (HPC) platform with the adoption of the portable message passing interface (MPI) standard. The algorithm is implemented in C++ with the use of the MPICH implementation of the MPI standard. The algorithm uses a patched-conic approach with two-body dynamics assumptions. New procedures are developed for determining trajectories in the Vinfinity-leveraging legs of the flight from the launch and non-launch planets and, deep-space maneuver legs of the flight from the launch and non-launch planets. The chromosome structure maintains the time of flight as a free parameter within certain boundaries. The fitness or the cost function of the algorithm uses only the mission Delta V, and does not include time of flight. The optimization is conducted with two variations for the minimum mission gravity-assist sequence, the 4-gravity-assist, and the 3-gravity-assist, with a maximum of 5 gravity-assists allowed in both the cases. The optimal trajectories discovered using the framework in both of the cases demonstrate the success of this framework.
Wilkinson, Karl A; Hine, Nicholas D M; Skylaris, Chris-Kriton
2014-11-11
We present a hybrid MPI-OpenMP implementation of Linear-Scaling Density Functional Theory within the ONETEP code. We illustrate its performance on a range of high performance computing (HPC) platforms comprising shared-memory nodes with fast interconnect. Our work has focused on applying OpenMP parallelism to the routines which dominate the computational load, attempting where possible to parallelize different loops from those already parallelized within MPI. This includes 3D FFT box operations, sparse matrix algebra operations, calculation of integrals, and Ewald summation. While the underlying numerical methods are unchanged, these developments represent significant changes to the algorithms used within ONETEP to distribute the workload across CPU cores. The new hybrid code exhibits much-improved strong scaling relative to the MPI-only code and permits calculations with a much higher ratio of cores to atoms. These developments result in a significantly shorter time to solution than was possible using MPI alone and facilitate the application of the ONETEP code to systems larger than previously feasible. We illustrate this with benchmark calculations from an amyloid fibril trimer containing 41,907 atoms. We use the code to study the mechanism of delamination of cellulose nanofibrils when undergoing sonification, a process which is controlled by a large number of interactions that collectively determine the structural properties of the fibrils. Many energy evaluations were needed for these simulations, and as these systems comprise up to 21,276 atoms this would not have been feasible without the developments described here.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moryakov, A. V., E-mail: sailor@orc.ru
2016-12-15
An algorithm for solving the linear Cauchy problem for large systems of ordinary differential equations is presented. The algorithm for systems of first-order differential equations is implemented in the EDELWEISS code with the possibility of parallel computations on supercomputers employing the MPI (Message Passing Interface) standard for the data exchange between parallel processes. The solution is represented by a series of orthogonal polynomials on the interval [0, 1]. The algorithm is characterized by simplicity and the possibility to solve nonlinear problems with a correction of the operator in accordance with the solution obtained in the previous iterative process.
2012-05-22
tabulation of the reduced space is performed using the In Situ Adaptive Tabulation ( ISAT ) algorithm. In addition, we use x2f mpi – a Fortran library...for parallel vector-valued function evaluation (used with ISAT in this context) – to efficiently redistribute the chemistry workload among the...Constrained-Equilibrium (RCCE) method, and tabulation of the reduced space is performed using the In Situ Adaptive Tabulation ( ISAT ) algorithm. In addition
Scalable Algorithms for Parallel Discrete Event Simulation Systems in Multicore Environments
2013-05-01
consolidated at the sender side. At the receiver side, the messages are deconsolidated and delivered to the appropriate thread. This approach bears some...Jiang, S. Kini, W. Yu, D. Buntinas, P. Wyckoff, and D. Panda . Performance comparison of mpi implementations over infiniband, myrinet and quadrics
ERIC Educational Resources Information Center
Haynes, Brenda J.
2010-01-01
The two-fold purpose of the study was to determine whether or not the Arkansas Leadership Academy's Master Principal Institute's (MPI) five curriculum content standards were aligned with the literature of effective educational leadership practices, and to ascertain if there were significant relationships among the principals' leadership practices…
NASA Astrophysics Data System (ADS)
Yan, Beichuan; Regueiro, Richard A.
2018-02-01
A three-dimensional (3D) DEM code for simulating complex-shaped granular particles is parallelized using message-passing interface (MPI). The concepts of link-block, ghost/border layer, and migration layer are put forward for design of the parallel algorithm, and theoretical scalability function of 3-D DEM scalability and memory usage is derived. Many performance-critical implementation details are managed optimally to achieve high performance and scalability, such as: minimizing communication overhead, maintaining dynamic load balance, handling particle migrations across block borders, transmitting C++ dynamic objects of particles between MPI processes efficiently, eliminating redundant contact information between adjacent MPI processes. The code executes on multiple US Department of Defense (DoD) supercomputers and tests up to 2048 compute nodes for simulating 10 million three-axis ellipsoidal particles. Performance analyses of the code including speedup, efficiency, scalability, and granularity across five orders of magnitude of simulation scale (number of particles) are provided, and they demonstrate high speedup and excellent scalability. It is also discovered that communication time is a decreasing function of the number of compute nodes in strong scaling measurements. The code's capability of simulating a large number of complex-shaped particles on modern supercomputers will be of value in both laboratory studies on micromechanical properties of granular materials and many realistic engineering applications involving granular materials.
DISP: Optimizations towards Scalable MPI Startup
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fu, Huansong; Pophale, Swaroop S; Gorentla Venkata, Manjunath
2016-01-01
Despite the popularity of MPI for high performance computing, the startup of MPI programs faces a scalability challenge as both the execution time and memory consumption increase drastically at scale. We have examined this problem using the collective modules of Cheetah and Tuned in Open MPI as representative implementations. Previous improvements for collectives have focused on algorithmic advances and hardware off-load. In this paper, we examine the startup cost of the collective module within a communicator and explore various techniques to improve its efficiency and scalability. Accordingly, we have developed a new scalable startup scheme with three internal techniques, namelymore » Delayed Initialization, Module Sharing and Prediction-based Topology Setup (DISP). Our DISP scheme greatly benefits the collective initialization of the Cheetah module. At the same time, it helps boost the performance of non-collective initialization in the Tuned module. We evaluate the performance of our implementation on Titan supercomputer at ORNL with up to 4096 processes. The results show that our delayed initialization can speed up the startup of Tuned and Cheetah by an average of 32.0% and 29.2%, respectively, our module sharing can reduce the memory consumption of Tuned and Cheetah by up to 24.1% and 83.5%, respectively, and our prediction-based topology setup can speed up the startup of Cheetah by up to 80%.« less
Mironov, Vladimir; Moskovsky, Alexander; D’Mello, Michael; ...
2017-10-04
The Hartree-Fock (HF) method in the quantum chemistry package GAMESS represents one of the most irregular algorithms in computation today. Major steps in the calculation are the irregular computation of electron repulsion integrals (ERIs) and the building of the Fock matrix. These are the central components of the main Self Consistent Field (SCF) loop, the key hotspot in Electronic Structure (ES) codes. By threading the MPI ranks in the official release of the GAMESS code, we not only speed up the main SCF loop (4x to 6x for large systems), but also achieve a significant (>2x) reduction in the overallmore » memory footprint. These improvements are a direct consequence of memory access optimizations within the MPI ranks. We benchmark our implementation against the official release of the GAMESS code on the Intel R Xeon PhiTM supercomputer. Here, scaling numbers are reported on up to 7,680 cores on Intel Xeon Phi coprocessors.« less
Procacci, Piero
2016-06-27
We present a new release (6.0β) of the ORAC program [Marsili et al. J. Comput. Chem. 2010, 31, 1106-1116] with a hybrid OpenMP/MPI (open multiprocessing message passing interface) multilevel parallelism tailored for generalized ensemble (GE) and fast switching double annihilation (FS-DAM) nonequilibrium technology aimed at evaluating the binding free energy in drug-receptor system on high performance computing platforms. The production of the GE or FS-DAM trajectories is handled using a weak scaling parallel approach on the MPI level only, while a strong scaling force decomposition scheme is implemented for intranode computations with shared memory access at the OpenMP level. The efficiency, simplicity, and inherent parallel nature of the ORAC implementation of the FS-DAM algorithm, project the code as a possible effective tool for a second generation high throughput virtual screening in drug discovery and design. The code, along with documentation, testing, and ancillary tools, is distributed under the provisions of the General Public License and can be freely downloaded at www.chim.unifi.it/orac .
Dynamic load balancing algorithm for molecular dynamics based on Voronoi cells domain decompositions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fattebert, J.-L.; Richards, D.F.; Glosli, J.N.
2012-12-01
We present a new algorithm for automatic parallel load balancing in classical molecular dynamics. It assumes a spatial domain decomposition of particles into Voronoi cells. It is a gradient method which attempts to minimize a cost function by displacing Voronoi sites associated with each processor/sub-domain along steepest descent directions. Excellent load balance has been obtained for quasi-2D and 3D practical applications, with up to 440·10 6 particles on 65,536 MPI tasks.
Evaluation of SuperLU on multicore architectures
NASA Astrophysics Data System (ADS)
Li, X. S.
2008-07-01
The Chip Multiprocessor (CMP) will be the basic building block for computer systems ranging from laptops to supercomputers. New software developments at all levels are needed to fully utilize these systems. In this work, we evaluate performance of different high-performance sparse LU factorization and triangular solution algorithms on several representative multicore machines. We included both Pthreads and MPI implementations in this study and found that the Pthreads implementation consistently delivers good performance and that a left-looking algorithm is usually superior.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bylaska, Eric J.; Jacquelin, Mathias; De Jong, Wibe A.
2017-10-20
Ab-initio Molecular Dynamics (AIMD) methods are an important class of algorithms, as they enable scientists to understand the chemistry and dynamics of molecular and condensed phase systems while retaining a first-principles-based description of their interactions. Many-core architectures such as the Intel® Xeon Phi™ processor are an interesting and promising target for these algorithms, as they can provide the computational power that is needed to solve interesting problems in chemistry. In this paper, we describe the efforts of refactoring the existing AIMD plane-wave method of NWChem from an MPI-only implementation to a scalable, hybrid code that employs MPI and OpenMP tomore » exploit the capabilities of current and future many-core architectures. We describe the optimizations required to get close to optimal performance for the multiplication of the tall-and-skinny matrices that form the core of the computational algorithm. We present strong scaling results on the complete AIMD simulation for a test case that simulates 256 water molecules and that strong-scales well on a cluster of 1024 nodes of Intel Xeon Phi processors. We compare the performance obtained with a cluster of dual-socket Intel® Xeon® E5–2698v3 processors.« less
BLESS 2: accurate, memory-efficient and fast error correction method.
Heo, Yun; Ramachandran, Anand; Hwu, Wen-Mei; Ma, Jian; Chen, Deming
2016-08-01
The most important features of error correction tools for sequencing data are accuracy, memory efficiency and fast runtime. The previous version of BLESS was highly memory-efficient and accurate, but it was too slow to handle reads from large genomes. We have developed a new version of BLESS to improve runtime and accuracy while maintaining a small memory usage. The new version, called BLESS 2, has an error correction algorithm that is more accurate than BLESS, and the algorithm has been parallelized using hybrid MPI and OpenMP programming. BLESS 2 was compared with five top-performing tools, and it was found to be the fastest when it was executed on two computing nodes using MPI, with each node containing twelve cores. Also, BLESS 2 showed at least 11% higher gain while retaining the memory efficiency of the previous version for large genomes. Freely available at https://sourceforge.net/projects/bless-ec dchen@illinois.edu Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
ALGORITHMS AND PROGRAMS FOR STRONG GRAVITATIONAL LENSING IN KERR SPACE-TIME INCLUDING POLARIZATION
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Bin; Maddumage, Prasad; Kantowski, Ronald
2015-05-15
Active galactic nuclei (AGNs) and quasars are important astrophysical objects to understand. Recently, microlensing observations have constrained the size of the quasar X-ray emission region to be of the order of 10 gravitational radii of the central supermassive black hole. For distances within a few gravitational radii, light paths are strongly bent by the strong gravity field of the central black hole. If the central black hole has nonzero angular momentum (spin), then a photon’s polarization plane will be rotated by the gravitational Faraday effect. The observed X-ray flux and polarization will then be influenced significantly by the strong gravitymore » field near the source. Consequently, linear gravitational lensing theory is inadequate for such extreme circumstances. We present simple algorithms computing the strong lensing effects of Kerr black holes, including the effects on polarization. Our algorithms are realized in a program “KERTAP” in two versions: MATLAB and Python. The key ingredients of KERTAP are a graphic user interface, a backward ray-tracing algorithm, a polarization propagator dealing with gravitational Faraday rotation, and algorithms computing observables such as flux magnification and polarization angles. Our algorithms can be easily realized in other programming languages such as FORTRAN, C, and C++. The MATLAB version of KERTAP is parallelized using the MATLAB Parallel Computing Toolbox and the Distributed Computing Server. The Python code was sped up using Cython and supports full implementation of MPI using the “mpi4py” package. As an example, we investigate the inclination angle dependence of the observed polarization and the strong lensing magnification of AGN X-ray emission. We conclude that it is possible to perform complex numerical-relativity related computations using interpreted languages such as MATLAB and Python.« less
Algorithms and Programs for Strong Gravitational Lensing In Kerr Space-time Including Polarization
NASA Astrophysics Data System (ADS)
Chen, Bin; Kantowski, Ronald; Dai, Xinyu; Baron, Eddie; Maddumage, Prasad
2015-05-01
Active galactic nuclei (AGNs) and quasars are important astrophysical objects to understand. Recently, microlensing observations have constrained the size of the quasar X-ray emission region to be of the order of 10 gravitational radii of the central supermassive black hole. For distances within a few gravitational radii, light paths are strongly bent by the strong gravity field of the central black hole. If the central black hole has nonzero angular momentum (spin), then a photon’s polarization plane will be rotated by the gravitational Faraday effect. The observed X-ray flux and polarization will then be influenced significantly by the strong gravity field near the source. Consequently, linear gravitational lensing theory is inadequate for such extreme circumstances. We present simple algorithms computing the strong lensing effects of Kerr black holes, including the effects on polarization. Our algorithms are realized in a program “KERTAP” in two versions: MATLAB and Python. The key ingredients of KERTAP are a graphic user interface, a backward ray-tracing algorithm, a polarization propagator dealing with gravitational Faraday rotation, and algorithms computing observables such as flux magnification and polarization angles. Our algorithms can be easily realized in other programming languages such as FORTRAN, C, and C++. The MATLAB version of KERTAP is parallelized using the MATLAB Parallel Computing Toolbox and the Distributed Computing Server. The Python code was sped up using Cython and supports full implementation of MPI using the “mpi4py” package. As an example, we investigate the inclination angle dependence of the observed polarization and the strong lensing magnification of AGN X-ray emission. We conclude that it is possible to perform complex numerical-relativity related computations using interpreted languages such as MATLAB and Python.
Oak Ridge Institutional Cluster Autotune Test Drive Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jibonananda, Sanyal; New, Joshua Ryan
2014-02-01
The Oak Ridge Institutional Cluster (OIC) provides general purpose computational resources for the ORNL staff to run computation heavy jobs that are larger than desktop applications but do not quite require the scale and power of the Oak Ridge Leadership Computing Facility (OLCF). This report details the efforts made and conclusions derived in performing a short test drive of the cluster resources on Phase 5 of the OIC. EnergyPlus was used in the analysis as a candidate user program and the overall software environment was evaluated against anticipated challenges experienced with resources such as the shared memory-Nautilus (JICS) and Titanmore » (OLCF). The OIC performed within reason and was found to be acceptable in the context of running EnergyPlus simulations. The number of cores per node and the availability of scratch space per node allow non-traditional desktop focused applications to leverage parallel ensemble execution. Although only individual runs of EnergyPlus were executed, the software environment on the OIC appeared suitable to run ensemble simulations with some modifications to the Autotune workflow. From a standpoint of general usability, the system supports common Linux libraries, compilers, standard job scheduling software (Torque/Moab), and the OpenMPI library (the only MPI library) for MPI communications. The file system is a Panasas file system which literature indicates to be an efficient file system.« less
Doyle, Mark; Pohost, Gerald M; Bairey Merz, C Noel; Shaw, Leslee J; Sopko, George; Rogers, William J; Sharaf, Barry L; Pepine, Carl J; Thompson, Diane V; Rayarao, Geetha; Tauxe, Lindsey; Kelsey, Sheryl F; Biederman, Robert W W
2016-10-01
We introduce an algorithmic approach to optimize diagnostic and prognostic value of gated cardiac single photon emission computed tomography (SPECT) and magnetic resonance (MR) myocardial perfusion imaging (MPI) modalities in women with suspected myocardial ischemia. The novel approach: bio-informatics assessment schema (BIAS) forms a mathematical model utilizing MPI data and cardiac metrics generated by one modality to predict the MPI status of another modality. The model identifies cardiac features that either enhance or mask the image-based evidence of ischemia. For each patient, the BIAS model value is used to set an appropriate threshold for the detection of ischemia. Women (n=130), with symptoms and signs of suspected myocardial ischemia, underwent MPI assessment for regional perfusion defects using two different modalities: gated SPECT and MR. To determine perfusion status, MR data were evaluated qualitatively (MRI QL ) and semi-quantitatively (MRI SQ ) while SPECT data were evaluated using conventional clinical criteria. Evaluators were masked to results of the alternate modality. These MPI status readings were designated "original". Two regression models designated "BIAS" models were generated to model MPI status obtained with one modality (e.g., MRI) compared with a second modality (e.g., SPECT), but importantly, the BIAS models did not include the primary Original MPI reading of the predicting modality. Instead, the BIAS models included auxiliary measurements like left ventricular chamber volumes and myocardial wall thickness. For each modality, the BIAS model was used to set a progressive threshold for interpretation of MPI status. Women were then followed for 38±14 months for the development of a first major adverse cardiovascular event [MACE: CV death, nonfatal myocardial infarction (MI) or hospitalization for heart failure]. Original and BIAS-augmented perfusion status were compared in their ability to detect coronary artery disease (CAD) and for prediction of MACE. Adverse events occurred in 14 (11%) women and CAD was present in 13 (10%). There was a positive correlation of maximum coronary artery stenosis and BIAS score for MRI and SPECT (P<0.001). Receiver operator characteristic (ROC) analysis was conducted and showed an increase in the area under the curve of the BIAS-augmented MPI interpretation of MACE vs . the original for MRI SQ (0.78 vs . 0.54), MRI QL (0.78 vs . 0.64), SPECT (0.82 vs . 0.63) and the average of the three readings (0.80±0.02 vs . 0.60±0.05, P<0.05). Increasing values of the BIAS score generated by both MRI and SPECT corresponded to the increasing prevalence of CAD and MACE. The BIAS-augmented detection of ischemia better predicted MACE compared with the Original reading for the MPI data for both MRI and SPECT.
Pohost, Gerald M.; Bairey Merz, C. Noel; Shaw, Leslee J.; Sopko, George; Rogers, William J.; Sharaf, Barry L.; Pepine, Carl J.; Thompson, Diane V.; Rayarao, Geetha; Tauxe, Lindsey; Kelsey, Sheryl F.; Biederman, Robert W. W.
2016-01-01
Background We introduce an algorithmic approach to optimize diagnostic and prognostic value of gated cardiac single photon emission computed tomography (SPECT) and magnetic resonance (MR) myocardial perfusion imaging (MPI) modalities in women with suspected myocardial ischemia. The novel approach: bio-informatics assessment schema (BIAS) forms a mathematical model utilizing MPI data and cardiac metrics generated by one modality to predict the MPI status of another modality. The model identifies cardiac features that either enhance or mask the image-based evidence of ischemia. For each patient, the BIAS model value is used to set an appropriate threshold for the detection of ischemia. Methods Women (n=130), with symptoms and signs of suspected myocardial ischemia, underwent MPI assessment for regional perfusion defects using two different modalities: gated SPECT and MR. To determine perfusion status, MR data were evaluated qualitatively (MRIQL) and semi-quantitatively (MRISQ) while SPECT data were evaluated using conventional clinical criteria. Evaluators were masked to results of the alternate modality. These MPI status readings were designated “original”. Two regression models designated “BIAS” models were generated to model MPI status obtained with one modality (e.g., MRI) compared with a second modality (e.g., SPECT), but importantly, the BIAS models did not include the primary Original MPI reading of the predicting modality. Instead, the BIAS models included auxiliary measurements like left ventricular chamber volumes and myocardial wall thickness. For each modality, the BIAS model was used to set a progressive threshold for interpretation of MPI status. Women were then followed for 38±14 months for the development of a first major adverse cardiovascular event [MACE: CV death, nonfatal myocardial infarction (MI) or hospitalization for heart failure]. Original and BIAS-augmented perfusion status were compared in their ability to detect coronary artery disease (CAD) and for prediction of MACE. Results Adverse events occurred in 14 (11%) women and CAD was present in 13 (10%). There was a positive correlation of maximum coronary artery stenosis and BIAS score for MRI and SPECT (P<0.001). Receiver operator characteristic (ROC) analysis was conducted and showed an increase in the area under the curve of the BIAS-augmented MPI interpretation of MACE vs. the original for MRISQ (0.78 vs. 0.54), MRIQL (0.78 vs. 0.64), SPECT (0.82 vs. 0.63) and the average of the three readings (0.80±0.02 vs. 0.60±0.05, P<0.05). Conclusions Increasing values of the BIAS score generated by both MRI and SPECT corresponded to the increasing prevalence of CAD and MACE. The BIAS-augmented detection of ischemia better predicted MACE compared with the Original reading for the MPI data for both MRI and SPECT. PMID:27747165
Lauzier, Pascal Theriault; Tang, Jie; Speidel, Michael A; Chen, Guang-Hong
2012-07-01
To achieve high temporal resolution in CT myocardial perfusion imaging (MPI), images are often reconstructed using filtered backprojection (FBP) algorithms from data acquired within a short-scan angular range. However, the variation in the central angle from one time frame to the next in gated short scans has been shown to create detrimental partial scan artifacts when performing quantitative MPI measurements. This study has two main purposes. (1) To demonstrate the existence of a distinct detrimental effect in short-scan FBP, i.e., the introduction of a nonuniform spatial image noise distribution; this nonuniformity can lead to unexpectedly high image noise and streaking artifacts, which may affect CT MPI quantification. (2) To demonstrate that statistical image reconstruction (SIR) algorithms can be a potential solution to address the nonuniform spatial noise distribution problem and can also lead to radiation dose reduction in the context of CT MPI. Projection datasets from a numerically simulated perfusion phantom and an in vivo animal myocardial perfusion CT scan were used in this study. In the numerical phantom, multiple realizations of Poisson noise were added to projection data at each time frame to investigate the spatial distribution of noise. Images from all datasets were reconstructed using both FBP and SIR reconstruction algorithms. To quantify the spatial distribution of noise, the mean and standard deviation were measured in several regions of interest (ROIs) and analyzed across time frames. In the in vivo study, two low-dose scans at tube currents of 25 and 50 mA were reconstructed using FBP and SIR. Quantitative perfusion metrics, namely, the normalized upslope (NUS), myocardial blood volume (MBV), and first moment transit time (FMT), were measured for two ROIs and compared to reference values obtained from a high-dose scan performed at 500 mA. Images reconstructed using FBP showed a highly nonuniform spatial distribution of noise. This spatial nonuniformity led to large fluctuations in the temporal direction. In the numerical phantom study, the level of noise was shown to vary by as much as 87% within a given image, and as much as 110% between different time frames for a ROI far from isocenter. The spatially nonuniform noise pattern was shown to correlate with the source trajectory and the object structure. In contrast, images reconstructed using SIR showed a highly uniform spatial distribution of noise, leading to smaller unexpected noise fluctuations in the temporal direction when a short scan angular range was used. In the numerical phantom study, the noise varied by less than 37% within a given image, and by less than 20% between different time frames. Also, the noise standard deviation in SIR images was on average half of that of FBP images. In the in vivo studies, the deviation observed between quantitative perfusion metrics measured from low-dose scans and high-dose scans was mitigated when SIR was used instead of FBP to reconstruct images. (1) Images reconstructed using FBP suffered from nonuniform spatial noise levels. This nonuniformity is another manifestation of the detrimental effects caused by short-scan reconstruction in CT MPI. (2) Images reconstructed using SIR had a much lower and more uniform noise level and thus can be used as a potential solution to address the FBP nonuniformity. (3) Given the improvement in the accuracy of the perfusion metrics when using SIR, it may be desirable to use a statistical reconstruction framework to perform low-dose dynamic CT MPI.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lauzier, Pascal Theriault; Tang Jie; Speidel, Michael A.
Purpose: To achieve high temporal resolution in CT myocardial perfusion imaging (MPI), images are often reconstructed using filtered backprojection (FBP) algorithms from data acquired within a short-scan angular range. However, the variation in the central angle from one time frame to the next in gated short scans has been shown to create detrimental partial scan artifacts when performing quantitative MPI measurements. This study has two main purposes. (1) To demonstrate the existence of a distinct detrimental effect in short-scan FBP, i.e., the introduction of a nonuniform spatial image noise distribution; this nonuniformity can lead to unexpectedly high image noise andmore » streaking artifacts, which may affect CT MPI quantification. (2) To demonstrate that statistical image reconstruction (SIR) algorithms can be a potential solution to address the nonuniform spatial noise distribution problem and can also lead to radiation dose reduction in the context of CT MPI. Methods: Projection datasets from a numerically simulated perfusion phantom and an in vivo animal myocardial perfusion CT scan were used in this study. In the numerical phantom, multiple realizations of Poisson noise were added to projection data at each time frame to investigate the spatial distribution of noise. Images from all datasets were reconstructed using both FBP and SIR reconstruction algorithms. To quantify the spatial distribution of noise, the mean and standard deviation were measured in several regions of interest (ROIs) and analyzed across time frames. In the in vivo study, two low-dose scans at tube currents of 25 and 50 mA were reconstructed using FBP and SIR. Quantitative perfusion metrics, namely, the normalized upslope (NUS), myocardial blood volume (MBV), and first moment transit time (FMT), were measured for two ROIs and compared to reference values obtained from a high-dose scan performed at 500 mA. Results: Images reconstructed using FBP showed a highly nonuniform spatial distribution of noise. This spatial nonuniformity led to large fluctuations in the temporal direction. In the numerical phantom study, the level of noise was shown to vary by as much as 87% within a given image, and as much as 110% between different time frames for a ROI far from isocenter. The spatially nonuniform noise pattern was shown to correlate with the source trajectory and the object structure. In contrast, images reconstructed using SIR showed a highly uniform spatial distribution of noise, leading to smaller unexpected noise fluctuations in the temporal direction when a short scan angular range was used. In the numerical phantom study, the noise varied by less than 37% within a given image, and by less than 20% between different time frames. Also, the noise standard deviation in SIR images was on average half of that of FBP images. In the in vivo studies, the deviation observed between quantitative perfusion metrics measured from low-dose scans and high-dose scans was mitigated when SIR was used instead of FBP to reconstruct images. Conclusions: (1) Images reconstructed using FBP suffered from nonuniform spatial noise levels. This nonuniformity is another manifestation of the detrimental effects caused by short-scan reconstruction in CT MPI. (2) Images reconstructed using SIR had a much lower and more uniform noise level and thus can be used as a potential solution to address the FBP nonuniformity. (3) Given the improvement in the accuracy of the perfusion metrics when using SIR, it may be desirable to use a statistical reconstruction framework to perform low-dose dynamic CT MPI.« less
Lauzier, Pascal Thériault; Tang, Jie; Speidel, Michael A.; Chen, Guang-Hong
2012-01-01
Purpose: To achieve high temporal resolution in CT myocardial perfusion imaging (MPI), images are often reconstructed using filtered backprojection (FBP) algorithms from data acquired within a short-scan angular range. However, the variation in the central angle from one time frame to the next in gated short scans has been shown to create detrimental partial scan artifacts when performing quantitative MPI measurements. This study has two main purposes. (1) To demonstrate the existence of a distinct detrimental effect in short-scan FBP, i.e., the introduction of a nonuniform spatial image noise distribution; this nonuniformity can lead to unexpectedly high image noise and streaking artifacts, which may affect CT MPI quantification. (2) To demonstrate that statistical image reconstruction (SIR) algorithms can be a potential solution to address the nonuniform spatial noise distribution problem and can also lead to radiation dose reduction in the context of CT MPI. Methods: Projection datasets from a numerically simulated perfusion phantom and an in vivo animal myocardial perfusion CT scan were used in this study. In the numerical phantom, multiple realizations of Poisson noise were added to projection data at each time frame to investigate the spatial distribution of noise. Images from all datasets were reconstructed using both FBP and SIR reconstruction algorithms. To quantify the spatial distribution of noise, the mean and standard deviation were measured in several regions of interest (ROIs) and analyzed across time frames. In the in vivo study, two low-dose scans at tube currents of 25 and 50 mA were reconstructed using FBP and SIR. Quantitative perfusion metrics, namely, the normalized upslope (NUS), myocardial blood volume (MBV), and first moment transit time (FMT), were measured for two ROIs and compared to reference values obtained from a high-dose scan performed at 500 mA. Results: Images reconstructed using FBP showed a highly nonuniform spatial distribution of noise. This spatial nonuniformity led to large fluctuations in the temporal direction. In the numerical phantom study, the level of noise was shown to vary by as much as 87% within a given image, and as much as 110% between different time frames for a ROI far from isocenter. The spatially nonuniform noise pattern was shown to correlate with the source trajectory and the object structure. In contrast, images reconstructed using SIR showed a highly uniform spatial distribution of noise, leading to smaller unexpected noise fluctuations in the temporal direction when a short scan angular range was used. In the numerical phantom study, the noise varied by less than 37% within a given image, and by less than 20% between different time frames. Also, the noise standard deviation in SIR images was on average half of that of FBP images. In the in vivo studies, the deviation observed between quantitative perfusion metrics measured from low-dose scans and high-dose scans was mitigated when SIR was used instead of FBP to reconstruct images. Conclusions: (1) Images reconstructed using FBP suffered from nonuniform spatial noise levels. This nonuniformity is another manifestation of the detrimental effects caused by short-scan reconstruction in CT MPI. (2) Images reconstructed using SIR had a much lower and more uniform noise level and thus can be used as a potential solution to address the FBP nonuniformity. (3) Given the improvement in the accuracy of the perfusion metrics when using SIR, it may be desirable to use a statistical reconstruction framework to perform low-dose dynamic CT MPI. PMID:22830741
The Wang Landau parallel algorithm for the simple grids. Optimizing OpenMPI parallel implementation
NASA Astrophysics Data System (ADS)
Kussainov, A. S.
2017-12-01
The Wang Landau Monte Carlo algorithm to calculate density of states for the different simple spin lattices was implemented. The energy space was split between the individual threads and balanced according to the expected runtime for the individual processes. Custom spin clustering mechanism, necessary for overcoming of the critical slowdown in the certain energy subspaces, was devised. Stable reconstruction of the density of states was of primary importance. Some data post-processing techniques were involved to produce the expected smooth density of states.
Mineccia, Michela; Zimmitti, Giuseppe; Ribero, Dario; Giraldi, Francesco; Bertolino, Franco; Brambilla, Romeo; Ferrero, Alessandro
2016-01-01
fecal peritonitis due to colorectal perforation is a dramatic event characterized by high mortality. Our study aims at determining how results of sigmoid resection (eventually extended to upper rectum) for colorectal perforation with fecal peritonitis changed in recent years and which factors affected eventual changes. Seventy-four patients were operated on at our institution (2005-2014) for colorectal perforation with fecal peritonitis and were divided into two numerically equal groups (operated on before (ERA1-group) and after (ERA2-group) May 2010). Mannheim Peritonitis Index (MPI) was calculated for each patient. Characteristics of two groups were compared. Predictors of postoperative outcomes were identified. Postoperative overall complications, major complications, and mortality occurred in 59%, 28%, and 18% of cases, respectively, and were less frequent in ERA2-group (51%, 16%, and 8%, respectively), compared to ERA1-group (68%, 41%, and 27%, respectively; p = .155, .02, and .032, respectively). Such results paralleled lower MPI values in ERA2-group, compared to ERA1-group (23(16-39) vs. 28(21-43), p = .006). Using receiver operating characteristic analysis, the best cut-off value for MPI for predicting postoperative complications and mortality was 28.5. MPI>28 was the only independent predictor of postoperative overall (p = .009, OR = 4.491) and major complications (p < .001, OR = 23.182) and was independently associated with a higher risk of mortality (p = .016, OR = 13.444), as well as duration of preoperative peritonitis longer than 24 h (p = .045, OR = 17.099). results of surgery for colorectal perforation with fecal peritonitis have improved over time, matching a concurrent decrease of MPI values and a better preoperative patient management. MPI value may help in selecting patients benefitting from surgical treatment. Copyright © 2015 IJS Publishing Group Limited. Published by Elsevier Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hjelm, Nathan Thomas; Pritchard, Howard Porter
These are a series of slides for a presentation for ExxonMobil's visit to Los Alamos National Laboratory. Topics covered are: Open MPI - The Release Story, MPI-3 RMA in Open MPI, MPI dynamic process management and Open MPI, and new options with CLE 6. Open MPI RMA features are: since v2.0.0 full support for the MPI-3.1 specification, support for non-contiguous datatypes, support for direct use of the RDMA capabilities of high performance networks (Cray Gemini/Aries, Infiniband), starting in v2.1.0 will have support for using network atomic operations for MPI_Fetch_and_op and MPI_Compare_and_swap, tested with MPI_THREAD_MULTIPLE.
Dynamic CT perfusion imaging of the myocardium: a technical note on improvement of image quality.
Muenzel, Daniela; Kabus, Sven; Gramer, Bettina; Leber, Vivian; Vembar, Mani; Schmitt, Holger; Wildgruber, Moritz; Fingerle, Alexander A; Rummeny, Ernst J; Huber, Armin; Noël, Peter B
2013-01-01
To improve image and diagnostic quality in dynamic CT myocardial perfusion imaging (MPI) by using motion compensation and a spatio-temporal filter. Dynamic CT MPI was performed using a 256-slice multidetector computed tomography scanner (MDCT). Data from two different patients-with and without myocardial perfusion defects-were evaluated to illustrate potential improvements for MPI (institutional review board approved). Three datasets for each patient were generated: (i) original data (ii) motion compensated data and (iii) motion compensated data with spatio-temporal filtering performed. In addition to the visual assessment of the tomographic slices, noise and contrast-to-noise-ratio (CNR) were measured for all data. Perfusion analysis was performed using time-density curves with regions-of-interest (ROI) placed in normal and hypoperfused myocardium. Precision in definition of normal and hypoperfused areas was determined in corresponding coloured perfusion maps. The use of motion compensation followed by spatio-temporal filtering resulted in better alignment of the cardiac volumes over time leading to a more consistent perfusion quantification and improved detection of the extend of perfusion defects. Additionally image noise was reduced by 78.5%, with CNR improvements by a factor of 4.7. The average effective radiation dose estimate was 7.1±1.1 mSv. The use of motion compensation and spatio-temporal smoothing will result in improved quantification of dynamic CT MPI using a latest generation CT scanner.
NASA Astrophysics Data System (ADS)
Coupon, Jean; Leauthaud, Alexie; Kilbinger, Martin; Medezinski, Elinor
2017-07-01
SWOT (Super W Of Theta) computes two-point statistics for very large data sets, based on “divide and conquer” algorithms, mainly, but not limited to data storage in binary trees, approximation at large scale, parellelization (open MPI), and bootstrap and jackknife resampling methods “on the fly”. It currently supports projected and 3D galaxy auto and cross correlations, galaxy-galaxy lensing, and weighted histograms.
NASA Astrophysics Data System (ADS)
Needham, Perri J.; Bhuiyan, Ashraf; Walker, Ross C.
2016-04-01
We present an implementation of explicit solvent particle mesh Ewald (PME) classical molecular dynamics (MD) within the PMEMD molecular dynamics engine, that forms part of the AMBER v14 MD software package, that makes use of Intel Xeon Phi coprocessors by offloading portions of the PME direct summation and neighbor list build to the coprocessor. We refer to this implementation as pmemd MIC offload and in this paper present the technical details of the algorithm, including basic models for MPI and OpenMP configuration, and analyze the resultant performance. The algorithm provides the best performance improvement for large systems (>400,000 atoms), achieving a ∼35% performance improvement for satellite tobacco mosaic virus (1,067,095 atoms) when 2 Intel E5-2697 v2 processors (2 ×12 cores, 30M cache, 2.7 GHz) are coupled to an Intel Xeon Phi coprocessor (Model 7120P-1.238/1.333 GHz, 61 cores). The implementation utilizes a two-fold decomposition strategy: spatial decomposition using an MPI library and thread-based decomposition using OpenMP. We also present compiler optimization settings that improve the performance on Intel Xeon processors, while retaining simulation accuracy.
NASA Astrophysics Data System (ADS)
Kracher, Daniela; Manzini, Elisa; Reick, Christian H.; Schultz, Martin; Stein, Olaf
2014-05-01
Climate change is driven by an increasing release of anthropogenic greenhouse gases (GHGs) such as carbon dioxide and nitrous oxide (N2O). Besides fossil fuel burning, also land use change and land management are anthropogenic sources of GHGs. Especially inputs of reactive nitrogen via fertilizer and deposition lead to enhanced emissions of N2O. One effect of a drastic future increase in surface temperature is a modification of atmospheric circulation, e.g. an accelerated Brewer Dobson circulation affecting the exchange between troposphere and stratosphere. N2O is inert in the troposphere and decayed only in the stratosphere. Thus, changes in atmospheric circulation, especially changes in the exchange between troposphere and stratosphere, will affect the atmospheric transport, decay, and distribution of N2O. In our study we assess the impact of global warming on atmospheric circulation and implied effects on the distribution and lifetime of atmospheric N2O. As terrestrial N2O emissions are highly determined by inputs of reactive nitrogen - the location of which being determined by human choice - we examine in particular the importance of latitudinal source regions of N2O for its global distribution. For this purpose we apply the Max Planck Institute Earth System Model, MPI-ESM. MPI-ESM consists of the atmospheric general circulation model ECHAM, the land surface model JSBACH, and MPIOM/HAMOCC representing ocean circulation and ocean biogeochemistry. Prognostic atmospheric N2O concentrations in MPI-ESM are determined by land N2O emissions, ocean N2O exchange and atmospheric tracer transport. As stratospheric chemistry is not explicitly represented in MPI-ESM, stratospheric decay rates of N2O are prescribed from a MACC MOZART simulation.
NASA Astrophysics Data System (ADS)
Schürmann, Gregor J.; Kaminski, Thomas; Köstler, Christoph; Carvalhais, Nuno; Voßbeck, Michael; Kattge, Jens; Giering, Ralf; Rödenbeck, Christian; Heimann, Martin; Zaehle, Sönke
2016-09-01
We describe the Max Planck Institute Carbon Cycle Data Assimilation System (MPI-CCDAS) built around the tangent-linear version of the JSBACH land-surface scheme, which is part of the MPI-Earth System Model v1. The simulated phenology and net land carbon balance were constrained by globally distributed observations of the fraction of absorbed photosynthetically active radiation (FAPAR, using the TIP-FAPAR product) and atmospheric CO2 at a global set of monitoring stations for the years 2005 to 2009. When constrained by FAPAR observations alone, the system successfully, and computationally efficiently, improved simulated growing-season average FAPAR, as well as its seasonality in the northern extra-tropics. When constrained by atmospheric CO2 observations alone, global net and gross carbon fluxes were improved, despite a tendency of the system to underestimate tropical productivity. Assimilating both data streams jointly allowed the MPI-CCDAS to match both observations (TIP-FAPAR and atmospheric CO2) equally well as the single data stream assimilation cases, thereby increasing the overall appropriateness of the simulated biosphere dynamics and underlying parameter values. Our study thus demonstrates the value of multiple-data-stream assimilation for the simulation of terrestrial biosphere dynamics. It further highlights the potential role of remote sensing data, here the TIP-FAPAR product, in stabilising the strongly underdetermined atmospheric inversion problem posed by atmospheric transport and CO2 observations alone. Notwithstanding these advances, the constraint of the observations on regional gross and net CO2 flux patterns on the MPI-CCDAS is limited through the coarse-scale parametrisation of the biosphere model. We expect improvement through a refined initialisation strategy and inclusion of further biosphere observations as constraints.
Distributed-Memory Computing With the Langley Aerothermodynamic Upwind Relaxation Algorithm (LAURA)
NASA Technical Reports Server (NTRS)
Riley, Christopher J.; Cheatwood, F. McNeil
1997-01-01
The Langley Aerothermodynamic Upwind Relaxation Algorithm (LAURA), a Navier-Stokes solver, has been modified for use in a parallel, distributed-memory environment using the Message-Passing Interface (MPI) standard. A standard domain decomposition strategy is used in which the computational domain is divided into subdomains with each subdomain assigned to a processor. Performance is examined on dedicated parallel machines and a network of desktop workstations. The effect of domain decomposition and frequency of boundary updates on performance and convergence is also examined for several realistic configurations and conditions typical of large-scale computational fluid dynamic analysis.
Processing MPI Datatypes Outside MPI
NASA Astrophysics Data System (ADS)
Ross, Robert; Latham, Robert; Gropp, William; Lusk, Ewing; Thakur, Rajeev
The MPI datatype functionality provides a powerful tool for describing structured memory and file regions in parallel applications, enabling noncontiguous data to be operated on by MPI communication and I/O routines. However, no facilities are provided by the MPI standard to allow users to efficiently manipulate MPI datatypes in their own codes.
X-space MPI: magnetic nanoparticles for safe medical imaging.
Goodwill, Patrick William; Saritas, Emine Ulku; Croft, Laura Rose; Kim, Tyson N; Krishnan, Kannan M; Schaffer, David V; Conolly, Steven M
2012-07-24
One quarter of all iodinated contrast X-ray clinical imaging studies are now performed on Chronic Kidney Disease (CKD) patients. Unfortunately, the iodine contrast agent used in X-ray is often toxic to CKD patients' weak kidneys, leading to significant morbidity and mortality. Hence, we are pioneering a new medical imaging method, called Magnetic Particle Imaging (MPI), to replace X-ray and CT iodinated angiography, especially for CKD patients. MPI uses magnetic nanoparticle contrast agents that are much safer than iodine for CKD patients. MPI already offers superb contrast and extraordinary sensitivity. The iron oxide nanoparticle tracers required for MPI are also used in MRI, and some are already approved for human use, but the contrast agents are far more effective at illuminating blood vessels when used in the MPI modality. We have recently developed a systems theoretic framework for MPI called x-space MPI, which has already dramatically improved the speed and robustness of MPI image reconstruction. X-space MPI has allowed us to optimize the hardware for fi ve MPI scanners. Moreover, x-space MPI provides a powerful framework for optimizing the size and magnetic properties of the iron oxide nanoparticle tracers used in MPI. Currently MPI nanoparticles have diameters in the 10-20 nanometer range, enabling millimeter-scale resolution in small animals. X-space MPI theory predicts that larger nanoparticles could enable up to 250 micrometer resolution imaging, which would represent a major breakthrough in safe imaging for CKD patients.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gittens, Alex; Devarakonda, Aditya; Racah, Evan
We explore the trade-offs of performing linear algebra using Apache Spark, compared to traditional C and MPI implementations on HPC platforms. Spark is designed for data analytics on cluster computing platforms with access to local disks and is optimized for data-parallel tasks. We examine three widely-used and important matrix factorizations: NMF (for physical plausibility), PCA (for its ubiquity) and CX (for data interpretability). We apply these methods to 1.6TB particle physics, 2.2TB and 16TB climate modeling and 1.1TB bioimaging data. The data matrices are tall-and-skinny which enable the algorithms to map conveniently into Spark’s data parallel model. We perform scalingmore » experiments on up to 1600 Cray XC40 nodes, describe the sources of slowdowns, and provide tuning guidance to obtain high performance.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sreepathi, Sarat; D'Azevedo, Eduardo; Philip, Bobby
On large supercomputers, the job scheduling systems may assign a non-contiguous node allocation for user applications depending on available resources. With parallel applications using MPI (Message Passing Interface), the default process ordering does not take into account the actual physical node layout available to the application. This contributes to non-locality in terms of physical network topology and impacts communication performance of the application. In order to mitigate such performance penalties, this work describes techniques to identify suitable task mapping that takes the layout of the allocated nodes as well as the application's communication behavior into account. During the first phasemore » of this research, we instrumented and collected performance data to characterize communication behavior of critical US DOE (United States - Department of Energy) applications using an augmented version of the mpiP tool. Subsequently, we developed several reordering methods (spectral bisection, neighbor join tree etc.) to combine node layout and application communication data for optimized task placement. We developed a tool called mpiAproxy to facilitate detailed evaluation of the various reordering algorithms without requiring full application executions. This work presents a comprehensive performance evaluation (14,000 experiments) of the various task mapping techniques in lowering communication costs on Titan, the leadership class supercomputer at Oak Ridge National Laboratory.« less
Addison, Daniel; Singh, Vinita; Okyere-Asante, K; Okafor, Henry
2014-01-01
Background: Patients presenting with chest pain and evidence of functional ischemia by myocardial perfusion imaging (MPI), but lacking commensurate angiographic disease pose a diagnostic and therapeutic dilemma. They are often dismissed as having ‘false-positive MPI’. Moreover, a majority of the available long-term outcome data for it has been derived from homogenous female populations. In this study, we sought to evaluate the long-term outcomes of this presentation in a multiethnic male-predominant cohort. Materials and Methods: We retrospectively identified 47 patients who presented to our institution between 2002 and 2005 with chest pain and evidence of ischemia on MPI, but with no significant angiographic disease on subsequent cardiac catheterization (cases). The occurrence of adverse cardiovascular outcomes (chest pain, congestive heart failure, acute myocardial infarction and stroke) post-index coronary angiogram was tracked. Similar data was collected for 37 patients who also presented with chest pain, but normal MPI over the same period (controls). Overall average follow-up was over 22 months. Results: Fifty-three percent (26/47) of the cases had one or more of the adverse outcomes as compared with 22% (8/37) of controls (P < 0.01). Of these, 13 (50.0%) and 3 (37.5%) were males, respectively. Conclusions: Ischemia on MPI is predictive of long-term adverse cardiovascular outcomes despite normal (‘false-negative’) coronary angiography. This appears to be gender-neutral. PMID:24970963
Enabling communication concurrency through flexible MPI endpoints
Dinan, James; Grant, Ryan E.; Balaji, Pavan; ...
2014-09-23
MPI defines a one-to-one relationship between MPI processes and ranks. This model captures many use cases effectively; however, it also limits communication concurrency and interoperability between MPI and programming models that utilize threads. Our paper describes the MPI endpoints extension, which relaxes the longstanding one-to-one relationship between MPI processes and ranks. Using endpoints, an MPI implementation can map separate communication contexts to threads, allowing them to drive communication independently. Also, endpoints enable threads to be addressable in MPI operations, enhancing interoperability between MPI and other programming models. Furthermore, these characteristics are illustrated through several examples and an empirical study thatmore » contrasts current multithreaded communication performance with the need for high degrees of communication concurrency to achieve peak communication performance.« less
Enabling communication concurrency through flexible MPI endpoints
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dinan, James; Grant, Ryan E.; Balaji, Pavan
MPI defines a one-to-one relationship between MPI processes and ranks. This model captures many use cases effectively; however, it also limits communication concurrency and interoperability between MPI and programming models that utilize threads. Our paper describes the MPI endpoints extension, which relaxes the longstanding one-to-one relationship between MPI processes and ranks. Using endpoints, an MPI implementation can map separate communication contexts to threads, allowing them to drive communication independently. Also, endpoints enable threads to be addressable in MPI operations, enhancing interoperability between MPI and other programming models. Furthermore, these characteristics are illustrated through several examples and an empirical study thatmore » contrasts current multithreaded communication performance with the need for high degrees of communication concurrency to achieve peak communication performance.« less
Enabling communication concurrency through flexible MPI endpoints
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dinan, James; Grant, Ryan E.; Balaji, Pavan
MPI defines a one-to-one relationship between MPI processes and ranks. This model captures many use cases effectively; however, it also limits communication concurrency and interoperability between MPI and programming models that utilize threads. This paper describes the MPI endpoints extension, which relaxes the longstanding one-to-one relationship between MPI processes and ranks. Using endpoints, an MPI implementation can map separate communication contexts to threads, allowing them to drive communication independently. Endpoints also enable threads to be addressable in MPI operations, enhancing interoperability between MPI and other programming models. These characteristics are illustrated through several examples and an empirical study that contrastsmore » current multithreaded communication performance with the need for high degrees of communication concurrency to achieve peak communication performance.« less
A parallel time integrator for noisy nonlinear oscillatory systems
NASA Astrophysics Data System (ADS)
Subber, Waad; Sarkar, Abhijit
2018-06-01
In this paper, we adapt a parallel time integration scheme to track the trajectories of noisy non-linear dynamical systems. Specifically, we formulate a parallel algorithm to generate the sample path of nonlinear oscillator defined by stochastic differential equations (SDEs) using the so-called parareal method for ordinary differential equations (ODEs). The presence of Wiener process in SDEs causes difficulties in the direct application of any numerical integration techniques of ODEs including the parareal algorithm. The parallel implementation of the algorithm involves two SDEs solvers, namely a fine-level scheme to integrate the system in parallel and a coarse-level scheme to generate and correct the required initial conditions to start the fine-level integrators. For the numerical illustration, a randomly excited Duffing oscillator is investigated in order to study the performance of the stochastic parallel algorithm with respect to a range of system parameters. The distributed implementation of the algorithm exploits Massage Passing Interface (MPI).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dorier, Matthieu; Mubarak, Misbah; Ross, Rob
Two-tiered direct network topologies such as Dragonflies have been proposed for future post-petascale and exascale machines, since they provide a high-radix, low-diameter, fast interconnection network. Such topologies call for redesigning MPI collective communication algorithms in order to attain the best performance. Yet as increasingly more applications share a machine, it is not clear how these topology-aware algorithms will react to interference with concurrent jobs accessing the same network. In this paper, we study three topology-aware broadcast algorithms, including one designed by ourselves. We evaluate their performance through event-driven simulation for small- and large-sized broadcasts (in terms of both data sizemore » and number of processes). We study the effect of different routing mechanisms on the topology-aware collective algorithms, as well as their sensitivity to network contention with other jobs. Our results show that while topology-aware algorithms dramatically reduce link utilization, their advantage in terms of latency is more limited.« less
Taking Limited English Proficient Adults into Account in the Federal Adult Education Funding Formula
ERIC Educational Resources Information Center
Capps, Randy; Fix, Michael; McHugh, Margie; Lin, Serena Yi-Ying
2009-01-01
This new report by Migration Policy Institute's (MPI's) National Center on Immigrant Integration Policy examines the funding formula used to distribute Workforce Investment Act (WIA) Title II federal funds for adult education, literacy, and English as a Second Language instruction. Though all adults with limited English proficiency (LEP) are…
Azad, Ariful; Buluç, Aydın
2016-05-16
We describe parallel algorithms for computing maximal cardinality matching in a bipartite graph on distributed-memory systems. Unlike traditional algorithms that match one vertex at a time, our algorithms process many unmatched vertices simultaneously using a matrix-algebraic formulation of maximal matching. This generic matrix-algebraic framework is used to develop three efficient maximal matching algorithms with minimal changes. The newly developed algorithms have two benefits over existing graph-based algorithms. First, unlike existing parallel algorithms, cardinality of matching obtained by the new algorithms stays constant with increasing processor counts, which is important for predictable and reproducible performance. Second, relying on bulk-synchronous matrix operations,more » these algorithms expose a higher degree of parallelism on distributed-memory platforms than existing graph-based algorithms. We report high-performance implementations of three maximal matching algorithms using hybrid OpenMP-MPI and evaluate the performance of these algorithm using more than 35 real and randomly generated graphs. On real instances, our algorithms achieve up to 200 × speedup on 2048 cores of a Cray XC30 supercomputer. Even higher speedups are obtained on larger synthetically generated graphs where our algorithms show good scaling on up to 16,384 cores.« less
Decadal prediction of European soil moisture from 1961 to 2010 using a regional climate model
NASA Astrophysics Data System (ADS)
Mieruch-Schnuelle, S.; Schädler, G.; Feldmann, H.
2014-12-01
The German national research program on decadal climate prediction(MiKlip) aims at the development of an operational decadal predictionsystem. To explore the potential of decadal predictions a hindcastensemble from 1961 to 2010 has been generated by the MPI-ESM, the newEarth system model of the Max Planck Institute for Meteorology. Toimprove the decadal predictions on higher spatial resolutions wedownscaled the MPI-ESM simulations by the regional model COSMO-CLM(CCLM) for Europe. In this study we will characterize and validatethe predictability of extreme states of soil moisture in Europesimulated by the MPI-ESM and the value added by the CCLM. The wateramount stored in the soil is a crucial component of the climate systemand especially important for agriculture, and has an influence onevaporation, groundwater and runoff. Thus, skillful prediction of soilmoisture in the order of years up to a decade could be used tomitigate risk and benefit society. Since soil moisture observationsare rare and validation of model output is difficult, we will ratherinvestigate the effective drought index (EDI), which can be retrievedsolely from precipitation data. Therefore we show that the EDI is agood estimator of the soil water content.
Hierarchical Petascale Simulation Framework For Stress Corrosion Cracking
DOE Office of Scientific and Technical Information (OSTI.GOV)
Grama, Ananth
2013-12-18
A number of major accomplishments resulted from the project. These include: • Data Structures, Algorithms, and Numerical Methods for Reactive Molecular Dynamics. We have developed a range of novel data structures, algorithms, and solvers (amortized ILU, Spike) for use with ReaxFF and charge equilibration. • Parallel Formulations of ReactiveMD (Purdue ReactiveMolecular Dynamics Package, PuReMD, PuReMD-GPU, and PG-PuReMD) for Messaging, GPU, and GPU Cluster Platforms. We have developed efficient serial, parallel (MPI), GPU (Cuda), and GPU Cluster (MPI/Cuda) implementations. Our implementations have been demonstrated to be significantly better than the state of the art, both in terms of performance and scalability.more » • Comprehensive Validation in the Context of Diverse Applications. We have demonstrated the use of our software in diverse systems, including silica-water, silicon-germanium nanorods, and as part of other projects, extended it to applications ranging from explosives (RDX) to lipid bilayers (biomembranes under oxidative stress). • Open Source Software Packages for Reactive Molecular Dynamics. All versions of our soft- ware have been released over the public domain. There are over 100 major research groups worldwide using our software. • Implementation into the Department of Energy LAMMPS Software Package. We have also integrated our software into the Department of Energy LAMMPS software package.« less
In situ and in-transit analysis of cosmological simulations
Friesen, Brian; Almgren, Ann; Lukic, Zarija; ...
2016-08-24
Modern cosmological simulations have reached the trillion-element scale, rendering data storage and subsequent analysis formidable tasks. To address this circumstance, we present a new MPI-parallel approach for analysis of simulation data while the simulation runs, as an alternative to the traditional workflow consisting of periodically saving large data sets to disk for subsequent ‘offline’ analysis. We demonstrate this approach in the compressible gasdynamics/N-body code Nyx, a hybrid MPI+OpenMP code based on the BoxLib framework, used for large-scale cosmological simulations. We have enabled on-the-fly workflows in two different ways: one is a straightforward approach consisting of all MPI processes periodically haltingmore » the main simulation and analyzing each component of data that they own (‘ in situ’). The other consists of partitioning processes into disjoint MPI groups, with one performing the simulation and periodically sending data to the other ‘sidecar’ group, which post-processes it while the simulation continues (‘in-transit’). The two groups execute their tasks asynchronously, stopping only to synchronize when a new set of simulation data needs to be analyzed. For both the in situ and in-transit approaches, we experiment with two different analysis suites with distinct performance behavior: one which finds dark matter halos in the simulation using merge trees to calculate the mass contained within iso-density contours, and another which calculates probability distribution functions and power spectra of various fields in the simulation. Both are common analysis tasks for cosmology, and both result in summary statistics significantly smaller than the original data set. We study the behavior of each type of analysis in each workflow in order to determine the optimal configuration for the different data analysis algorithms.« less
NASA Astrophysics Data System (ADS)
Bellerby, Tim
2015-04-01
PM (Parallel Models) is a new parallel programming language specifically designed for writing environmental and geophysical models. The language is intended to enable implementers to concentrate on the science behind the model rather than the details of running on parallel hardware. At the same time PM leaves the programmer in control - all parallelisation is explicit and the parallel structure of any given program may be deduced directly from the code. This paper describes a PM implementation based on the Message Passing Interface (MPI) and Open Multi-Processing (OpenMP) standards, looking at issues involved with translating the PM parallelisation model to MPI/OpenMP protocols and considering performance in terms of the competing factors of finer-grained parallelisation and increased communication overhead. In order to maximise portability, the implementation stays within the MPI 1.3 standard as much as possible, with MPI-2 MPI-IO file handling the only significant exception. Moreover, it does not assume a thread-safe implementation of MPI. PM adopts a two-tier abstract representation of parallel hardware. A PM processor is a conceptual unit capable of efficiently executing a set of language tasks, with a complete parallel system consisting of an abstract N-dimensional array of such processors. PM processors may map to single cores executing tasks using cooperative multi-tasking, to multiple cores or even to separate processing nodes, efficiently sharing tasks using algorithms such as work stealing. While tasks may move between hardware elements within a PM processor, they may not move between processors without specific programmer intervention. Tasks are assigned to processors using a nested parallelism approach, building on ideas from Reyes et al. (2009). The main program owns all available processors. When the program enters a parallel statement then either processors are divided out among the newly generated tasks (number of new tasks < number of processors) or tasks are divided out among the available processors (number of tasks > number of processors). Nested parallel statements may further subdivide the processor set owned by a given task. Tasks or processors are distributed evenly by default, but uneven distributions are possible under programmer control. It is also possible to explicitly enable child tasks to migrate within the processor set owned by their parent task, reducing load unbalancing at the potential cost of increased inter-processor message traffic. PM incorporates some programming structures from the earlier MIST language presented at a previous EGU General Assembly, while adopting a significantly different underlying parallelisation model and type system. PM code is available at www.pm-lang.org under an unrestrictive MIT license. Reference Ruymán Reyes, Antonio J. Dorta, Francisco Almeida, Francisco de Sande, 2009. Automatic Hybrid MPI+OpenMP Code Generation with llc, Recent Advances in Parallel Virtual Machine and Message Passing Interface, Lecture Notes in Computer Science Volume 5759, 185-195
An MPI-1 Compliant Thread-Based Implementation
NASA Astrophysics Data System (ADS)
Díaz Martín, J. C.; Rico Gallego, J. A.; Álvarez Llorente, J. M.; Perogil Duque, J. F.
This work presents AzequiaMPI, the first full compliant implementation of the MPI-1 standard where the MPI node is a thread. Performance comparisons with MPICH2-Nemesis show that thread-based implementations exploit adequately the multicore architectures under oversubscription, what could make MPI competitive with OpenMP-like solutions.
An implementation and evaluation of the MPI 3.0 one-sided communication interface
Dinan, James S.; Balaji, Pavan; Buntinas, Darius T.; ...
2016-01-09
The Q1 Message Passing Interface (MPI) 3.0 standard includes a significant revision to MPI’s remote memory access (RMA) interface, which provides support for one-sided communication. MPI-3 RMA is expected to greatly enhance the usability and performance ofMPI RMA.We present the first complete implementation of MPI-3 RMA and document implementation techniques and performance optimization opportunities enabled by the new interface. Our implementation targets messaging-based networks and is publicly available in the latest release of the MPICH MPI implementation. Here using this implementation, we explore the performance impact of new MPI-3 functionality and semantics. Results indicate that the MPI-3 RMA interface providesmore » significant advantages over the MPI-2 interface by enabling increased communication concurrency through relaxed semantics in the interface and additional routines that provide new window types, synchronization modes, and atomic operations.« less
An implementation and evaluation of the MPI 3.0 one-sided communication interface
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dinan, James S.; Balaji, Pavan; Buntinas, Darius T.
The Q1 Message Passing Interface (MPI) 3.0 standard includes a significant revision to MPI’s remote memory access (RMA) interface, which provides support for one-sided communication. MPI-3 RMA is expected to greatly enhance the usability and performance ofMPI RMA.We present the first complete implementation of MPI-3 RMA and document implementation techniques and performance optimization opportunities enabled by the new interface. Our implementation targets messaging-based networks and is publicly available in the latest release of the MPICH MPI implementation. Here using this implementation, we explore the performance impact of new MPI-3 functionality and semantics. Results indicate that the MPI-3 RMA interface providesmore » significant advantages over the MPI-2 interface by enabling increased communication concurrency through relaxed semantics in the interface and additional routines that provide new window types, synchronization modes, and atomic operations.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Collins, Benjamin S.
The Futility package contains the following: 1) Definition of the size of integers and real numbers; 2) A generic Unit test harness; 3) Definitions for some basic extensions to the Fortran language: arbitrary length strings, a parameter list construct, exception handlers, command line processor, timers; 4) Geometry definitions: point, line, plane, box, cylinder, polyhedron; 5) File wrapper functions: standard Fortran input/output files, Fortran binary files, HDF5 files; 6) Parallel wrapper functions: MPI, and Open MP abstraction layers, partitioning algorithms; 7) Math utilities: BLAS, Matrix and Vector definitions, Linear Solver methods and wrappers for other TPLs (PETSC, MKL, etc), preconditioner classes;more » 8) Misc: random number generator, water saturation properties, sorting algorithms.« less
Regional-scale calculation of the LS factor using parallel processing
NASA Astrophysics Data System (ADS)
Liu, Kai; Tang, Guoan; Jiang, Ling; Zhu, A.-Xing; Yang, Jianyi; Song, Xiaodong
2015-05-01
With the increase of data resolution and the increasing application of USLE over large areas, the existing serial implementation of algorithms for computing the LS factor is becoming a bottleneck. In this paper, a parallel processing model based on message passing interface (MPI) is presented for the calculation of the LS factor, so that massive datasets at a regional scale can be processed efficiently. The parallel model contains algorithms for calculating flow direction, flow accumulation, drainage network, slope, slope length and the LS factor. According to the existence of data dependence, the algorithms are divided into local algorithms and global algorithms. Parallel strategy are designed according to the algorithm characters including the decomposition method for maintaining the integrity of the results, optimized workflow for reducing the time taken for exporting the unnecessary intermediate data and a buffer-communication-computation strategy for improving the communication efficiency. Experiments on a multi-node system show that the proposed parallel model allows efficient calculation of the LS factor at a regional scale with a massive dataset.
A De-centralized Scheduling and Load Balancing Algorithm for Heterogeneous Grid Environments
NASA Technical Reports Server (NTRS)
Arora, Manish; Das, Sajal K.; Biswas, Rupak
2002-01-01
In the past two decades, numerous scheduling and load balancing techniques have been proposed for locally distributed multiprocessor systems. However, they all suffer from significant deficiencies when extended to a Grid environment: some use a centralized approach that renders the algorithm unscalable, while others assume the overhead involved in searching for appropriate resources to be negligible. Furthermore, classical scheduling algorithms do not consider a Grid node to be N-resource rich and merely work towards maximizing the utilization of one of the resources. In this paper, we propose a new scheduling and load balancing algorithm for a generalized Grid model of N-resource nodes that not only takes into account the node and network heterogeneity, but also considers the overhead involved in coordinating among the nodes. Our algorithm is decentralized, scalable, and overlaps the node coordination time with that of the actual processing of ready jobs, thus saving valuable clock cycles needed for making decisions. The proposed algorithm is studied by conducting simulations using the Message Passing Interface (MPI) paradigm.
A De-Centralized Scheduling and Load Balancing Algorithm for Heterogeneous Grid Environments
NASA Technical Reports Server (NTRS)
Arora, Manish; Das, Sajal K.; Biswas, Rupak; Biegel, Bryan (Technical Monitor)
2002-01-01
In the past two decades, numerous scheduling and load balancing techniques have been proposed for locally distributed multiprocessor systems. However, they all suffer from significant deficiencies when extended to a Grid environment: some use a centralized approach that renders the algorithm unscalable, while others assume the overhead involved in searching for appropriate resources to be negligible. Furthermore, classical scheduling algorithms do not consider a Grid node to be N-resource rich and merely work towards maximizing the utilization of one of the resources. In this paper we propose a new scheduling and load balancing algorithm for a generalized Grid model of N-resource nodes that not only takes into account the node and network heterogeneity, but also considers the overhead involved in coordinating among the nodes. Our algorithm is de-centralized, scalable, and overlaps the node coordination time of the actual processing of ready jobs, thus saving valuable clock cycles needed for making decisions. The proposed algorithm is studied by conducting simulations using the Message Passing Interface (MPI) paradigm.
Scalable Domain Decomposed Monte Carlo Particle Transport
NASA Astrophysics Data System (ADS)
O'Brien, Matthew Joseph
In this dissertation, we present the parallel algorithms necessary to run domain decomposed Monte Carlo particle transport on large numbers of processors (millions of processors). Previous algorithms were not scalable, and the parallel overhead became more computationally costly than the numerical simulation. The main algorithms we consider are: • Domain decomposition of constructive solid geometry: enables extremely large calculations in which the background geometry is too large to fit in the memory of a single computational node. • Load Balancing: keeps the workload per processor as even as possible so the calculation runs efficiently. • Global Particle Find: if particles are on the wrong processor, globally resolve their locations to the correct processor based on particle coordinate and background domain. • Visualizing constructive solid geometry, sourcing particles, deciding that particle streaming communication is completed and spatial redecomposition. These algorithms are some of the most important parallel algorithms required for domain decomposed Monte Carlo particle transport. We demonstrate that our previous algorithms were not scalable, prove that our new algorithms are scalable, and run some of the algorithms up to 2 million MPI processes on the Sequoia supercomputer.
NASA Astrophysics Data System (ADS)
Al-Refaie, Ahmed F.; Tennyson, Jonathan
2017-12-01
Construction and diagonalization of the Hamiltonian matrix is the rate-limiting step in most low-energy electron - molecule collision calculations. Tennyson (1996) implemented a novel algorithm for Hamiltonian construction which took advantage of the structure of the wavefunction in such calculations. This algorithm is re-engineered to make use of modern computer architectures and the use of appropriate diagonalizers is considered. Test calculations demonstrate that significant speed-ups can be gained using multiple CPUs. This opens the way to calculations which consider higher collision energies, larger molecules and / or more target states. The methodology, which is implemented as part of the UK molecular R-matrix codes (UKRMol and UKRMol+) can also be used for studies of bound molecular Rydberg states, photoionization and positron-molecule collisions.
Parallel deterministic neutronics with AMR in 3D
DOE Office of Scientific and Technical Information (OSTI.GOV)
Clouse, C.; Ferguson, J.; Hendrickson, C.
1997-12-31
AMTRAN, a three dimensional Sn neutronics code with adaptive mesh refinement (AMR) has been parallelized over spatial domains and energy groups and runs on the Meiko CS-2 with MPI message passing. Block refined AMR is used with linear finite element representations for the fluxes, which allows for a straight forward interpretation of fluxes at block interfaces with zoning differences. The load balancing algorithm assumes 8 spatial domains, which minimizes idle time among processors.
MPI Runtime Error Detection with MUST: Advances in Deadlock Detection
Hilbrich, Tobias; Protze, Joachim; Schulz, Martin; ...
2013-01-01
The widely used Message Passing Interface (MPI) is complex and rich. As a result, application developers require automated tools to avoid and to detect MPI programming errors. We present the Marmot Umpire Scalable Tool (MUST) that detects such errors with significantly increased scalability. We present improvements to our graph-based deadlock detection approach for MPI, which cover future MPI extensions. Our enhancements also check complex MPI constructs that no previous graph-based detection approach handled correctly. Finally, we present optimizations for the processing of MPI operations that reduce runtime deadlock detection overheads. Existing approaches often require ( p ) analysis time permore » MPI operation, for p processes. We empirically observe that our improvements lead to sub-linear or better analysis time per operation for a wide range of real world applications.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moody, Adam
2007-05-22
MpiGraph consists of an MPI application called mpiGraph written in C to measure message bandwidth and an associated crunch_mpiGraph script written in Perl to process the application output into an HTMO report. The mpiGraph application is designed to inspect the health and scalability of a high-performance interconnect while under heavy load. This is useful to detect hardware and software problems in a system, such as slow nodes, links, switches, or contention in switch routing. It is also useful to characterize how interconnect performance changes with different settings or how one interconnect type compares to another.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fang, Aiman; Laguna, Ignacio; Sato, Kento
Future high-performance computing systems may face frequent failures with their rapid increase in scale and complexity. Resilience to faults has become a major challenge for large-scale applications running on supercomputers, which demands fault tolerance support for prevalent MPI applications. Among failure scenarios, process failures are one of the most severe issues as they usually lead to termination of applications. However, the widely used MPI implementations do not provide mechanisms for fault tolerance. We propose FTA-MPI (Fault Tolerance Assistant MPI), a programming model that provides support for failure detection, failure notification and recovery. Specifically, FTA-MPI exploits a try/catch model that enablesmore » failure localization and transparent recovery of process failures in MPI applications. We demonstrate FTA-MPI with synthetic applications and a molecular dynamics code CoMD, and show that FTA-MPI provides high programmability for users and enables convenient and flexible recovery of process failures.« less
New Kind of Ellis Island as Second-Generation Immigrants Land on College Campuses
ERIC Educational Resources Information Center
Hohn, Marcia Drew; Mohammed, Denzil
2012-01-01
Demographics in American higher education are changing dramatically. A recent study by the Migration Policy Institute (MPI) reveals that 11.3 million people ages 16 to 26 (one in four) are first- and second-generation immigrants. Moreover, the report continues, between 1995 and 2010, immigrant-origin youth accounted for half of all growth in the…
NASA Astrophysics Data System (ADS)
Kracher, D.; Manzini, E.; Reick, C. H.; Schultz, M. G.; Stein, O.
2014-12-01
Greenhouse gas induced climate change will modify the physical conditions of the atmosphere. One of the projected changes is an acceleration of the Brewer-Dobson circulation in the stratosphere, as it has been shown in many model studies. This change in the stratospheric circulation consequently bears an effect on the transport and distribution of atmospheric components such as N2O. Since N2O is involved in ozone destruction, a modified distribution of N2O can be of importance for ozone chemistry. N2O is inert in the troposphere and decays only in the stratosphere. Thus, changes in the exchange between troposphere and stratosphere can also affect the stratospheric sink of N2O, and consequently its atmospheric lifetime. N2O is a potent greenhouse gas with a global warming potential of currently approximately 300 CO2-equivalents in a 100-year perspective. A faster decay in atmospheric N2O mixing ratios, i.e. a decreased atmospheric lifetime of N2O, will also reduce its global warming potential. In order to assess the impact of climate change on atmospheric circulation and implied effects on the distribution and lifetime of atmospheric N2O, we apply the Max Planck Institute Earth System Model, MPI-ESM. MPI-ESM consists of the atmospheric general circulation model ECHAM, the land surface model JSBACH, and MPIOM/HAMOCC representing ocean circulation and ocean biogeochemistry. Prognostic atmospheric N2O concentrations in MPI-ESM are determined by land N2O emissions, ocean-atmosphere N2O exchange and atmospheric tracer transport. As stratospheric chemistry is not explicitly represented in MPI-ESM, stratospheric decay rates of N2O are prescribed from a MACC MOZART simulation. Increasing surface temperatures and CO2 concentrations in the stratosphere impact atmospheric circulation differently. Thus, we conduct a series of transient runs with the atmospheric model of MPI-ESM to isolate different factors governing a shift in atmospheric circulation. From those transient simulations we diagnose decreasing tropospheric N2O concentrations, increased transport of N2O from the troposphere to the stratosphere, and increasing stratospheric decay of N2O leading to a reduction in atmospheric lifetime of N2O, in dependency to climate change evolution.
NASA Astrophysics Data System (ADS)
Schuster, Mareike; Thürkow, Markus; Weiher, Stefan; Kirchner, Ingo; Ulbrich, Uwe; Will, Andreas
2016-04-01
A general bias of global atmosphere ocean models, and also of the MPI-ESM, is an under-representation of the high latitude cyclone activity and an overestimation of the mid latitude cyclone activity in the North Atlantic, thus representing the extra-tropical storm track too zonal. We will show, that this effect can be antagonized by applying an atmospheric Two-Way Coupling (TWC). In this study we present a newly developed Two-Way Coupled model system, which is based on the MPI-ESM, and show that it is able to capture the mean storm track location more accurate. It also influences the sub-decadal deterministic predictability of extra-tropical cyclones and shows significantly enhanced skill compared to the "uncoupled" MPI-ESM standalone system. This study evaluates a set of hindcast experiments performed with said Two-Way Coupled model system. The regional model COSMO CLM is Two-Way Coupled to the atmosphere of the global Max-Plack-Institute Earth System Model (MPI-ESM) and therefore integrates and exchanges the state of the atmosphere every 10 minutes (MPI-TWC-ESM). In the coupled source region (North Atlantic), mesoscale processes which are relevant for the formation and early-stage development of cyclones are expected to be better represented, and therefore influence the large scale dynamics of the target region (Europe). The database covers 102 "uncoupled" years and 102 Two-Way Coupled years of the recent climate (1960-2010). Results are validated against the ERA-Interim reanalysis. Besides the climatological point of view, the design of this single model ensemble allows for an analysis of the predictability of the first and second leadyears of the hindcasts. As a first step to understand the improved predictability of cyclones, we will show a detailed analysis of climatologies for specific cyclone categories, sorted by season and region. Especially for cyclones affecting Europe, the TWC is capable to counteract the AOGCM's biases in the North Atlantic. Also, cyclones which are generated in the northern North Atlantic and the Labrador Sea are to an extraordinary extent underestimated in the "uncoupled" MPI-ESM - for the latter region the TWC can balance this shortcoming. In the Northern Hemisphere annual mean statistics the TWC does not change the distribution of the strength of cyclones, but it changes the distribution of the lifetime of cyclones.
Using WEED to simulate the global wetland distribution in a ESM
NASA Astrophysics Data System (ADS)
Stacke, Tobias; Hagemann, Stefan
2016-04-01
Lakes and wetlands are an important land surface feature. In terms of hydrology, they regulate river discharge, mitigate flood events and constitute a significant surface water storage. Considering physical processes, they link the surface water and energy balances by altering the separation of incoming energy into sensible and latent heat fluxes. Finally, they impact biogeochemical processes and may act as carbon sinks or sources. Most global hydrology and climate models regard wetland extent and properties as constant in time. However, to study interactions between wetlands and different states of climate, it is necessary to implement surface water bodies (thereafter referred to as wetlands) with dynamical behavior into these models. Besides an improved representation of geophysical feedbacks between wetlands, land surface and atmosphere, a dynamical wetland scheme could also provide estimates of soil wetness as input for biogeochemical models, which are used to compute methane production in wetlands. Recently, a model for the representation of wetland extent dynamics (WEED) was developed as part of the hydrology model (MPI-HM) of the Max-Planck-Institute for Meteorology (MPI-M). The WEED scheme computes wetland extent in agreement with the range of observations for the high northern latitudes. It simulates a realistic seasonal cycle which shows sensitivity to northern snow-melt as well as rainy seasons in the tropics. Furthermore, flood peaks in river discharge are mitigated. However, the WEED scheme overestimates wetland extent in the Tropics which might be related to the MPI-HM's simplified potential evapotranspiration computation. In order to overcome this limitation, the WEED scheme is implemented into the MPI-M's land surface model JSBACH. Thus, not only its effect on water fluxes can be investigated but also its impact on the energy cycle, which is not included in the MPI-HM. Furthermore, it will be possible to analyze the physical effects of wetlands in a coupled land-atmosphere simulation. First simulations with JSBACH-WEED show results similar to the MPI-HM simulations. As the next step, the scheme is modified to account for energy cycle relevant issues such as the dynamical alteration of surface albedo as well as the allocation of appropriate thermal properties to the wetlands. In our presentation, we will give an overview on the functionality of the WEED scheme and the effect of wetlands in coupled land-atmosphere simulations.
HPC-NMF: A High-Performance Parallel Algorithm for Nonnegative Matrix Factorization
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kannan, Ramakrishnan; Sukumar, Sreenivas R.; Ballard, Grey M.
NMF is a useful tool for many applications in different domains such as topic modeling in text mining, background separation in video analysis, and community detection in social networks. Despite its popularity in the data mining community, there is a lack of efficient distributed algorithms to solve the problem for big data sets. We propose a high-performance distributed-memory parallel algorithm that computes the factorization by iteratively solving alternating non-negative least squares (NLS) subproblems formore » $$\\WW$$ and $$\\HH$$. It maintains the data and factor matrices in memory (distributed across processors), uses MPI for interprocessor communication, and, in the dense case, provably minimizes communication costs (under mild assumptions). As opposed to previous implementation, our algorithm is also flexible: It performs well for both dense and sparse matrices, and allows the user to choose any one of the multiple algorithms for solving the updates to low rank factors $$\\WW$$ and $$\\HH$$ within the alternating iterations.« less
An efficicient data structure for three-dimensional vertex based finite volume method
NASA Astrophysics Data System (ADS)
Akkurt, Semih; Sahin, Mehmet
2017-11-01
A vertex based three-dimensional finite volume algorithm has been developed using an edge based data structure.The mesh data structure of the given algorithm is similar to ones that exist in the literature. However, the data structures are redesigned and simplied in order to fit requirements of the vertex based finite volume method. In order to increase the cache efficiency, the data access patterns for the vertex based finite volume method are investigated and these datas are packed/allocated in a way that they are close to each other in the memory. The present data structure is not limited with tetrahedrons, arbitrary polyhedrons are also supported in the mesh without putting any additional effort. Furthermore, the present data structure also supports adaptive refinement and coarsening. For the implicit and parallel implementation of the FVM algorithm, PETSc and MPI libraries are employed. The performance and accuracy of the present algorithm are tested for the classical benchmark problems by comparing the CPU time for the open source algorithms.
Parallelization of Nullspace Algorithm for the computation of metabolic pathways
Jevremović, Dimitrije; Trinh, Cong T.; Srienc, Friedrich; Sosa, Carlos P.; Boley, Daniel
2011-01-01
Elementary mode analysis is a useful metabolic pathway analysis tool in understanding and analyzing cellular metabolism, since elementary modes can represent metabolic pathways with unique and minimal sets of enzyme-catalyzed reactions of a metabolic network under steady state conditions. However, computation of the elementary modes of a genome- scale metabolic network with 100–1000 reactions is very expensive and sometimes not feasible with the commonly used serial Nullspace Algorithm. In this work, we develop a distributed memory parallelization of the Nullspace Algorithm to handle efficiently the computation of the elementary modes of a large metabolic network. We give an implementation in C++ language with the support of MPI library functions for the parallel communication. Our proposed algorithm is accompanied with an analysis of the complexity and identification of major bottlenecks during computation of all possible pathways of a large metabolic network. The algorithm includes methods to achieve load balancing among the compute-nodes and specific communication patterns to reduce the communication overhead and improve efficiency. PMID:22058581
Characterizing MPI matching via trace-based simulation
Ferreira, Kurt Brian; Levy, Scott Larson Nicoll; Pedretti, Kevin; ...
2017-01-01
With the increased scale expected on future leadership-class systems, detailed information about the resource usage and performance of MPI message matching provides important insights into how to maintain application performance on next-generation systems. However, obtaining MPI message matching performance data is often not possible without significant effort. A common approach is to instrument an MPI implementation to collect relevant statistics. While this approach can provide important data, collecting matching data at runtime perturbs the application's execution, including its matching performance, and is highly dependent on the MPI library's matchlist implementation. In this paper, we introduce a trace-based simulation approach tomore » obtain detailed MPI message matching performance data for MPI applications without perturbing their execution. Using a number of key parallel workloads, we demonstrate that this simulator approach can rapidly and accurately characterize matching behavior. Specifically, we use our simulator to collect several important statistics about the operation of the MPI posted and unexpected queues. For example, we present data about search lengths and the duration that messages spend in the queues waiting to be matched. Here, data gathered using this simulation-based approach have significant potential to aid hardware designers in determining resource allocation for MPI matching functions and provide application and middleware developers with insight into the scalability issues associated with MPI message matching.« less
High-performance iron oxide nanoparticles for magnetic particle imaging - guided hyperthermia (hMPI)
NASA Astrophysics Data System (ADS)
Bauer, Lisa M.; Situ, Shu F.; Griswold, Mark A.; Samia, Anna Cristina S.
2016-06-01
Magnetic particle imaging (MPI) is an emerging imaging modality that allows the direct and quantitative mapping of iron oxide nanoparticles. In MPI, the development of tailored iron oxide nanoparticle tracers is paramount to achieving high sensitivity and good spatial resolution. To date, most MPI tracers being developed for potential clinical applications are based on spherical undoped magnetite nanoparticles. For the first time, we report on the systematic investigation of the effects of changes in chemical composition and shape anisotropy on the MPI performance of iron oxide nanoparticle tracers. We observed a 2-fold enhancement in MPI signal through selective doping of magnetite nanoparticles with zinc. Moreover, we demonstrated focused magnetic hyperthermia heating by adapting the field gradient used in MPI. By saturating the iron oxide nanoparticles outside of a field free region (FFR) with an external static field, we can selectively heat a target region in our test sample. By comparing zinc-doped magnetite cubic nanoparticles with undoped spherical nanoparticles, we could show a 5-fold improvement in the specific absorption rate (SAR) in magnetic hyperthermia while providing good MPI signal, thereby demonstrating the potential for high-performance focused hyperthermia therapy through an MPI-guided approach (hMPI).Magnetic particle imaging (MPI) is an emerging imaging modality that allows the direct and quantitative mapping of iron oxide nanoparticles. In MPI, the development of tailored iron oxide nanoparticle tracers is paramount to achieving high sensitivity and good spatial resolution. To date, most MPI tracers being developed for potential clinical applications are based on spherical undoped magnetite nanoparticles. For the first time, we report on the systematic investigation of the effects of changes in chemical composition and shape anisotropy on the MPI performance of iron oxide nanoparticle tracers. We observed a 2-fold enhancement in MPI signal through selective doping of magnetite nanoparticles with zinc. Moreover, we demonstrated focused magnetic hyperthermia heating by adapting the field gradient used in MPI. By saturating the iron oxide nanoparticles outside of a field free region (FFR) with an external static field, we can selectively heat a target region in our test sample. By comparing zinc-doped magnetite cubic nanoparticles with undoped spherical nanoparticles, we could show a 5-fold improvement in the specific absorption rate (SAR) in magnetic hyperthermia while providing good MPI signal, thereby demonstrating the potential for high-performance focused hyperthermia therapy through an MPI-guided approach (hMPI). Electronic supplementary information (ESI) available: Detailed IONP synthetic methods, description of magnetic particle relaxometer set-up, TEM of reference IONP (Senior Scientific PrecisionMRX™ 25 nm oleic acid-coated nanoparticles), concentration dependent PSF of all IONP samples, PSF and SAR of Zn-Sph and Zn-Cube mixture sample, upper right quadrant of field-dependent hysteresis curve labelled with static field strengths, and the magnetic hyperthermia temperature profiles with and without the presence of external magnetic fields. See DOI: 10.1039/c6nr01877g
Petascale turbulence simulation using a highly parallel fast multipole method on GPUs
NASA Astrophysics Data System (ADS)
Yokota, Rio; Barba, L. A.; Narumi, Tetsu; Yasuoka, Kenji
2013-03-01
This paper reports large-scale direct numerical simulations of homogeneous-isotropic fluid turbulence, achieving sustained performance of 1.08 petaflop/s on GPU hardware using single precision. The simulations use a vortex particle method to solve the Navier-Stokes equations, with a highly parallel fast multipole method (FMM) as numerical engine, and match the current record in mesh size for this application, a cube of 40963 computational points solved with a spectral method. The standard numerical approach used in this field is the pseudo-spectral method, relying on the FFT algorithm as the numerical engine. The particle-based simulations presented in this paper quantitatively match the kinetic energy spectrum obtained with a pseudo-spectral method, using a trusted code. In terms of parallel performance, weak scaling results show the FMM-based vortex method achieving 74% parallel efficiency on 4096 processes (one GPU per MPI process, 3 GPUs per node of the TSUBAME-2.0 system). The FFT-based spectral method is able to achieve just 14% parallel efficiency on the same number of MPI processes (using only CPU cores), due to the all-to-all communication pattern of the FFT algorithm. The calculation time for one time step was 108 s for the vortex method and 154 s for the spectral method, under these conditions. Computing with 69 billion particles, this work exceeds by an order of magnitude the largest vortex-method calculations to date.
Fault Tolerant Frequent Pattern Mining
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shohdy, Sameh; Vishnu, Abhinav; Agrawal, Gagan
FP-Growth algorithm is a Frequent Pattern Mining (FPM) algorithm that has been extensively used to study correlations and patterns in large scale datasets. While several researchers have designed distributed memory FP-Growth algorithms, it is pivotal to consider fault tolerant FP-Growth, which can address the increasing fault rates in large scale systems. In this work, we propose a novel parallel, algorithm-level fault-tolerant FP-Growth algorithm. We leverage algorithmic properties and MPI advanced features to guarantee an O(1) space complexity, achieved by using the dataset memory space itself for checkpointing. We also propose a recovery algorithm that can use in-memory and disk-based checkpointing,more » though in many cases the recovery can be completed without any disk access, and incurring no memory overhead for checkpointing. We evaluate our FT algorithm on a large scale InfiniBand cluster with several large datasets using up to 2K cores. Our evaluation demonstrates excellent efficiency for checkpointing and recovery in comparison to the disk-based approach. We have also observed 20x average speed-up in comparison to Spark, establishing that a well designed algorithm can easily outperform a solution based on a general fault-tolerant programming model.« less
Detection and Correction of Silent Data Corruption for Large-Scale High-Performance Computing
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fiala, David J; Mueller, Frank; Engelmann, Christian
Faults have become the norm rather than the exception for high-end computing on clusters with 10s/100s of thousands of cores. Exacerbating this situation, some of these faults remain undetected, manifesting themselves as silent errors that corrupt memory while applications continue to operate and report incorrect results. This paper studies the potential for redundancy to both detect and correct soft errors in MPI message-passing applications. Our study investigates the challenges inherent to detecting soft errors within MPI application while providing transparent MPI redundancy. By assuming a model wherein corruption in application data manifests itself by producing differing MPI message data betweenmore » replicas, we study the best suited protocols for detecting and correcting MPI data that is the result of corruption. To experimentally validate our proposed detection and correction protocols, we introduce RedMPI, an MPI library which resides in the MPI profiling layer. RedMPI is capable of both online detection and correction of soft errors that occur in MPI applications without requiring any modifications to the application source by utilizing either double or triple redundancy. Our results indicate that our most efficient consistency protocol can successfully protect applications experiencing even high rates of silent data corruption with runtime overheads between 0% and 30% as compared to unprotected applications without redundancy. Using our fault injector within RedMPI, we observe that even a single soft error can have profound effects on running applications, causing a cascading pattern of corruption in most cases causes that spreads to all other processes. RedMPI's protection has been shown to successfully mitigate the effects of soft errors while allowing applications to complete with correct results even in the face of errors.« less
NASA Astrophysics Data System (ADS)
Bunzel, Felix; Müller, Wolfgang A.; Dobrynin, Mikhail; Fröhlich, Kristina; Hagemann, Stefan; Pohlmann, Holger; Stacke, Tobias; Baehr, Johanna
2018-01-01
We evaluate the impact of a new five-layer soil-hydrology scheme on seasonal hindcast skill of 2 m temperatures over Europe obtained with the Max Planck Institute Earth System Model (MPI-ESM). Assimilation experiments from 1981 to 2010 and 10-member seasonal hindcasts initialized on 1 May each year are performed with MPI-ESM in two soil configurations, one using a bucket scheme and one a new five-layer soil-hydrology scheme. We find the seasonal hindcast skill for European summer temperatures to improve with the five-layer scheme compared to the bucket scheme and investigate possible causes for these improvements. First, improved indirect soil moisture assimilation allows for enhanced soil moisture-temperature feedbacks in the hindcasts. Additionally, this leads to improved prediction of anomalies in the 500 hPa geopotential height surface, reflecting more realistic atmospheric circulation patterns over Europe.
Towards quantitative magnetic particle imaging: A comparison with magnetic particle spectroscopy
NASA Astrophysics Data System (ADS)
Paysen, Hendrik; Wells, James; Kosch, Olaf; Steinhoff, Uwe; Trahms, Lutz; Schaeffter, Tobias; Wiekhorst, Frank
2018-05-01
Magnetic Particle Imaging (MPI) is a quantitative imaging modality with promising features for several biomedical applications. Here, we study quantitatively the raw data obtained during MPI measurements. We present a method for the calibration of the MPI scanner output using measurements from a magnetic particle spectrometer (MPS) to yield data in units of magnetic moments. The calibration technique is validated in a simplified MPI mode with a 1D excitation field. Using the calibrated results from MPS and MPI, we determine and compare the detection limits for each system. The detection limits were found to be 5.10-12 Am2 for MPS and 3.6.10-10 Am2 for MPI. Finally, the quantitative information contained in a standard MPI measurement with a 3D excitation is analyzed and compared to the previous results, showing a decrease in signal amplitudes of the odd harmonics related to the case of 1D excitation. We propose physical explanations for all acquired results; and discuss the possible benefits for the improvement of MPI technology.
Performance Analysis of Scientific and Engineering Applications Using MPInside and TAU
NASA Technical Reports Server (NTRS)
Saini, Subhash; Mehrotra, Piyush; Taylor, Kenichi Jun Haeng; Shende, Sameer Suresh; Biswas, Rupak
2010-01-01
In this paper, we present performance analysis of two NASA applications using performance tools like Tuning and Analysis Utilities (TAU) and SGI MPInside. MITgcmUV and OVERFLOW are two production-quality applications used extensively by scientists and engineers at NASA. MITgcmUV is a global ocean simulation model, developed by the Estimating the Circulation and Climate of the Ocean (ECCO) Consortium, for solving the fluid equations of motion using the hydrostatic approximation. OVERFLOW is a general-purpose Navier-Stokes solver for computational fluid dynamics (CFD) problems. Using these tools, we analyze the MPI functions (MPI_Sendrecv, MPI_Bcast, MPI_Reduce, MPI_Allreduce, MPI_Barrier, etc.) with respect to message size of each rank, time consumed by each function, and how ranks communicate. MPI communication is further analyzed by studying the performance of MPI functions used in these two applications as a function of message size and number of cores. Finally, we present the compute time, communication time, and I/O time as a function of the number of cores.
Parallel algorithm of VLBI software correlator under multiprocessor environment
NASA Astrophysics Data System (ADS)
Zheng, Weimin; Zhang, Dong
2007-11-01
The correlator is the key signal processing equipment of a Very Lone Baseline Interferometry (VLBI) synthetic aperture telescope. It receives the mass data collected by the VLBI observatories and produces the visibility function of the target, which can be used to spacecraft position, baseline length measurement, synthesis imaging, and other scientific applications. VLBI data correlation is a task of data intensive and computation intensive. This paper presents the algorithms of two parallel software correlators under multiprocessor environments. A near real-time correlator for spacecraft tracking adopts the pipelining and thread-parallel technology, and runs on the SMP (Symmetric Multiple Processor) servers. Another high speed prototype correlator using the mixed Pthreads and MPI (Massage Passing Interface) parallel algorithm is realized on a small Beowulf cluster platform. Both correlators have the characteristic of flexible structure, scalability, and with 10-station data correlating abilities.
Hybrid Optimization Parallel Search PACKage
DOE Office of Scientific and Technical Information (OSTI.GOV)
2009-11-10
HOPSPACK is open source software for solving optimization problems without derivatives. Application problems may have a fully nonlinear objective function, bound constraints, and linear and nonlinear constraints. Problem variables may be continuous, integer-valued, or a mixture of both. The software provides a framework that supports any derivative-free type of solver algorithm. Through the framework, solvers request parallel function evaluation, which may use MPI (multiple machines) or multithreading (multiple processors/cores on one machine). The framework provides a Cache and Pending Cache of saved evaluations that reduces execution time and facilitates restarts. Solvers can dynamically create other algorithms to solve subproblems, amore » useful technique for handling multiple start points and integer-valued variables. HOPSPACK ships with the Generating Set Search (GSS) algorithm, developed at Sandia as part of the APPSPACK open source software project.« less
2009-08-01
event of a fire. The mesh prevents cracking to the steel substrate, which would reduce the insulating properties of the char. The procedure is as...Top Coats: MPI #9, Exterior Alkyd Enamel , Gloss, MPI Gloss Level 6 (i.e., a semi-gloss) • System 2: o Primer: MPI #23, Surface Tolerant Metal...Metal Primer X X MPI Paint #9 Exterior Alkyd Enamel , Gloss X MPI Paint #94 Exterior Alkyd
MPI-Defrost: Extension of Defrost to MPI-based Cluster Environment
NASA Astrophysics Data System (ADS)
Amin, Mustafa A.; Easther, Richard; Finkel, Hal
2011-06-01
MPI-Defrost extends Frolov’s Defrost to an MPI-based cluster environment. This version has been restricted to a single field. Restoring two-field support should be straightforward, but will require some code changes. Some output options may also not be fully supported under MPI. This code was produced to support our own work, and has been made available for the benefit of anyone interested in either oscillon simulations or an MPI capable version of Defrost, and it is provided on an "as-is" basis. Andrei Frolov is the primary developer of Defrost and we thank him for placing his work under the GPL (GNU Public License), and thus allowing us to distribute this modified version.
Optimizing R with SparkR on a commodity cluster for biomedical research.
Sedlmayr, Martin; Würfl, Tobias; Maier, Christian; Häberle, Lothar; Fasching, Peter; Prokosch, Hans-Ulrich; Christoph, Jan
2016-12-01
Medical researchers are challenged today by the enormous amount of data collected in healthcare. Analysis methods such as genome-wide association studies (GWAS) are often computationally intensive and thus require enormous resources to be performed in a reasonable amount of time. While dedicated clusters and public clouds may deliver the desired performance, their use requires upfront financial efforts or anonymous data, which is often not possible for preliminary or occasional tasks. We explored the possibilities to build a private, flexible cluster for processing scripts in R based on commodity, non-dedicated hardware of our department. For this, a GWAS-calculation in R on a single desktop computer, a Message Passing Interface (MPI)-cluster, and a SparkR-cluster were compared with regards to the performance, scalability, quality, and simplicity. The original script had a projected runtime of three years on a single desktop computer. Optimizing the script in R already yielded a significant reduction in computing time (2 weeks). By using R-MPI and SparkR, we were able to parallelize the computation and reduce the time to less than three hours (2.6 h) on already available, standard office computers. While MPI is a proven approach in high-performance clusters, it requires rather static, dedicated nodes. SparkR and its Hadoop siblings allow for a dynamic, elastic environment with automated failure handling. SparkR also scales better with the number of nodes in the cluster than MPI due to optimized data communication. R is a popular environment for clinical data analysis. The new SparkR solution offers elastic resources and allows supporting big data analysis using R even on non-dedicated resources with minimal change to the original code. To unleash the full potential, additional efforts should be invested to customize and improve the algorithms, especially with regards to data distribution. Copyright © 2016 The Authors. Published by Elsevier Ireland Ltd.. All rights reserved.
On the Suitability of MPI as a PGAS Runtime
DOE Office of Scientific and Technical Information (OSTI.GOV)
Daily, Jeffrey A.; Vishnu, Abhinav; Palmer, Bruce J.
2014-12-18
Partitioned Global Address Space (PGAS) models are emerging as a popular alternative to MPI models for designing scalable applications. At the same time, MPI remains a ubiquitous communication subsystem due to its standardization, high performance, and availability on leading platforms. In this paper, we explore the suitability of using MPI as a scalable PGAS communication subsystem. We focus on the Remote Memory Access (RMA) communication in PGAS models which typically includes {\\em get, put,} and {\\em atomic memory operations}. We perform an in-depth exploration of design alternatives based on MPI. These alternatives include using a semantically-matching interface such as MPI-RMA,more » as well as not-so-intuitive interfaces such as MPI two-sided with a combination of multi-threading and dynamic process management. With an in-depth exploration of these alternatives and their shortcomings, we propose a novel design which is facilitated by the data-centric view in PGAS models. This design leverages a combination of highly tuned MPI two-sided semantics and an automatic, user-transparent split of MPI communicators to provide asynchronous progress. We implement the asynchronous progress ranks approach and other approaches within the Communication Runtime for Exascale which is a communication subsystem for Global Arrays. Our performance evaluation spans pure communication benchmarks, graph community detection and sparse matrix-vector multiplication kernels, and a computational chemistry application. The utility of our proposed PR-based approach is demonstrated by a 2.17x speed-up on 1008 processors over the other MPI-based designs.« less
Magnetic Particle Imaging for Real-Time Perfusion Imaging in Acute Stroke.
Ludewig, Peter; Gdaniec, Nadine; Sedlacik, Jan; Forkert, Nils D; Szwargulski, Patryk; Graeser, Matthias; Adam, Gerhard; Kaul, Michael G; Krishnan, Kannan M; Ferguson, R Matthew; Khandhar, Amit P; Walczak, Piotr; Fiehler, Jens; Thomalla, Götz; Gerloff, Christian; Knopp, Tobias; Magnus, Tim
2017-10-24
The fast and accurate assessment of cerebral perfusion is fundamental for the diagnosis and successful treatment of stroke patients. Magnetic particle imaging (MPI) is a new radiation-free tomographic imaging method with a superior temporal resolution, compared to other conventional imaging methods. In addition, MPI scanners can be built as prehospital mobile devices, which require less complex infrastructure than computed tomography (CT) and magnetic resonance imaging (MRI). With these advantages, MPI could accelerate the stroke diagnosis and treatment, thereby improving outcomes. Our objective was to investigate the capabilities of MPI to detect perfusion deficits in a murine model of ischemic stroke. Cerebral ischemia was induced by inserting of a microfilament in the internal carotid artery in C57BL/6 mice, thereby blocking the blood flow into the medial cerebral artery. After the injection of a contrast agent (superparamagnetic iron oxide nanoparticles) specifically tailored for MPI, cerebral perfusion and vascular anatomy were assessed by the MPI scanner within seconds. To validate and compare our MPI data, we performed perfusion imaging with a small animal MRI scanner. MPI detected the perfusion deficits in the ischemic brain, which were comparable to those with MRI but in real-time. For the first time, we showed that MPI could be used as a diagnostic tool for relevant diseases in vivo, such as an ischemic stroke. Due to its shorter image acquisition times and increased temporal resolution compared to that of MRI or CT, we expect that MPI offers the potential to improve stroke imaging and treatment.
Dantas, Roberto Nery; Assuncao, Antonildes Nascimento; Marques, Ismar Aguiar; Fahel, Mateus Guimaraes; Nomura, Cesar Higa; Avila, Luiz Francisco Rodrigues; Giorgi, Maria Clementina Pinto; Soares, Jose; Meneghetti, Jose Claudio; Parga, Jose Rodrigues
2018-06-01
Despite advances in non-invasive myocardial perfusion imaging (MPI) evaluation, computed tomography (CT) multiphase MPI protocols have not yet been compared with the highly accurate rubidium-82 positron emission tomography ( 82 RbPET) MPI. Thus, this study aimed to evaluate agreement between 82 RbPET and 320-detector row CT (320-CT) MPI using a multiphase protocol in suspected CAD patients. Forty-four patients referred for MPI evaluation were prospectively enrolled and underwent dipyridamole stress 82 RbPET and multiphase 320-CT MPI (five consecutive volumetric acquisitions during stress). Statistical analyses were performed using the R software. There was high agreement for recognizing summed stress scores ≥ 4 (kappa 0.77, 95% CI 0.55-0.98, p < 0.001) and moderate for detecting SDS ≥ 2 (kappa 0.51, 95% CI 0.23-0.80, p < 0.001). In a per segment analysis, agreement was high for the presence of perfusion defects during stress and rest (kappa 0.75 and 0.82, respectively) and was moderate for impairment severity (kappa 0.58 and 0.65, respectively). The 320-CT protocol was safe, with low radiation burden (9.3 ± 2.4 mSv). There was a significant agreement between dipyridamole stress 320-CT MPI and 82 RbPET MPI in the evaluation of suspected CAD patients of intermediate risk. The multiphase 320-CT MPI protocol was feasible, diagnostic and with relatively low radiation exposure. • Rubidium-82 PET and 320-MDCT can perform MPI studies for CAD investigation. • There is high agreement between rubidium-82 PET and 320-MDCT for MPI assessment. • Multiphase CT perfusion protocols are feasible and with low radiation. • Multiphase CT perfusion protocols can identify image artefacts.
Salamon, Johannes; Hofmann, Martin; Jung, Caroline; Kaul, Michael Gerhard; Werner, Franziska; Them, Kolja; Reimer, Rudolph; Nielsen, Peter; Vom Scheidt, Annika; Adam, Gerhard; Knopp, Tobias; Ittrich, Harald
2016-01-01
In-vitro evaluation of the feasibility of 4D real time tracking of endovascular devices and stenosis treatment with a magnetic particle imaging (MPI) / magnetic resonance imaging (MRI) road map approach and an MPI-guided approach using a blood pool tracer. A guide wire and angioplasty-catheter were labeled with a thin layer of magnetic lacquer. For real time MPI a custom made software framework was developed. A stenotic vessel phantom filled with saline or superparamagnetic iron oxide nanoparticles (MM4) was equipped with bimodal fiducial markers for co-registration in preclinical 7T MRI and MPI. In-vitro angioplasty was performed inflating the balloon with saline or MM4. MPI data were acquired using a field of view of 37.3×37.3×18.6 mm3 and a frame rate of 46 volumes/sec. Analysis of the magnetic lacquer-marks on the devices were performed with electron microscopy, atomic absorption spectrometry and micro-computed tomography. Magnetic marks allowed for MPI/MRI guidance of interventional devices. Bimodal fiducial markers enable MPI/MRI image fusion for MRI based roadmapping. MRI roadmapping and the blood pool tracer approach facilitate MPI real time monitoring of in-vitro angioplasty. Successful angioplasty was verified with MPI and MRI. Magnetic marks consist of micrometer sized ferromagnetic plates mainly composed of iron and iron oxide. 4D real time MP imaging, tracking and guiding of endovascular instruments and in-vitro angioplasty is feasible. In addition to an approach that requires a blood pool tracer, MRI based roadmapping might emerge as a promising tool for radiation free 4D MPI-guided interventions.
NASA Astrophysics Data System (ADS)
Jung, C.; Salamon, J.; Hofmann, M.; Kaul, M. G.; Adam, G.; Ittrich, H.; Knopp, T.
2016-03-01
Purpose: The goal of this study was to achieve a real time 3D visualisation of the murine cardiovascular system by intravenously injected superparamagnetic nanoparticles using Magnetic particle imaging (MPI). Material and Methods: MPI scans of FVB mice were performed using a 3D imaging sequence (1T/m gradient strength, 10mT drive-field strength). A dynamic scan with a temporal resolution of 21.5ms per 3D volume acquisition was performed. 50μl ferucarbotran (Resovist®, Bayer Healthcare AG) were injected into the tail vein after baseline MPI measurements. As MPI delivers no anatomic information, MRI scans at a 7T ClinScan (Bruker) were performed using a T2-weighted 2D TSE sequence. The reconstruction of the MPI data was performed on the MPI console (ParaVision 6.0/MPI, Bruker). Image fusion was done using additional image processing software (Imalytics, Philips). The dynamic information was extracted using custom software developed in the Julia programming environment. Results: The combined MRI-MPI measurements were carried out successfully. MPI data clearly demonstrated the passage of the SPIO tracer through the inferior vena cava, the heart and finally the liver. By co-registration with MRI the anatomical regions were identified. Due to the volume frame rate of about 46 volumes per second a signal modulation with the frequency of the heart beat was detectable and a heart beat of 520 beats per minute (bpm) has been assumed. Moreover, the blood flow velocity of approximately 5cm/s in the vena cava has been estimated. Conclusions: The high temporal resolution of MPI allows real-time imaging and bolus tracking of intravenous injected nanoparticles and offers a real time tool to assess blood flow velocity.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oxberry, Geoffrey
Google Test MPI Listener is a plugin for the Google Test c++ unit testing library that organizes test output of software that uses both the MPI parallel programming model and Google Test. Typically, such output is ordered arbitrarily and disorganized, making difficult the process of interpreting test output. This plug organizes output in MPI rank order, enabling easy interpretation of test results.
Final report: Compiled MPI. Cost-Effective Exascale Application Development
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gropp, William Douglas
2015-12-21
This is the final report on Compiled MPI: Cost-Effective Exascale Application Development, and summarizes the results under this project. The project investigated runtime enviroments that improve the performance of MPI (Message-Passing Interface) programs; work at Illinois in the last period of this project looked at optimizing data access optimizations expressed with MPI datatypes.
SBML-PET-MPI: a parallel parameter estimation tool for Systems Biology Markup Language based models.
Zi, Zhike
2011-04-01
Parameter estimation is crucial for the modeling and dynamic analysis of biological systems. However, implementing parameter estimation is time consuming and computationally demanding. Here, we introduced a parallel parameter estimation tool for Systems Biology Markup Language (SBML)-based models (SBML-PET-MPI). SBML-PET-MPI allows the user to perform parameter estimation and parameter uncertainty analysis by collectively fitting multiple experimental datasets. The tool is developed and parallelized using the message passing interface (MPI) protocol, which provides good scalability with the number of processors. SBML-PET-MPI is freely available for non-commercial use at http://www.bioss.uni-freiburg.de/cms/sbml-pet-mpi.html or http://sites.google.com/site/sbmlpetmpi/.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Carrington, David Bradley; Waters, Jiajia
KIVA-hpFE is a high performance computer software for solving the physics of multi-species and multiphase turbulent reactive flow in complex geometries having immersed moving parts. The code is written in Fortran 90/95 and can be used on any computer platform with any popular complier. The code is in two versions, a serial version and a parallel version utilizing MPICH2 type Message Passing Interface (MPI or Intel MPI) for solving distributed domains. The parallel version is at least 30x faster than the serial version and much faster than our previous generation of parallel engine modeling software, by many factors. The 5thmore » generation algorithm construction is a Galerkin type Finite Element Method (FEM) solving conservative momentum, species, and energy transport equations along with two-equation turbulent model k-ω Reynolds Averaged Navier-Stokes (RANS) model and a Vreman type dynamic Large Eddy Simulation (LES) method. The LES method is capable modeling transitional flow from laminar to fully turbulent; therefore, this LES method does not require special hybrid or blending to walls. The FEM projection method also uses a Petrov-Galerkin (P-G) stabilization along with pressure stabilization. We employ hierarchical basis sets, constructed on the fly with enrichment in areas associated with relatively larger error as determined by error estimation methods. In addition, when not using the hp-adaptive module, the code employs Lagrangian basis or shape functions. The shape functions are constructed for hexahedral, prismatic and tetrahedral elements. The software is designed to solve many types of reactive flow problems, from burners to internal combustion engines and turbines. In addition, the formulation allows for direct integration of solid bodies (conjugate heat transfer), as in heat transfer through housings, parts, cylinders. It can also easily be extended to stress modeling of solids, used in fluid structure interactions problems, solidification, porous media modeling and magneto hydrodynamics.« less
Parallelization of PANDA discrete ordinates code using spatial decomposition
DOE Office of Scientific and Technical Information (OSTI.GOV)
Humbert, P.
2006-07-01
We present the parallel method, based on spatial domain decomposition, implemented in the 2D and 3D versions of the discrete Ordinates code PANDA. The spatial mesh is orthogonal and the spatial domain decomposition is Cartesian. For 3D problems a 3D Cartesian domain topology is created and the parallel method is based on a domain diagonal plane ordered sweep algorithm. The parallel efficiency of the method is improved by directions and octants pipelining. The implementation of the algorithm is straightforward using MPI blocking point to point communications. The efficiency of the method is illustrated by an application to the 3D-Ext C5G7more » benchmark of the OECD/NEA. (authors)« less
NASA Astrophysics Data System (ADS)
Boyko, Oleksiy; Zheleznyak, Mark
2015-04-01
The original numerical code TOPKAPI-IMMS of the distributed rainfall-runoff model TOPKAPI ( Todini et al, 1996-2014) is developed and implemented in Ukraine. The parallel version of the code has been developed recently to be used on multiprocessors systems - multicore/processors PC and clusters. Algorithm is based on binary-tree decomposition of the watershed for the balancing of the amount of computation for all processors/cores. Message passing interface (MPI) protocol is used as a parallel computing framework. The numerical efficiency of the parallelization algorithms is demonstrated for the case studies for the flood predictions of the mountain watersheds of the Ukrainian Carpathian regions. The modeling results is compared with the predictions based on the lumped parameters models.
Shared Memory Parallelization of an Implicit ADI-type CFD Code
NASA Technical Reports Server (NTRS)
Hauser, Th.; Huang, P. G.
1999-01-01
A parallelization study designed for ADI-type algorithms is presented using the OpenMP specification for shared-memory multiprocessor programming. Details of optimizations specifically addressed to cache-based computer architectures are described and performance measurements for the single and multiprocessor implementation are summarized. The paper demonstrates that optimization of memory access on a cache-based computer architecture controls the performance of the computational algorithm. A hybrid MPI/OpenMP approach is proposed for clusters of shared memory machines to further enhance the parallel performance. The method is applied to develop a new LES/DNS code, named LESTool. A preliminary DNS calculation of a fully developed channel flow at a Reynolds number of 180, Re(sub tau) = 180, has shown good agreement with existing data.
Accelerating atomistic calculations of quantum energy eigenstates on graphic cards
NASA Astrophysics Data System (ADS)
Rodrigues, Walter; Pecchia, A.; Lopez, M.; Auf der Maur, M.; Di Carlo, A.
2014-10-01
Electronic properties of nanoscale materials require the calculation of eigenvalues and eigenvectors of large matrices. This bottleneck can be overcome by parallel computing techniques or the introduction of faster algorithms. In this paper we report a custom implementation of the Lanczos algorithm with simple restart, optimized for graphical processing units (GPUs). The whole algorithm has been developed using CUDA and runs entirely on the GPU, with a specialized implementation that spares memory and reduces at most machine-to-device data transfers. Furthermore parallel distribution over several GPUs has been attained using the standard message passing interface (MPI). Benchmark calculations performed on a GaN/AlGaN wurtzite quantum dot with up to 600,000 atoms are presented. The empirical tight-binding (ETB) model with an sp3d5s∗+spin-orbit parametrization has been used to build the system Hamiltonian (H).
Exploiting Multiple Levels of Parallelism in Sparse Matrix-Matrix Multiplication
Azad, Ariful; Ballard, Grey; Buluc, Aydin; ...
2016-11-08
Sparse matrix-matrix multiplication (or SpGEMM) is a key primitive for many high-performance graph algorithms as well as for some linear solvers, such as algebraic multigrid. The scaling of existing parallel implementations of SpGEMM is heavily bound by communication. Even though 3D (or 2.5D) algorithms have been proposed and theoretically analyzed in the flat MPI model on Erdös-Rényi matrices, those algorithms had not been implemented in practice and their complexities had not been analyzed for the general case. In this work, we present the first implementation of the 3D SpGEMM formulation that exploits multiple (intranode and internode) levels of parallelism, achievingmore » significant speedups over the state-of-the-art publicly available codes at all levels of concurrencies. We extensively evaluate our implementation and identify bottlenecks that should be subject to further research.« less
Enhancing PC Cluster-Based Parallel Branch-and-Bound Algorithms for the Graph Coloring Problem
NASA Astrophysics Data System (ADS)
Taoka, Satoshi; Takafuji, Daisuke; Watanabe, Toshimasa
A branch-and-bound algorithm (BB for short) is the most general technique to deal with various combinatorial optimization problems. Even if it is used, computation time is likely to increase exponentially. So we consider its parallelization to reduce it. It has been reported that the computation time of a parallel BB heavily depends upon node-variable selection strategies. And, in case of a parallel BB, it is also necessary to prevent increase in communication time. So, it is important to pay attention to how many and what kind of nodes are to be transferred (called sending-node selection strategy). In this paper, for the graph coloring problem, we propose some sending-node selection strategies for a parallel BB algorithm by adopting MPI for parallelization and experimentally evaluate how these strategies affect computation time of a parallel BB on a PC cluster network.
Li, J; Guo, L-X; Zeng, H; Han, X-B
2009-06-01
A message-passing-interface (MPI)-based parallel finite-difference time-domain (FDTD) algorithm for the electromagnetic scattering from a 1-D randomly rough sea surface is presented. The uniaxial perfectly matched layer (UPML) medium is adopted for truncation of FDTD lattices, in which the finite-difference equations can be used for the total computation domain by properly choosing the uniaxial parameters. This makes the parallel FDTD algorithm easier to implement. The parallel performance with different processors is illustrated for one sea surface realization, and the computation time of the parallel FDTD algorithm is dramatically reduced compared to a single-process implementation. Finally, some numerical results are shown, including the backscattering characteristics of sea surface for different polarization and the bistatic scattering from a sea surface with large incident angle and large wind speed.
Augmentation in the treatment of restless legs syndrome with transdermal rotigotine.
Beneš, Heike; García-Borreguero, Diego; Ferini-Strambi, Luigi; Schollmayer, Erwin; Fichtner, Andreas; Kohnen, Ralf
2012-06-01
To assess the risk of augmentation under treatment with the transdermally delivered dopamine agonist rotigotine for restless legs syndrome (RLS). Experts in RLS augmentation retrospectively reviewed data from two double-blind, placebo-controlled 6-month trials (745 rotigotine and 214 placebo subjects, NCT00136045 and NCT00135993) and from two open-label 1-year trials (620 rotigotine subjects, NCT00498108 and NCT00263068). All study visits were systematically evaluated applying the Max Planck Institute (MPI) criteria for the diagnosis of both augmentation and clinically relevant augmentation. MPI criteria for augmentation were met on at least one visit by 8.2% of all subjects in the double-blind trials with 12 subjects meeting the criteria for clinically relevant augmentation: 11 under rotigotine (1.5%) and one under placebo treatment. In the open-label trials, 9.7% of all subjects met the MPI criteria for augmentation and 2.9% met the criteria for clinically relevant augmentation. None of the patients treated with rotigotine for up to 1.5 years (double-blind plus open-label trial) discontinued prematurely owing to augmentation. Neither could dose-dependency or a time pattern for clinically relevant augmentation episodes be detected. Our analyses suggest that the risk for clinically relevant augmentation for the duration of up to 18 months of rotigotine treatment is low. Copyright © 2012 Elsevier B.V. All rights reserved.
Relaxation in x-space magnetic particle imaging.
Croft, Laura R; Goodwill, Patrick W; Conolly, Steven M
2012-12-01
Magnetic particle imaging (MPI) is a new imaging modality that noninvasively images the spatial distribution of superparamagnetic iron oxide nanoparticles (SPIOs). MPI has demonstrated high contrast and zero attenuation with depth, and MPI promises superior safety compared to current angiography methods, X-ray, computed tomography, and magnetic resonance imaging angiography. Nanoparticle relaxation can delay the SPIO magnetization, and in this work we investigate the open problem of the role relaxation plays in MPI scanning and its effect on the image. We begin by amending the x-space theory of MPI to include nanoparticle relaxation effects. We then validate the amended theory with experiments from a Berkeley x-space relaxometer and a Berkeley x-space projection MPI scanner. Our theory and experimental data indicate that relaxation reduces SNR and asymmetrically blurs the image in the scanning direction. While relaxation effects can have deleterious effects on the MPI scan, we show theoretically and experimentally that x-space reconstruction remains robust in the presence of relaxation. Furthermore, the role of relaxation in x-space theory provides guidance as we develop methods to minimize relaxation-induced blurring. This will be an important future area of research for the MPI community.
Javadi, Hamid; Jallalat, Sara; Semnani, Shahriar; Mogharrabi, Mehdi; Nabipour, Iraj; Abbaszadeh, Moloud; Assadi, Majid
2013-01-01
False-positive findings with myocardial perfusion imaging (MPI) have frequently been identified in the presence of left bundle branch block (LBBB) and tend to lower the accuracy of MPI in individuals with normal coronary angiographs. Pharmacologic stress is recognized as the preferred method for MPI in patients with LBBB. In contrast, very few studies have evaluated the effect of right bundle branch block (RBBB) on MPI, and there is no consensus regarding the selection of pharmacologic versus exercise stress during MPI for the RBBB patient. In this study, we present a 45-year-old man with RBBB, who has a normal coronary artery angiography, but who showed abnormal myocardial perfusion with exercise MPI, and normal perfusion on dipyridamole MPI. The aim of the study is to stimulate awareness that the stress method selected for patients with RBBB can potentially interfere with the accuracy of the data.
Collective Framework and Performance Optimizations to Open MPI for Cray XT Platforms
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ladd, Joshua S; Gorentla Venkata, Manjunath; Shamis, Pavel
2011-01-01
The performance and scalability of collective operations plays a key role in the performance and scalability of many scientific applications. Within the Open MPI code base we have developed a general purpose hierarchical collective operations framework called Cheetah, and applied it at large scale on the Oak Ridge Leadership Computing Facility's Jaguar (OLCF) platform, obtaining better performance and scalability than the native MPI implementation. This paper discuss Cheetah's design and implementation, and optimizations to the framework for Cray XT 5 platforms. Our results show that the Cheetah's Broadcast and Barrier perform better than the native MPI implementation. For medium data,more » the Cheetah's Broadcast outperforms the native MPI implementation by 93% for 49,152 processes problem size. For small and large data, it out performs the native MPI implementation by 10% and 9%, respectively, at 24,576 processes problem size. The Cheetah's Barrier performs 10% better than the native MPI implementation for 12,288 processes problem size.« less
NDL-v2.0: A new version of the numerical differentiation library for parallel architectures
NASA Astrophysics Data System (ADS)
Hadjidoukas, P. E.; Angelikopoulos, P.; Voglis, C.; Papageorgiou, D. G.; Lagaris, I. E.
2014-07-01
We present a new version of the numerical differentiation library (NDL) used for the numerical estimation of first and second order partial derivatives of a function by finite differencing. In this version we have restructured the serial implementation of the code so as to achieve optimal task-based parallelization. The pure shared-memory parallelization of the library has been based on the lightweight OpenMP tasking model allowing for the full extraction of the available parallelism and efficient scheduling of multiple concurrent library calls. On multicore clusters, parallelism is exploited by means of TORC, an MPI-based multi-threaded tasking library. The new MPI implementation of NDL provides optimal performance in terms of function calls and, furthermore, supports asynchronous execution of multiple library calls within legacy MPI programs. In addition, a Python interface has been implemented for all cases, exporting the functionality of our library to sequential Python codes. Catalog identifier: AEDG_v2_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEDG_v2_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 63036 No. of bytes in distributed program, including test data, etc.: 801872 Distribution format: tar.gz Programming language: ANSI Fortran-77, ANSI C, Python. Computer: Distributed systems (clusters), shared memory systems. Operating system: Linux, Unix. Has the code been vectorized or parallelized?: Yes. RAM: The library uses O(N) internal storage, N being the dimension of the problem. It can use up to O(N2) internal storage for Hessian calculations, if a task throttling factor has not been set by the user. Classification: 4.9, 4.14, 6.5. Catalog identifier of previous version: AEDG_v1_0 Journal reference of previous version: Comput. Phys. Comm. 180(2009)1404 Does the new version supersede the previous version?: Yes Nature of problem: The numerical estimation of derivatives at several accuracy levels is a common requirement in many computational tasks, such as optimization, solution of nonlinear systems, and sensitivity analysis. For a large number of scientific and engineering applications, the underlying functions correspond to simulation codes for which analytical estimation of derivatives is difficult or almost impossible. A parallel implementation that exploits systems with multiple CPUs is very important for large scale and computationally expensive problems. Solution method: Finite differencing is used with a carefully chosen step that minimizes the sum of the truncation and round-off errors. The parallel versions employ both OpenMP and MPI libraries. Reasons for new version: The updated version was motivated by our endeavors to extend a parallel Bayesian uncertainty quantification framework [1], by incorporating higher order derivative information as in most state-of-the-art stochastic simulation methods such as Stochastic Newton MCMC [2] and Riemannian Manifold Hamiltonian MC [3]. The function evaluations are simulations with significant time-to-solution, which also varies with the input parameters such as in [1, 4]. The runtime of the N-body-type of problem changes considerably with the introduction of a longer cut-off between the bodies. In the first version of the library, the OpenMP-parallel subroutines spawn a new team of threads and distribute the function evaluations with a PARALLEL DO directive. This limits the functionality of the library as multiple concurrent calls require nested parallelism support from the OpenMP environment. Therefore, either their function evaluations will be serialized or processor oversubscription is likely to occur due to the increased number of OpenMP threads. In addition, the Hessian calculations include two explicit parallel regions that compute first the diagonal and then the off-diagonal elements of the array. Due to the barrier between the two regions, the parallelism of the calculations is not fully exploited. These issues have been addressed in the new version by first restructuring the serial code and then running the function evaluations in parallel using OpenMP tasks. Although the MPI-parallel implementation of the first version is capable of fully exploiting the task parallelism of the PNDL routines, it does not utilize the caching mechanism of the serial code and, therefore, performs some redundant function evaluations in the Hessian and Jacobian calculations. This can lead to: (a) higher execution times if the number of available processors is lower than the total number of tasks, and (b) significant energy consumption due to wasted processor cycles. Overcoming these drawbacks, which become critical as the time of a single function evaluation increases, was the primary goal of this new version. Due to the code restructure, the MPI-parallel implementation (and the OpenMP-parallel in accordance) avoids redundant calls, providing optimal performance in terms of the number of function evaluations. Another limitation of the library was that the library subroutines were collective and synchronous calls. In the new version, each MPI process can issue any number of subroutines for asynchronous execution. We introduce two library calls that provide global and local task synchronizations, similarly to the BARRIER and TASKWAIT directives of OpenMP. The new MPI-implementation is based on TORC, a new tasking library for multicore clusters [5-7]. TORC improves the portability of the software, as it relies exclusively on the POSIX-Threads and MPI programming interfaces. It allows MPI processes to utilize multiple worker threads, offering a hybrid programming and execution environment similar to MPI+OpenMP, in a completely transparent way. Finally, to further improve the usability of our software, a Python interface has been implemented on top of both the OpenMP and MPI versions of the library. This allows sequential Python codes to exploit shared and distributed memory systems. Summary of revisions: The revised code improves the performance of both parallel (OpenMP and MPI) implementations. The functionality and the user-interface of the MPI-parallel version have been extended to support the asynchronous execution of multiple PNDL calls, issued by one or multiple MPI processes. A new underlying tasking library increases portability and allows MPI processes to have multiple worker threads. For both implementations, an interface to the Python programming language has been added. Restrictions: The library uses only double precision arithmetic. The MPI implementation assumes the homogeneity of the execution environment provided by the operating system. Specifically, the processes of a single MPI application must have identical address space and a user function resides at the same virtual address. In addition, address space layout randomization should not be used for the application. Unusual features: The software takes into account bound constraints, in the sense that only feasible points are used to evaluate the derivatives, and given the level of the desired accuracy, the proper formula is automatically employed. Running time: Running time depends on the function's complexity. The test run took 23 ms for the serial distribution, 25 ms for the OpenMP with 2 threads, 53 ms and 1.01 s for the MPI parallel distribution using 2 threads and 2 processes respectively and yield-time for idle workers equal to 10 ms. References: [1] P. Angelikopoulos, C. Paradimitriou, P. Koumoutsakos, Bayesian uncertainty quantification and propagation in molecular dynamics simulations: a high performance computing framework, J. Chem. Phys 137 (14). [2] H.P. Flath, L.C. Wilcox, V. Akcelik, J. Hill, B. van Bloemen Waanders, O. Ghattas, Fast algorithms for Bayesian uncertainty quantification in large-scale linear inverse problems based on low-rank partial Hessian approximations, SIAM J. Sci. Comput. 33 (1) (2011) 407-432. [3] M. Girolami, B. Calderhead, Riemann manifold Langevin and Hamiltonian Monte Carlo methods, J. R. Stat. Soc. Ser. B (Stat. Methodol.) 73 (2) (2011) 123-214. [4] P. Angelikopoulos, C. Paradimitriou, P. Koumoutsakos, Data driven, predictive molecular dynamics for nanoscale flow simulations under uncertainty, J. Phys. Chem. B 117 (47) (2013) 14808-14816. [5] P.E. Hadjidoukas, E. Lappas, V.V. Dimakopoulos, A runtime library for platform-independent task parallelism, in: PDP, IEEE, 2012, pp. 229-236. [6] C. Voglis, P.E. Hadjidoukas, D.G. Papageorgiou, I. Lagaris, A parallel hybrid optimization algorithm for fitting interatomic potentials, Appl. Soft Comput. 13 (12) (2013) 4481-4492. [7] P.E. Hadjidoukas, C. Voglis, V.V. Dimakopoulos, I. Lagaris, D.G. Papageorgiou, Supporting adaptive and irregular parallelism for non-linear numerical optimization, Appl. Math. Comput. 231 (2014) 544-559.
MPI_XSTAR: MPI-based parallelization of XSTAR program
NASA Astrophysics Data System (ADS)
Danehkar, A.
2017-12-01
MPI_XSTAR parallelizes execution of multiple XSTAR runs using Message Passing Interface (MPI). XSTAR (ascl:9910.008), part of the HEASARC's HEAsoft (ascl:1408.004) package, calculates the physical conditions and emission spectra of ionized gases. MPI_XSTAR invokes XSTINITABLE from HEASoft to generate a job list of XSTAR commands for given physical parameters. The job list is used to make directories in ascending order, where each individual XSTAR is spawned on each processor and outputs are saved. HEASoft's XSTAR2TABLE program is invoked upon the contents of each directory in order to produce table model FITS files for spectroscopy analysis tools.
Katouda, Michio; Naruse, Akira; Hirano, Yukihiko; Nakajima, Takahito
2016-11-15
A new parallel algorithm and its implementation for the RI-MP2 energy calculation utilizing peta-flop-class many-core supercomputers are presented. Some improvements from the previous algorithm (J. Chem. Theory Comput. 2013, 9, 5373) have been performed: (1) a dual-level hierarchical parallelization scheme that enables the use of more than 10,000 Message Passing Interface (MPI) processes and (2) a new data communication scheme that reduces network communication overhead. A multi-node and multi-GPU implementation of the present algorithm is presented for calculations on a central processing unit (CPU)/graphics processing unit (GPU) hybrid supercomputer. Benchmark results of the new algorithm and its implementation using the K computer (CPU clustering system) and TSUBAME 2.5 (CPU/GPU hybrid system) demonstrate high efficiency. The peak performance of 3.1 PFLOPS is attained using 80,199 nodes of the K computer. The peak performance of the multi-node and multi-GPU implementation is 514 TFLOPS using 1349 nodes and 4047 GPUs of TSUBAME 2.5. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Sloan, Gregory James
The direct numerical simulation (DNS) offers the most accurate approach to modeling the behavior of a physical system, but carries an enormous computation cost. There exists a need for an accurate DNS to model the coupled solid-fluid system seen in targeted drug delivery (TDD), nanofluid thermal energy storage (TES), as well as other fields where experiments are necessary, but experiment design may be costly. A parallel DNS can greatly reduce the large computation times required, while providing the same results and functionality of the serial counterpart. A D2Q9 lattice Boltzmann method approach was implemented to solve the fluid phase. The use of domain decomposition with message passing interface (MPI) parallelism resulted in an algorithm that exhibits super-linear scaling in testing, which may be attributed to the caching effect. Decreased performance on a per-node basis for a fixed number of processes confirms this observation. A multiscale approach was implemented to model the behavior of nanoparticles submerged in a viscous fluid, and used to examine the mechanisms that promote or inhibit clustering. Parallelization of this model using a masterworker algorithm with MPI gives less-than-linear speedup for a fixed number of particles and varying number of processes. This is due to the inherent inefficiency of the master-worker approach. Lastly, these separate simulations are combined, and two-way coupling is implemented between the solid and fluid.
Solar wind interaction with Venus and Mars in a parallel hybrid code
NASA Astrophysics Data System (ADS)
Jarvinen, Riku; Sandroos, Arto
2013-04-01
We discuss the development and applications of a new parallel hybrid simulation, where ions are treated as particles and electrons as a charge-neutralizing fluid, for the interaction between the solar wind and Venus and Mars. The new simulation code under construction is based on the algorithm of the sequential global planetary hybrid model developed at the Finnish Meteorological Institute (FMI) and on the Corsair parallel simulation platform also developed at the FMI. The FMI's sequential hybrid model has been used for studies of plasma interactions of several unmagnetized and weakly magnetized celestial bodies for more than a decade. Especially, the model has been used to interpret in situ particle and magnetic field observations from plasma environments of Mars, Venus and Titan. Further, Corsair is an open source MPI (Message Passing Interface) particle and mesh simulation platform, mainly aimed for simulations of diffusive shock acceleration in solar corona and interplanetary space, but which is now also being extended for global planetary hybrid simulations. In this presentation we discuss challenges and strategies of parallelizing a legacy simulation code as well as possible applications and prospects of a scalable parallel hybrid model for the solar wind interactions of Venus and Mars.
Underlying-event sensitive observables in Drell–Yan production using GENEVA
Alioli, Simone; Bauer, Christian W.; Guns, Sam; ...
2016-11-09
We present an extension of the Geneva Monte Carlo framework to include multiple parton interactions (MPI) provided by Pythia8. This allows us to obtain predictions for underlying-event sensitive measurements in Drell–Yan production, in conjunction with Geneva ’s fully differential NNLO calculation, NNLL' resummation for the 0-jet resolution variable (beam thrust), and NLL resummation for the 1-jet resolution variable. We describe the interface with the parton-shower algorithm and MPI model of Pythia8, which preserves both the precision of the partonic N-jet cross sections in Geneva as well as the shower accuracy and good description of soft hadronic physics of Pythia8. Wemore » present results for several underlying-event sensitive observables and compare to data from ATLAS and CMS as well as to standalone Pythia8 predictions. This includes a comparison with the recent ATLAS measurement of the beam thrust spectrum, which provides a potential avenue to fully disentangle the physical effects from the primary hard interaction, primary soft radiation, multiple parton interactions, and nonperturbative hadronization.« less
High Resolution Aerospace Applications using the NASA Columbia Supercomputer
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.; Aftosmis, Michael J.; Berger, Marsha
2005-01-01
This paper focuses on the parallel performance of two high-performance aerodynamic simulation packages on the newly installed NASA Columbia supercomputer. These packages include both a high-fidelity, unstructured, Reynolds-averaged Navier-Stokes solver, and a fully-automated inviscid flow package for cut-cell Cartesian grids. The complementary combination of these two simulation codes enables high-fidelity characterization of aerospace vehicle design performance over the entire flight envelope through extensive parametric analysis and detailed simulation of critical regions of the flight envelope. Both packages. are industrial-level codes designed for complex geometry and incorpor.ats. CuStomized multigrid solution algorithms. The performance of these codes on Columbia is examined using both MPI and OpenMP and using both the NUMAlink and InfiniBand interconnect fabrics. Numerical results demonstrate good scalability on up to 2016 CPUs using the NUMAIink4 interconnect, with measured computational rates in the vicinity of 3 TFLOP/s, while InfiniBand showed some performance degradation at high CPU counts, particularly with multigrid. Nonetheless, the results are encouraging enough to indicate that larger test cases using combined MPI/OpenMP communication should scale well on even more processors.
Underlying-event sensitive observables in Drell–Yan production using GENEVA
DOE Office of Scientific and Technical Information (OSTI.GOV)
Alioli, Simone; Bauer, Christian W.; Guns, Sam
We present an extension of the Geneva Monte Carlo framework to include multiple parton interactions (MPI) provided by Pythia8. This allows us to obtain predictions for underlying-event sensitive measurements in Drell–Yan production, in conjunction with Geneva ’s fully differential NNLO calculation, NNLL' resummation for the 0-jet resolution variable (beam thrust), and NLL resummation for the 1-jet resolution variable. We describe the interface with the parton-shower algorithm and MPI model of Pythia8, which preserves both the precision of the partonic N-jet cross sections in Geneva as well as the shower accuracy and good description of soft hadronic physics of Pythia8. Wemore » present results for several underlying-event sensitive observables and compare to data from ATLAS and CMS as well as to standalone Pythia8 predictions. This includes a comparison with the recent ATLAS measurement of the beam thrust spectrum, which provides a potential avenue to fully disentangle the physical effects from the primary hard interaction, primary soft radiation, multiple parton interactions, and nonperturbative hadronization.« less
Mathematical analysis of the 1D model and reconstruction schemes for magnetic particle imaging
NASA Astrophysics Data System (ADS)
Erb, W.; Weinmann, A.; Ahlborg, M.; Brandt, C.; Bringout, G.; Buzug, T. M.; Frikel, J.; Kaethner, C.; Knopp, T.; März, T.; Möddel, M.; Storath, M.; Weber, A.
2018-05-01
Magnetic particle imaging (MPI) is a promising new in vivo medical imaging modality in which distributions of super-paramagnetic nanoparticles are tracked based on their response in an applied magnetic field. In this paper we provide a mathematical analysis of the modeled MPI operator in the univariate situation. We provide a Hilbert space setup, in which the MPI operator is decomposed into simple building blocks and in which these building blocks are analyzed with respect to their mathematical properties. In turn, we obtain an analysis of the MPI forward operator and, in particular, of its ill-posedness properties. We further get that the singular values of the MPI core operator decrease exponentially. We complement our analytic results by some numerical studies which, in particular, suggest a rapid decay of the singular values of the MPI operator.
Parallelization of Rocket Engine Simulator Software (PRESS)
NASA Technical Reports Server (NTRS)
Cezzar, Ruknet
1998-01-01
We have outlined our work in the last half of the funding period. We have shown how a demo package for RESSAP using MPI can be done. However, we also mentioned the difficulties with the UNIX platform. We have reiterated some of the suggestions made during the presentation of the progress of the at Fourth Annual HBCU Conference. Although we have discussed, in some detail, how TURBDES/PUMPDES software can be run in parallel using MPI, at present, we are unable to experiment any further with either MPI or PVM. Due to X windows not being implemented, we are also not able to experiment further with XPVM, which it will be recalled, has a nice GUI interface. There are also some concerns, on our part, about MPI being an appropriate tool. The best thing about MPr is that it is public domain. Although and plenty of documentation exists for the intricacies of using MPI, little information is available on its actual implementations. Other than very typical, somewhat contrived examples, such as Jacobi algorithm for solving Laplace's equation, there are few examples which can readily be applied to real situations, such as in our case. In effect, the review of literature on both MPI and PVM, and there is a lot, indicate something similar to the enormous effort which was spent on LISP and LISP-like languages as tools for artificial intelligence research. During the development of a book on programming languages [12], when we searched the literature for very simple examples like taking averages, reading and writing records, multiplying matrices, etc., we could hardly find a any! Yet, so much was said and done on that topic in academic circles. It appears that we faced the same problem with MPI, where despite significant documentation, we could not find even a simple example which supports course-grain parallelism involving only a few processes. From the foregoing, it appears that a new direction may be required for more productive research during the extension period (10/19/98 - 10/18/99). At the least, the research would need to be done on Windows 95/Windows NT based platforms. Moreover, with the acquisition of Lahey Fortran package for PC platform, and the existing Borland C + + 5. 0, we can do work on C + + wrapper issues. We have carefully studied the blueprint for Space Transportation Propulsion Integrated Design Environment for the next 25 years [13] and found the inclusion of HBCUs in that effort encouraging. Especially in the long period for which a map is provided, there is no doubt that HBCUs will grow and become better equipped to do meaningful research. In the shorter period, as was suggested in our presentation at the HBCU conference, some key decisions regarding the aging Fortran based software for rocket propellants will need to be made. One important issue is whether or not object oriented languages such as C + + or Java should be used for distributed computing. Whether or not "distributed computing" is necessary for the existing software is yet another, larger, question to be tackled with.
Temperature dependence in magnetic particle imaging
NASA Astrophysics Data System (ADS)
Wells, James; Paysen, Hendrik; Kosch, Olaf; Trahms, Lutz; Wiekhorst, Frank
2018-05-01
Experimental results are presented demonstrating how temperature can influence the dynamics of magnetic nanoparticles (MNPs) in liquid suspension, when exposed to alternating magnetic fields in the kilohertz frequency range. The measurements used to probe the nanoparticle systems are directly linked to both the emerging biomedical technique of magnetic particle imaging (MPI), and to the recently proposed concept of remote nanoscale thermometry using MNPs under AC field excitation. Here, we report measurements on three common types of MNPs, two of which are currently leading candidates for use as tracers in MPI. Using highly-sensitive magnetic particle spectroscopy (MPS), we demonstrate significant and divergent thermal dependences in several key measures used in the evaluation of MNP dynamics for use in MPI and other applications. The temperature range studied was between 296 and 318 Kelvin, making our findings of particular importance for MPI and other biomedical technologies. Furthermore, we report the detection of the same temperature dependences in measurements conducted using the detection coils within an operational preclinical MPI scanner. This clearly shows the importance of considering temperature during MPI development, and the potential for temperature-resolved MPI using this system. We propose possible physical explanations for the differences in the behaviors observed between the different particle types, and discuss our results in terms of the opportunities and concerns they raise for MPI and other MNP based technologies.
NASA Technical Reports Server (NTRS)
Lawson, Gary; Poteat, Michael; Sosonkina, Masha; Baurle, Robert; Hammond, Dana
2016-01-01
In this work, several mini-apps have been created to enhance a real-world application performance, namely the VULCAN code for complex flow analysis developed at the NASA Langley Research Center. These mini-apps explore hybrid parallel programming paradigms with Message Passing Interface (MPI) for distributed memory access and either Shared MPI (SMPI) or OpenMP for shared memory accesses. Performance testing shows that MPI+SMPI yields the best execution performance, while requiring the largest number of code changes. A maximum speedup of 23X was measured for MPI+SMPI, but only 10X was measured for MPI+OpenMP.
Effect of central hypothyroidism on Doppler-derived myocardial performance index.
Doin, Fabio Luiz Casanova; Borges, Mariana da Rosa; Campos, Orlando; de Camargo Carvalho, Antonio Carlos; de Paola, Angelo Amato Vincenzo; Paiva, Marcelo Goulart; Abucham, Julio; Moises, Valdir Ambrosio
2004-06-01
Myocardial performance index (MPI) has been used to assess global ventricular function in different types of cardiac disease. Thyroid hormones influence cardiac performance directly and indirectly by changes in peripheral circulation. The aim of this study was to evaluate the possible effect of central hypothyroidism (CH) on MPI. The study included 28 control subjects and 7 patients with CH without cardiac disease. MPI was defined as the sum of isovolumetric contraction time (ICT) and isovolumetric relaxation time divided by ejection time. Patients were submitted to hormonal therapy with thyroxin and the study was repeated after 35 to 42 days. MPI was significantly higher in patients with CH (0.54 +/- 0.08) than in control subjects (0.40 +/- 0.05) (P =.002). The increase in MPI was caused by the prolongation of ICT without a significant variation of isovolumetric relaxation time and ejection time. After hormonal therapy there was a significant reduction of MPI (0.54 +/- 0.08 vs 0.42 +/- 0.07; P =.028) and ICT. MPI was increased in patients with untreated CH. The increase was related to prolongation of ICT and reverted by hormonal therapy.
MPI Enhancements in John the Ripper
NASA Astrophysics Data System (ADS)
Sykes, Edward R.; Lin, Michael; Skoczen, Wesley
2010-11-01
John the Ripper (JtR) is an open source software package commonly used by system administrators to enforce password policy. JtR is designed to attack (i.e., crack) passwords encrypted in a wide variety of commonly used formats. While parallel implementations of JtR exist, there are several limitations to them. This research reports on two distinct algorithms that enhance this password cracking tool using the Message Passing Interface. The first algorithm is a novel approach that uses numerous processors to crack one password by using an innovative approach to workload distribution. In this algorithm the candidate password is distributed to all participating processors and the word list is divided based on probability so that each processor has the same likelihood of cracking the password while eliminating overlapping operations. The second algorithm developed in this research involves dividing the passwords within a password file equally amongst available processors while ensuring load-balanced and fault-tolerant behavior. This paper describes John the Ripper, the design of these two algorithms and preliminary results. Given the same amount of time, the original JtR can crack 29 passwords, whereas our algorithms 1 and 2 can crack an additional 35 and 45 passwords respectively.
Current and Future Decadal Trends in the Oceanic Carbon Uptake Are Dominated by Internal Variability
NASA Astrophysics Data System (ADS)
Li, Hongmei; Ilyina, Tatiana
2018-01-01
We investigate the internal decadal variability of the ocean carbon uptake using 100 ensemble simulations based on the Max Planck Institute Earth system model (MPI-ESM). We find that on decadal time scales, internal variability (ensemble spread) is as large as the forced temporal variability (ensemble mean), and the largest internal variability is found in major carbon sink regions, that is, the 50-65°S band of the Southern Ocean, the North Pacific, and the North Atlantic. The MPI-ESM ensemble produces both positive and negative 10 year trends in the ocean carbon uptake in agreement with observational estimates. Negative decadal trends are projected to occur in the future under RCP4.5 scenario. Due to the large internal variability, the Southern Ocean and the North Pacific require the most ensemble members (more than 53 and 46, respectively) to reproduce the forced decadal trends. This number increases up to 79 in future decades as CO2 emission trajectory changes.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sayan Ghosh, Jeff Hammond
OpenSHMEM is a community effort to unifyt and standardize the SHMEM programming model. MPI (Message Passing Interface) is a well-known community standard for parallel programming using distributed memory. The most recen t release of MPI, version 3.0, was designed in part to support programming models like SHMEM.OSHMPI is an implementation of the OpenSHMEM standard using MPI-3 for the Linux operating system. It is the first implementation of SHMEM over MPI one-sided communication and has the potential to be widely adopted due to the portability and widely availability of Linux and MPI-3. OSHMPI has been tested on a variety of systemsmore » and implementations of MPI-3, includingInfiniBand clusters using MVAPICH2 and SGI shared-memory supercomputers using MPICH. Current support is limited to Linux but may be extended to Apple OSX if there is sufficient interest. The code is opensource via https://github.com/jeffhammond/oshmpi« less
Large-scale parallel lattice Boltzmann-cellular automaton model of two-dimensional dendritic growth
NASA Astrophysics Data System (ADS)
Jelinek, Bohumir; Eshraghi, Mohsen; Felicelli, Sergio; Peters, John F.
2014-03-01
An extremely scalable lattice Boltzmann (LB)-cellular automaton (CA) model for simulations of two-dimensional (2D) dendritic solidification under forced convection is presented. The model incorporates effects of phase change, solute diffusion, melt convection, and heat transport. The LB model represents the diffusion, convection, and heat transfer phenomena. The dendrite growth is driven by a difference between actual and equilibrium liquid composition at the solid-liquid interface. The CA technique is deployed to track the new interface cells. The computer program was parallelized using the Message Passing Interface (MPI) technique. Parallel scaling of the algorithm was studied and major scalability bottlenecks were identified. Efficiency loss attributable to the high memory bandwidth requirement of the algorithm was observed when using multiple cores per processor. Parallel writing of the output variables of interest was implemented in the binary Hierarchical Data Format 5 (HDF5) to improve the output performance, and to simplify visualization. Calculations were carried out in single precision arithmetic without significant loss in accuracy, resulting in 50% reduction of memory and computational time requirements. The presented solidification model shows a very good scalability up to centimeter size domains, including more than ten million of dendrites. Catalogue identifier: AEQZ_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEQZ_v1_0.html Program obtainable from: CPC Program Library, Queen’s University, Belfast, UK Licensing provisions: Standard CPC license, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 29,767 No. of bytes in distributed program, including test data, etc.: 3131,367 Distribution format: tar.gz Programming language: Fortran 90. Computer: Linux PC and clusters. Operating system: Linux. Has the code been vectorized or parallelized?: Yes. Program is parallelized using MPI. Number of processors used: 1-50,000 RAM: Memory requirements depend on the grid size Classification: 6.5, 7.7. External routines: MPI (http://www.mcs.anl.gov/research/projects/mpi/), HDF5 (http://www.hdfgroup.org/HDF5/) Nature of problem: Dendritic growth in undercooled Al-3 wt% Cu alloy melt under forced convection. Solution method: The lattice Boltzmann model solves the diffusion, convection, and heat transfer phenomena. The cellular automaton technique is deployed to track the solid/liquid interface. Restrictions: Heat transfer is calculated uncoupled from the fluid flow. Thermal diffusivity is constant. Unusual features: Novel technique, utilizing periodic duplication of a pre-grown “incubation” domain, is applied for the scaleup test. Running time: Running time varies from minutes to days depending on the domain size and number of computational cores.
Tycho 2: A Proxy Application for Kinetic Transport Sweeps
DOE Office of Scientific and Technical Information (OSTI.GOV)
Garrett, Charles Kristopher; Warsa, James S.
2016-09-14
Tycho 2 is a proxy application that implements discrete ordinates (SN) kinetic transport sweeps on unstructured, 3D, tetrahedral meshes. It has been designed to be small and require minimal dependencies to make collaboration and experimentation as easy as possible. Tycho 2 has been released as open source software. The software is currently in a beta release with plans for a stable release (version 1.0) before the end of the year. The code is parallelized via MPI across spatial cells and OpenMP across angles. Currently, several parallelization algorithms are implemented.
Ruisi, Michael; Levine, Michael; Finkielstein, Dennis
2013-12-01
The myocardial performance index (MPI) first described by Chuwa Tei in 1995 is a relatively new echocardiographic variable used for assessment of overall cardiac function. Previous studies have demonstrated the MPI to be a sum representation of both left ventricular systolic and diastolic function with prognostic value in patients with coronary artery disease as well as symptomatic heart failure. Ninety patients with either established coronary artery disease (CAD) or CAD risk factors underwent routine treadmill exercise stress testing with two-dimensional Doppler echocardiography using the standard Bruce protocol. Both resting and stress MPI values were measured for all 90 of the patients. Using a normal MPI cut off of ≤ 0.47, the prevalence of an abnormal resting MPI in our 90 subjects was 72/90 or 80% and the prevalence of an abnormal stress MPI in our 90 subjects was 48/90 or 53.33%. The average MPI observed in the resting portion of the stress test for the cohort was: 0.636 with a standard deviation of 0.182. The average MPI in the stress portion of the stress test for the cohort was 0.530 with a standard deviation of 0.250. The P value with the use of a one-tailed dependent T test was calculated to be < 0.05. We postulate that these findings reflect that the MPI (Tei) index assessed during exercise may be a sensitive indicator of occult coronary disease in an at risk group independent of wall motion assessment.
Modified personal interviews: resurrecting reliable personal interviews for admissions?
Hanson, Mark D; Kulasegaram, Kulamakan Mahan; Woods, Nicole N; Fechtig, Lindsey; Anderson, Geoff
2012-10-01
Traditional admissions personal interviews provide flexible faculty-student interactions but are plagued by low inter-interview reliability. Axelson and Kreiter (2009) retrospectively showed that multiple independent sampling (MIS) may improve reliability of personal interviews; thus, the authors incorporated MIS into the admissions process for medical students applying to the University of Toronto's Leadership Education and Development Program (LEAD). They examined the reliability and resource demands of this modified personal interview (MPI) format. In 2010-2011, LEAD candidates submitted written applications, which were used to screen for participation in the MPI process. Selected candidates completed four brief (10-12 minutes) independent MPIs each with a different interviewer. The authors blueprinted MPI questions to (i.e., aligned them with) leadership attributes, and interviewers assessed candidates' eligibility on a five-point Likert-type scale. The authors analyzed inter-interview reliability using the generalizability theory. Sixteen candidates submitted applications; 10 proceeded to the MPI stage. Reliability of the written application components was 0.75. The MPI process had overall inter-interview reliability of 0.79. Correlation between the written application and MPI scores was 0.49. A decision study showed acceptable reliability of 0.74 with only three MPIs scored using one global rating. Furthermore, a traditional admissions interview format would take 66% more time than the MPI format. The MPI format, used during the LEAD admissions process, achieved high reliability with minimal faculty resources. The MPI format's reliability and effective resource use were possible through MIS and employment of expert interviewers. MPIs may be useful for other admissions tasks.
Design analysis of an MPI human functional brain scanner
Mason, Erica E.; Cooley, Clarissa Z.; Cauley, Stephen F.; Griswold, Mark A.; Conolly, Steven M.; Wald, Lawrence L.
2017-01-01
MPI’s high sensitivity makes it a promising modality for imaging brain function. Functional contrast is proposed based on blood SPION concentration changes due to Cerebral Blood Volume (CBV) increases during activation, a mechanism utilized in fMRI studies. MPI offers the potential for a direct and more sensitive measure of SPION concentration, and thus CBV, than fMRI. As such, fMPI could surpass fMRI in sensitivity, enhancing the scientific and clinical value of functional imaging. As human-sized MPI systems have not been attempted, we assess the technical challenges of scaling MPI from rodent to human brain. We use a full-system MPI simulator to test arbitrary hardware designs and encoding practices, and we examine tradeoffs imposed by constraints that arise when scaling to human size as well as safety constraints (PNS and central nervous system stimulation) not considered in animal scanners, thereby estimating spatial resolutions and sensitivities achievable with current technology. Using a projection FFL MPI system, we examine coil hardware options and their implications for sensitivity and spatial resolution. We estimate that an fMPI brain scanner is feasible, although with reduced sensitivity (20×) and spatial resolution (5×) compared to existing rodent systems. Nonetheless, it retains sufficient sensitivity and spatial resolution to make it an attractive future instrument for studying the human brain; additional technical innovations can result in further improvements. PMID:28752130
Lee, M-Y; Won, H-S; Jeon, E-J; Yoon, H C; Choi, J Y; Hong, S J; Kim, M-J
2014-06-01
To evaluate the reproducibility of measurement of the fetal left modified myocardial performance index (Mod-MPI) determined using a novel automated system. This was a prospective study of 116 ultrasound examinations from 110 normal singleton pregnancies at 12 + 1 to 37 + 1 weeks' gestation. Two experienced operators each measured the left Mod-MPI twice manually and twice automatically using the Auto Mod-MPI system. Intra- and interoperator reproducibility were assessed using intraclass correlation coefficients (ICCs) and the manual and automated measurements obtained by the more experienced operator were compared using Bland-Altman plots and ICCs. Both operators successfully measured the left Mod-MPI in all cases using the Auto Mod-MPI system. For both operators, intraoperator reproducibility was higher when performing automated measurements (ICC = 0.967 and 0.962 for Operators 1 and 2, respectively) than when performing manual measurements (ICC = 0.857 and 0.856 for Operators 1 and 2, respectively). Interoperator agreement was also better for automated than for manual measurements (ICC = 0.930 vs 0.723, respectively). There was good agreement between the automated and manual values measured by the more experienced operator. The Auto Mod-MPI system is a reliable technique for measuring fetal left Mod-MPI and demonstrates excellent reproducibility. Copyright © 2013 ISUOG. Published by John Wiley & Sons Ltd.
Tejani, Furqan H; Thompson, Randall C; Iskandrian, Ami E; McNutt, Bruce E; Franks, Billy
2011-02-01
Caffeine attenuates the coronary hyperemic response to adenosine by competitive A₂(A) receptor blockade. This study aims to determine whether oral caffeine administration compromises diagnostic accuracy in patients undergoing vasodilator stress myocardial perfusion imaging (MPI) with regadenoson, a selective adenosine A(2A) agonist. This multicenter, randomized, double-blind, placebo-controlled, parallel-group study includes patients with suspected coronary artery disease who regularly consume caffeine. Each participant undergoes three SPECT MPI studies: a rest study on day 1 (MPI-1); a regadenoson stress study on day 3 (MPI-2), and a regadenoson stress study on day 5 with double-blind administration of oral caffeine 200 or 400 mg or placebo capsules (MPI-3; n = 90 per arm). Only participants with ≥ 1 reversible defect on the second MPI study undergo the subsequent stress MPI test. The primary endpoint is the difference in the number of reversible defects on the two stress tests using a 17-segment model. Pharmacokinetic/pharmacodynamic analyses will evaluate the effect of caffeine on the regadenoson exposure-response relationship. Safety will also be assessed. The results of this study will show whether the consumption of caffeine equivalent to 2-4 cups of coffee prior to an MPI study with regadenoson affects the diagnostic validity of stress testing (ClinicalTrials.gov number, NCT00826280).
Fernandes, José Maria G; Rivera, Ivan Romero; de Oliveira Romão, Benício; Mendonça, Maria Alayde; Vasconcelos, Miriam Lira Castro; Carvalho, Antônio Carlos; Campos, Orlando; De Paola, Angelo Amato V; Moisés, Valdir A
2009-09-01
The Doppler-derived myocardial performance index (MPI) has been used in the evaluation of left ventricular (LV) function in several diseases. In patients with isolated diastolic dysfunction, the diagnostic utility of this index remains unclear. The aim of this study was to determine the diagnostic utility of MPI in patients with systemic hypertension, impaired LV relaxation, and normal ejection fraction. Thirty hypertensive patients with impaired LV relaxation were compared to 30 control subjects. MPI and its components, isovolumetric relaxation time (IRT), isovolumetric contraction time (ICT), and the ejection time (ET), were measured from LV outflow and mitral inflow Doppler velocity profiles. MPI was higher in patients than in control subjects (0.45 +/- 0.13 vs 0.37 +/- 0.07 P < 0.0029). The increase in MPI was due to the prolongation of IRT without significant change of ICT and ET. MPI cutoff value of > or =0.40 identified impaired LV relaxation with a sensitivity of 63% and specificity of 70% while an IRT >94 ms had a sensitivity of 67% and specificity of 80%. Multivariate analysis identified relative wall thickness, mitral early filling wave velocity (E), and systolic myocardial velocity (Sm) as independent predictors of MPI in patients with hypertension. MPI was increase in patients with hypertension, diastolic dysfunction, and normal ejection fraction but was not superior to IRT to detect impaired LV relaxation.
Yao, Zhiming; Zhu, Hui; Li, Wenchan; Chen, Congxia; Wang, Hua; Shi, Lei; Zhang, Wenjie
2017-04-01
We investigated the cardiac risk stratification value of adenosine triphosphate stress myocardial perfusion imaging (ATP-MPI) in patients aged 70 years and older with suspected coronary artery disease (CAD). We identified a series of 415 consecutive patients aged 70 years and older with suspected CAD, who had undergone ATP-MPI with 99m Tc-MIBI. The presence of a fixed and/or reversible perfusion defect was considered as an abnormal MPI. Follow-up was available in 399 patients (96.1%) over 3.45 ± 1.71 years after excluding 16 patients who underwent early coronary revascularization <60 days after MPI. The major adverse cardiac events (MACE), including cardiac death, nonfatal infarction, and late coronary revascularization, were recorded. One hundred twenty-five (31.3%) patients had abnormal MPI and the remaining had normal MPI. A multivariable analysis using Cox regression demonstrated that abnormal MPI was independently associated with MACE (hazard ratio 19.50 and 95% confidence interval 5.91-64.31, P value .000). The patients with SSS > 8 had significantly higher cumulative MACE rate than patients with SSS ≤ 8 had (37.8% vs 5.2%, respectively, P < .001). The Kaplan-Meier cumulative MACE-free survival in patients with abnormal MPI (57.0%) was significantly lower than that in patients with normal MPI (89.6%), P < .0001. Among patients with SSS > 8, the Kaplan-Meier cumulative MACE-free survival were 36.9% in patients ≥80 years old and 49.5% in patients 70-79 years old, respectively, P < .05. However, among patients with SSS ≤ 8, there was no difference between the Kaplan-Meier cumulative MACE-free survivals of these two age groups. ATP-MPI data are useful for the prediction of major adverse cardiac events in patients aged 70 years and older with suspected CAD.
Initialization shock in decadal hindcasts due to errors in wind stress over the tropical Pacific
NASA Astrophysics Data System (ADS)
Pohlmann, Holger; Kröger, Jürgen; Greatbatch, Richard J.; Müller, Wolfgang A.
2017-10-01
Low prediction skill in the tropical Pacific is a common problem in decadal prediction systems, especially for lead years 2-5 which, in many systems, is lower than in uninitialized experiments. On the other hand, the tropical Pacific is of almost worldwide climate relevance through its teleconnections with other tropical and extratropical regions and also of importance for global mean temperature. Understanding the causes of the reduced prediction skill is thus of major interest for decadal climate predictions. We look into the problem of reduced prediction skill by analyzing the Max Planck Institute Earth System Model (MPI-ESM) decadal hindcasts for the fifth phase of the Climate Model Intercomparison Project and performing a sensitivity experiment in which hindcasts are initialized from a model run forced only by surface wind stress. In both systems, sea surface temperature variability in the tropical Pacific is successfully initialized, but most skill is lost at lead years 2-5. Utilizing the sensitivity experiment enables us to pin down the reason for the reduced prediction skill in MPI-ESM to errors in wind stress used for the initialization. A spurious trend in the wind stress forcing displaces the equatorial thermocline in MPI-ESM unrealistically. When the climate model is then switched into its forecast mode, the recovery process triggers artificial El Niño and La Niña events at the surface. Our results demonstrate the importance of realistic wind stress products for the initialization of decadal predictions.
Functional Properties of a High Protein Beverage Stabilized with Oat-β-Glucan.
Vasquez-Orejarena, Eva; Simons, Christopher T; Litchfield, John H; Alvarez, Valente B
2018-05-01
This study evaluated the effect of oat flour and milk protein on the functional properties and sensory acceptability of shelf stable high protein dairy beverages containing at least 0.75 g of oat-β-glucan per serving size. Formulations adjusted to levels of 1.50% to 2.30% oat flour and 2.50% to 4.00% milk protein isolate (MPI) were thermal processed in a rotary retort. The finished product exhibited good suspension stability (>80%). The increase of oat and MPI contents lead to nectar-like beverages (51 to 100 mPas). However, oat flour was the component showing the highest effect on the viscosity coefficient values of the beverages. Sensory evaluation indicated that formulations with less than 1.9% oat flour and 2.5% MPI (thin liquid, <50 mPas) were the most accepted. Mouthfeel (perceived thickness), sweetness and aftertaste had the most influence on overall liking of the beverages. Overall, this study comprises the development of a functional food product. Supplementation of beverages with fiber from oats is an innovative approach to stabilize high protein beverages. Ready to drink protein beverage formulations use gums to stabilize the product and provide a desirable mouthfeel. The levels of oat-β-glucan used in the beverage increased the thickness and meet the requirement of the FDA approved health claim for reduction of the cardiovascular disease risk (21 CFR 101.81). © 2018 Institute of Food Technologists®.
Ojeda-May, Pedro; Nam, Kwangho
2017-08-08
The strategy and implementation of scalable and efficient semiempirical (SE) QM/MM methods in CHARMM are described. The serial version of the code was first profiled to identify routines that required parallelization. Afterward, the code was parallelized and accelerated with three approaches. The first approach was the parallelization of the entire QM/MM routines, including the Fock matrix diagonalization routines, using the CHARMM message passage interface (MPI) machinery. In the second approach, two different self-consistent field (SCF) energy convergence accelerators were implemented using density and Fock matrices as targets for their extrapolations in the SCF procedure. In the third approach, the entire QM/MM and MM energy routines were accelerated by implementing the hybrid MPI/open multiprocessing (OpenMP) model in which both the task- and loop-level parallelization strategies were adopted to balance loads between different OpenMP threads. The present implementation was tested on two solvated enzyme systems (including <100 QM atoms) and an S N 2 symmetric reaction in water. The MPI version exceeded existing SE QM methods in CHARMM, which include the SCC-DFTB and SQUANTUM methods, by at least 4-fold. The use of SCF convergence accelerators further accelerated the code by ∼12-35% depending on the size of the QM region and the number of CPU cores used. Although the MPI version displayed good scalability, the performance was diminished for large numbers of MPI processes due to the overhead associated with MPI communications between nodes. This issue was partially overcome by the hybrid MPI/OpenMP approach which displayed a better scalability for a larger number of CPU cores (up to 64 CPUs in the tested systems).
Magnetic Particle Imaging (MPI) for NMR and MRI researchers
NASA Astrophysics Data System (ADS)
Saritas, Emine U.; Goodwill, Patrick W.; Croft, Laura R.; Konkle, Justin J.; Lu, Kuan; Zheng, Bo; Conolly, Steven M.
2013-04-01
Magnetic Particle Imaging (MPI) is a new tracer imaging modality that is gaining significant interest from NMR and MRI researchers. While the physics of MPI differ substantially from MRI, it employs hardware and imaging concepts that are familiar to MRI researchers, such as magnetic excitation and detection, pulse sequences, and relaxation effects. Furthermore, MPI employs the same superparamagnetic iron oxide (SPIO) contrast agents that are sometimes used for MR angiography and are often used for MRI cell tracking studies. These SPIOs are much safer for humans than iodine or gadolinium, especially for Chronic Kidney Disease (CKD) patients. The weak kidneys of CKD patients cannot safely excrete iodine or gadolinium, leading to increased morbidity and mortality after iodinated X-ray or CT angiograms, or after gadolinium-MRA studies. Iron oxides, on the other hand, are processed in the liver, and have been shown to be safe even for CKD patients. Unlike the “black blood” contrast generated by SPIOs in MRI due to increased T2∗ dephasing, SPIOs in MPI generate positive, “bright blood” contrast. With this ideal contrast, even prototype MPI scanners can already achieve fast, high-sensitivity, and high-contrast angiograms with millimeter-scale resolutions in phantoms and in animals. Moreover, MPI shows great potential for an exciting array of applications, including stem cell tracking in vivo, first-pass contrast studies to diagnose or stage cancer, and inflammation imaging in vivo. So far, only a handful of prototype small-animal MPI scanners have been constructed worldwide. Hence, MPI is open to great advances, especially in hardware, pulse sequence, and nanoparticle improvements, with the potential to revolutionize the biomedical imaging field.
Performance comparison analysis library communication cluster system using merge sort
NASA Astrophysics Data System (ADS)
Wulandari, D. A. R.; Ramadhan, M. E.
2018-04-01
Begins by using a single processor, to increase the speed of computing time, the use of multi-processor was introduced. The second paradigm is known as parallel computing, example cluster. The cluster must have the communication potocol for processing, one of it is message passing Interface (MPI). MPI have many library, both of them OPENMPI and MPICH2. Performance of the cluster machine depend on suitable between performance characters of library communication and characters of the problem so this study aims to analyze the comparative performances libraries in handling parallel computing process. The case study in this research are MPICH2 and OpenMPI. This case research execute sorting’s problem to know the performance of cluster system. The sorting problem use mergesort method. The research method is by implementing OpenMPI and MPICH2 on a Linux-based cluster by using five computer virtual then analyze the performance of the system by different scenario tests and three parameters for to know the performance of MPICH2 and OpenMPI. These performances are execution time, speedup and efficiency. The results of this study showed that the addition of each data size makes OpenMPI and MPICH2 have an average speed-up and efficiency tend to increase but at a large data size decreases. increased data size doesn’t necessarily increased speed up and efficiency but only execution time example in 100000 data size. OpenMPI has a execution time greater than MPICH2 example in 1000 data size average execution time with MPICH2 is 0,009721 and OpenMPI is 0,003895 OpenMPI can customize communication needs.
Martin, Wade H; Xian, Hong; Chandiramani, Pooja; Bainter, Emily; Klein, Andrew J P
2015-08-01
No data exist comparing outcome prediction from arm exercise vs pharmacologic myocardial perfusion imaging (MPI) stress test variables in patients unable to perform treadmill exercise. In this retrospective study, 2,173 consecutive lower extremity disabled veterans aged 65.4 ± 11.0years (mean ± SD) underwent either pharmacologic MPI (1730 patients) or arm exercise stress tests (443 patients) with MPI (n = 253) or electrocardiography alone (n = 190) between 1997 and 2002. Cox multivariate regression models and reclassification analysis by integrated discrimination improvement (IDI) were used to characterize stress test and MPI predictors of cardiovascular mortality at ≥10-year follow-up after inclusion of significant demographic, clinical, and other variables. Cardiovascular death occurred in 561 pharmacologic MPI and 102 arm exercise participants. Multivariate-adjusted cardiovascular mortality was predicted by arm exercise resting metabolic equivalents (hazard ratio [HR] 0.52, 95% CI 0.39-0.69, P < .001), 1-minute heart rate recovery (HR 0.61, 95% CI 0.44-0.86, P < .001), and pharmacologic and arm exercise delta (peak-rest) heart rate (both P < .001). Only an abnormal arm exercise MPI prognosticated cardiovascular death by multivariate Cox analysis (HR 1.98, 95% CI 1.04-3.77, P < .05). Arm exercise MPI defect number, type, and size provided IDI over covariates for prediction of cardiovascular mortality (IDI = 0.074-0.097). Only pharmacologic defect size prognosticated cardiovascular mortality (IDI = 0.022). Arm exercise capacity, heart rate recovery, and pharmacologic and arm exercise heart rate responses are robust predictors of cardiovascular mortality. Arm exercise MPI results are equivalent and possibly superior to pharmacologic MPI for cardiovascular mortality prediction in patients unable to perform treadmill exercise. Published by Elsevier Inc.
Forberg, Jakob L; Hilmersson, Catarina E; Carlsson, Marcus; Arheden, Håkan; Björk, Jonas; Hjalte, Krister; Ekelund, Ulf
2009-01-01
Background Previous studies from the USA have shown that acute nuclear myocardial perfusion imaging (MPI) in low risk emergency department (ED) patients with suspected acute coronary syndrome (ACS) can be of clinical value. The aim of this study was to evaluate the utility and hospital economics of acute MPI in Swedish ED patients with suspected ACS. Methods We included 40 patients (mean age 55 ± 2 years, 50% women) who were admitted from the ED at Lund University Hospital for chest pain suspicious of ACS, and who had a normal or non-ischemic ECG and no previous myocardial infarction. All patients underwent MPI from the ED, and the results were analyzed only after patient discharge. The current diagnostic practice of admitting the included patients for observation and further evaluation was compared to a theoretical "MPI strategy", where patients with a normal MPI test would have been discharged home from the ED. Results Twenty-seven patients had normal MPI results, and none of them had ACS. MPI thus had a negative predictive value for ACS of 100%. With the MPI strategy, 2/3 of the patients would thus have been discharged from the ED, resulting in a reduction of total hospital cost by some 270 EUR and of bed occupancy by 0.8 days per investigated patient. Conclusion Our findings in a Swedish ED support the results of larger American trials that acute MPI has the potential to safely reduce the number of admissions and decrease overall costs for low-risk ED patients with suspected ACS. PMID:19545365
Fortran code for SU(3) lattice gauge theory with and without MPI checkerboard parallelization
NASA Astrophysics Data System (ADS)
Berg, Bernd A.; Wu, Hao
2012-10-01
We document plain Fortran and Fortran MPI checkerboard code for Markov chain Monte Carlo simulations of pure SU(3) lattice gauge theory with the Wilson action in D dimensions. The Fortran code uses periodic boundary conditions and is suitable for pedagogical purposes and small scale simulations. For the Fortran MPI code two geometries are covered: the usual torus with periodic boundary conditions and the double-layered torus as defined in the paper. Parallel computing is performed on checkerboards of sublattices, which partition the full lattice in one, two, and so on, up to D directions (depending on the parameters set). For updating, the Cabibbo-Marinari heatbath algorithm is used. We present validations and test runs of the code. Performance is reported for a number of currently used Fortran compilers and, when applicable, MPI versions. For the parallelized code, performance is studied as a function of the number of processors. Program summary Program title: STMC2LSU3MPI Catalogue identifier: AEMJ_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEMJ_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 26666 No. of bytes in distributed program, including test data, etc.: 233126 Distribution format: tar.gz Programming language: Fortran 77 compatible with the use of Fortran 90/95 compilers, in part with MPI extensions. Computer: Any capable of compiling and executing Fortran 77 or Fortran 90/95, when needed with MPI extensions. Operating system: Red Hat Enterprise Linux Server 6.1 with OpenMPI + pgf77 11.8-0, Centos 5.3 with OpenMPI + gfortran 4.1.2, Cray XT4 with MPICH2 + pgf90 11.2-0. Has the code been vectorised or parallelized?: Yes, parallelized using MPI extensions. Number of processors used: 2 to 11664 RAM: 200 Mega bytes per process. Classification: 11.5. Nature of problem: Physics of pure SU(3) Quantum Field Theory (QFT). This is relevant for our understanding of Quantum Chromodynamics (QCD). It includes the glueball spectrum, topological properties and the deconfining phase transition of pure SU(3) QFT. For instance, Relativistic Heavy Ion Collision (RHIC) experiments at the Brookhaven National Laboratory provide evidence that quarks confined in hadrons undergo at high enough temperature and pressure a transition into a Quark-Gluon Plasma (QGP). Investigations of its thermodynamics in pure SU(3) QFT are of interest. Solution method: Markov Chain Monte Carlo (MCMC) simulations of SU(3) Lattice Gauge Theory (LGT) with the Wilson action. This is a regularization of pure SU(3) QFT on a hypercubic lattice, which allows approaching the continuum SU(3) QFT by means of Finite Size Scaling (FSS) studies. Specifically, we provide updating routines for the Cabibbo-Marinari heatbath with and without checkerboard parallelization. While the first is suitable for pedagogical purposes and small scale projects, the latter allows for efficient parallel processing. Targetting the geometry of RHIC experiments, we have implemented a Double-Layered Torus (DLT) lattice geometry, which has previously not been used in LGT MCMC simulations and enables inside and outside layers at distinct temperatures, the lower-temperature layer acting as the outside boundary for the higher-temperature layer, where the deconfinement transition goes on. Restrictions: The checkerboard partition of the lattice makes the development of measurement programs more tedious than is the case for an unpartitioned lattice. Presently, only one measurement routine for Polyakov loops is provided. Unusual features: We provide three different versions for the send/receive function of the MPI library, which work for different operating system +compiler +MPI combinations. This involves activating the correct row in the last three rows of our latmpi.par parameter file. The underlying reason is distinct buffer conventions. Running time: For a typical run using an Intel i7 processor, it takes (1.8-6) E-06 seconds to update one link of the lattice, depending on the compiler used. For example, if we do a simulation on a small (4 * 83) DLT lattice with a statistics of 221 sweeps (i.e., update the two lattice layers of 4 * (4 * 83) links each 221 times), the total CPU time needed can be 2 * 4 * (4 * 83) * 221 * 3 E-06 seconds = 1.7 minutes, where 2 — two layers of lattice 4 — four dimensions 83 * 4 — lattice size 221 — sweeps of updating 6 E-06 s mdash; average time to update one link variable. If we divide the job into 8 parallel processes, then the real time is (for negligible communication overhead) 1.7 mins / 8 = 0.2 mins.
Development and Initial Validation of the Multicultural Personality Inventory (MPI).
Ponterotto, Joseph G; Fietzer, Alexander W; Fingerhut, Esther C; Woerner, Scott; Stack, Lauren; Magaldi-Dopman, Danielle; Rust, Jonathan; Nakao, Gen; Tsai, Yu-Ting; Black, Natasha; Alba, Renaldo; Desai, Miraj; Frazier, Chantel; LaRue, Alyse; Liao, Pei-Wen
2014-01-01
Two studies summarize the development and initial validation of the Multicultural Personality Inventory (MPI). In Study 1, the 115-item prototype MPI was administered to 415 university students where exploratory factor analysis resulted in a 70-item, 7-factor model. In Study 2, the 70-item MPI and theoretically related companion instruments were administered to a multisite sample of 576 university students. Confirmatory factory analysis found the 7-factor structure to be a relatively good fit to the data (Comparative Fit Index =.954; root mean square error of approximation =.057), and MPI factors predicted variance in criterion variables above and beyond the variance accounted for by broad personality traits (i.e., Big Five). Study limitations and directions for further validation research are specified.
Hughes, Tyler; Shcherbinin, Sergey; Celler, Anna
2011-07-01
Normal patient databases (NPDs) are used to distinguish between normal and abnormal perfusion in SPECT myocardial perfusion imaging (MPI) and have gained wide acceptance in the clinical environment, yet there are limitations to this approach. This study introduces a template-based method for semi-quantitative MPI, which attempts to overcome some of the NPD limitations. Our approach involves the construction of a 3D digital healthy heart template from the delineation of the patient's left ventricle in the SPECT image. This patient-specific template of the heart, filled with uniform activity, is then analytically projected and reconstructed using the same algorithm as the original image. Subsequent to generating bulls-eye maps for the patient image (PB) and the template image (TB), a ratio (PB/TB) is calculated, which produces a reconstruction-artifact corrected image (CB). Finally, a threshold is used to define defects within CB enabling measurements of the perfusion defect extent (EXT). The SPECT-based template (Ts) measurements were compared to those of a CT-based "ideal" template (TI). Twenty digital phantoms were simulated: male and female, each with one healthy heart and nine hearts with various defects. Four physical phantom studies were performed modeling a healthy heart and three hearts with different defects. The phantom represented a thorax with spine, lung, and left ventricle inserts. Images were acquired on General Electric's (GE) Infinia Hawkeye SPECT/CT camera using standard clinical MPI protocol. Finally, our method was applied to 14 patient MPI rest/stress studies acquired on the GE Infinia Hawkeye SPECT/CT camera and compared to the results obtained from Cedars-Sinai's QPS software. In the simulation studies, the true EXT correlated well with the TI (slope= 1.08; offset = -0.40%; r = 0.99) and Ts (slope = 0.90; offset = 0.27%; r = 0.99) methods with no significant differences between them. Similarly, strong correlations were measured for EXT obtained from QPS and the template method for patient studies (slope =0.91; offset = 0.45%; r = 0.98). Mean errors in extent for the Ts method using simulation, physical phantom, and patient data were 2.7% +/- 2.4%, 0.9% +/- 0.5%, 2.0% +/- 2.7%, respectively. The authors introduced a method for semi-quantitative SPECT MPI, which offers a patient-specific approach to define the perfusion defect regions within the heart, as opposed to the patient-averaged NPD methodology.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gyllenhaal, J.
CLOMP is the C version of the Livermore OpenMP benchmark developed to measure OpenMP overheads and other performance impacts due to threading. For simplicity, it does not use MPI by default but it is expected to be run on the resources a threaded MPI task would use (e.g., a portion of a shared memory compute node). Compiling with -DWITH_MPI allows packing one or more nodes with CLOMP tasks and having CLOMP report OpenMP performance for the slowest MPI task. On current systems, the strong scaling performance results for 4, 8, or 16 threads are of the most interest. Suggested weakmore » scaling inputs are provided for evaluating future systems. Since MPI is often used to place at least one MPI task per coherence or NUMA domain, it is recommended to focus OpenMP runtime measurements on a subset of node hardware where it is most possible to have low OpenMP overheads (e.g., within one coherence domain or NUMA domain).« less
NASA Astrophysics Data System (ADS)
Behrens, Jörg; Hanke, Moritz; Jahns, Thomas
2014-05-01
In this talk we present a way to facilitate efficient use of MPI communication for developers of climate models. Exploitation of the performance potential of today's highly parallel supercomputers with real world simulations is a complex task. This is partly caused by the low level nature of the MPI communication library which is the dominant communication tool at least for inter-node communication. In order to manage the complexity of the task, climate simulations with non-trivial communication patterns often use an internal abstraction layer above MPI without exploiting the benefits of communication aggregation or MPI-datatypes. The solution for the complexity and performance problem we propose is the communication library YAXT. This library is built on top of MPI and takes high level descriptions of arbitrary domain decompositions and automatically derives an efficient collective data exchange. Several exchanges can be aggregated in order to reduce latency costs. Examples are given which demonstrate the simplicity and the performance gains for selected climate applications.
Eriksson, Mathilda; Andreasson, Kalle; Weidmann, Joachim; Lundberg, Kajsa; Tegerstedt, Karin
2011-01-01
Virus-like particles (VLPs) consist of capsid proteins from viruses and have been shown to be usable as carriers of protein and peptide antigens for immune therapy. In this study, we have produced and assayed murine polyomavirus (MPyV) VLPs carrying the entire human Prostate Specific Antigen (PSA) (PSA-MPyVLPs) for their potential use for immune therapy in a mouse model system. BALB/c mice immunized with PSA-MPyVLPs were only marginally protected against outgrowth of a PSA-expressing tumor. To improve protection, PSA-MPyVLPs were co-injected with adjuvant CpG, either alone or loaded onto murine dendritic cells (DCs). Immunization with PSA-MPyVLPs loaded onto DCs in the presence of CpG was shown to efficiently protect mice from tumor outgrowth. In addition, cellular and humoral immune responses after immunization were examined. PSA-specific CD4+ and CD8+ cells were demonstrated, but no PSA-specific IgG antibodies. Vaccination with DCs loaded with PSA-MPyVLPs induced an eight-fold lower titre of anti-VLP antibodies than vaccination with PSA-MPyVLPs alone. In conclusion, immunization of BALB/c mice with PSA-MPyVLPs, loaded onto DCs and co-injected with CpG, induces an efficient PSA-specific tumor protective immune response, including both CD4+ and CD8+ cells with a low induction of anti-VLP antibodies. PMID:21858228
Pilotto, Alberto; Ferrucci, Luigi; Scarcelli, Carlo; Niro, Valeria; Di Mario, Francesco; Seripa, Davide; Andriulli, Angelo; Leandro, Gioacchino; Franceschi, Marilisa
2007-01-01
The potential usefulness of standardized comprehensive geriatric assessment (CGA) in evaluating treatment and follow-up of older patients with upper gastrointestinal bleeding is unknown. To evaluate the usefulness of the CGA as a 2-year mortality multidimensional prognostic index (MPI) in older patients hospitalized for upper gastrointestinal bleeding. Patients aged > or =65 years consecutively hospitalized for acute upper gastrointestinal bleeding were included. Diagnosis of bleeding was based on clinical and endoscopic features. All patients underwent a CGA that included six standardized scales, i.e., Activities of Daily Living (ADL), Instrumental Activities of Daily Living (IADL), Short Portable Mental Status Questionnaire (SPMSQ), Mini Nutritional Assessment (MNA), Exton-Smith Score (ESS) and Comorbity Index Rating Scale (CIRS), as well as information on medication history and cohabitation, for a total of 63 items. A MPI was calculated from the integrated total scores and expressed as MPI 1 = low risk, MPI 2 = moderate risk, and MPI 3 = severe risk. The predictive value of the MPI for mortality over a 24-month follow-up was calculated. 36 elderly patients (M 16/F 20, mean age 82.8 +/- 7.9 years, range 70-101 years) were included in the study. A significant difference in mean age was observed between males and females (M 80.1 +/- 4.8 vs. F 84.9 +/- 9.3 years; p < 0.05). The causes of upper gastrointestinal bleeding were duodenal ulcer in 38.8%, gastric ulcer in 22.2%, and erosive gastritis in 16.6% of the patients, while 16.6% had gastrointestinal bleeding from unknown origin. The overall 2-year mortality rate was 30.5%. 18 patients (50%) were classified as having a low-risk MPI (mean value 0.18 +/- 0.09), 12 (33.3%) as having a moderate-risk MPI (mean value 0.48 +/- 0.08) and 6 (16.6%) as having a severe-risk MPI (mean value 0.83 +/- 0.06). Higher MPI grades were significantly associated with higher mortality (grade 1 = 12.5%, grade 2 = 41.6%, grade 3 = 83.3%; p = 0.001). Adjusting for age and sex, the prognostic efficacy of MPI for mortality was confirmed and highly significant (odds ratio 10.47, 95% CI 2.04-53.6). CGA is a useful tool for calculating a MPI that significantly predicts the risk of 2-year mortality in older patients with upper gastrointestinal bleeding. Copyright 2007 S. Karger AG, Basel.
MultiNest: Efficient and Robust Bayesian Inference
NASA Astrophysics Data System (ADS)
Feroz, F.; Hobson, M. P.; Bridges, M.
2011-09-01
We present further development and the first public release of our multimodal nested sampling algorithm, called MultiNest. This Bayesian inference tool calculates the evidence, with an associated error estimate, and produces posterior samples from distributions that may contain multiple modes and pronounced (curving) degeneracies in high dimensions. The developments presented here lead to further substantial improvements in sampling efficiency and robustness, as compared to the original algorithm presented in Feroz & Hobson (2008), which itself significantly outperformed existing MCMC techniques in a wide range of astrophysical inference problems. The accuracy and economy of the MultiNest algorithm is demonstrated by application to two toy problems and to a cosmological inference problem focusing on the extension of the vanilla LambdaCDM model to include spatial curvature and a varying equation of state for dark energy. The MultiNest software is fully parallelized using MPI and includes an interface to CosmoMC. It will also be released as part of the SuperBayeS package, for the analysis of supersymmetric theories of particle physics, at this http URL.
Design and implementation of a hybrid MPI-CUDA model for the Smith-Waterman algorithm.
Khaled, Heba; Faheem, Hossam El Deen Mostafa; El Gohary, Rania
2015-01-01
This paper provides a novel hybrid model for solving the multiple pair-wise sequence alignment problem combining message passing interface and CUDA, the parallel computing platform and programming model invented by NVIDIA. The proposed model targets homogeneous cluster nodes equipped with similar Graphical Processing Unit (GPU) cards. The model consists of the Master Node Dispatcher (MND) and the Worker GPU Nodes (WGN). The MND distributes the workload among the cluster working nodes and then aggregates the results. The WGN performs the multiple pair-wise sequence alignments using the Smith-Waterman algorithm. We also propose a modified implementation to the Smith-Waterman algorithm based on computing the alignment matrices row-wise. The experimental results demonstrate a considerable reduction in the running time by increasing the number of the working GPU nodes. The proposed model achieved a performance of about 12 Giga cell updates per second when we tested against the SWISS-PROT protein knowledge base running on four nodes.
Employing multi-GPU power for molecular dynamics simulation: an extension of GALAMOST
NASA Astrophysics Data System (ADS)
Zhu, You-Liang; Pan, Deng; Li, Zhan-Wei; Liu, Hong; Qian, Hu-Jun; Zhao, Yang; Lu, Zhong-Yuan; Sun, Zhao-Yan
2018-04-01
We describe the algorithm of employing multi-GPU power on the basis of Message Passing Interface (MPI) domain decomposition in a molecular dynamics code, GALAMOST, which is designed for the coarse-grained simulation of soft matters. The code of multi-GPU version is developed based on our previous single-GPU version. In multi-GPU runs, one GPU takes charge of one domain and runs single-GPU code path. The communication between neighbouring domains takes a similar algorithm of CPU-based code of LAMMPS, but is optimised specifically for GPUs. We employ a memory-saving design which can enlarge maximum system size at the same device condition. An optimisation algorithm is employed to prolong the update period of neighbour list. We demonstrate good performance of multi-GPU runs on the simulation of Lennard-Jones liquid, dissipative particle dynamics liquid, polymer and nanoparticle composite, and two-patch particles on workstation. A good scaling of many nodes on cluster for two-patch particles is presented.
Singapore Students' Performance on Australian and Singapore Assessment Items
ERIC Educational Resources Information Center
Ho, Siew Yin; Lowrie, Tom
2012-01-01
This study describes Singapore students' (N = 607) performance on a recently developed Mathematics Processing Instrument (MPI). The MPI comprised tasks sourced from Australia's NAPLAN and Singapore's PSLE. In addition, the MPI had a corresponding question which encouraged students to describe how they solved the respective tasks. In particular,…
Zhou, Xinyi Y; Tay, Zhi Wei; Chandrasekharan, Prashant; Yu, Elaine Y; Hensley, Daniel W; Orendorff, Ryan; Jeffris, Kenneth E; Mai, David; Zheng, Bo; Goodwill, Patrick W; Conolly, Steven M
2018-05-10
Magnetic particle imaging (MPI) is an emerging ionizing radiation-free biomedical tracer imaging technique that directly images the intense magnetization of superparamagnetic iron oxide nanoparticles (SPIOs). MPI offers ideal image contrast because MPI shows zero signal from background tissues. Moreover, there is zero attenuation of the signal with depth in tissue, allowing for imaging deep inside the body quantitatively at any location. Recent work has demonstrated the potential of MPI for robust, sensitive vascular imaging and cell tracking with high contrast and dose-limited sensitivity comparable to nuclear medicine. To foster future applications in MPI, this new biomedical imaging field is welcoming researchers with expertise in imaging physics, magnetic nanoparticle synthesis and functionalization, nanoscale physics, and small animal imaging applications. Copyright © 2018 Elsevier Ltd. All rights reserved.
MPI_XSTAR: MPI-based Parallelization of the XSTAR Photoionization Program
NASA Astrophysics Data System (ADS)
Danehkar, Ashkbiz; Nowak, Michael A.; Lee, Julia C.; Smith, Randall K.
2018-02-01
We describe a program for the parallel implementation of multiple runs of XSTAR, a photoionization code that is used to predict the physical properties of an ionized gas from its emission and/or absorption lines. The parallelization program, called MPI_XSTAR, has been developed and implemented in the C++ language by using the Message Passing Interface (MPI) protocol, a conventional standard of parallel computing. We have benchmarked parallel multiprocessing executions of XSTAR, using MPI_XSTAR, against a serial execution of XSTAR, in terms of the parallelization speedup and the computing resource efficiency. Our experience indicates that the parallel execution runs significantly faster than the serial execution, however, the efficiency in terms of the computing resource usage decreases with increasing the number of processors used in the parallel computing.
OPAL: An Open-Source MPI-IO Library over Cray XT
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yu, Weikuan; Vetter, Jeffrey S; Canon, Richard Shane
Parallel IO over Cray XT is supported by a vendor-supplied MPI-IO package. This package contains a proprietary ADIO implementation built on top of the sysio library. While it is reasonable to maintain a stable code base for application scientists' convenience, it is also very important to the system developers and researchers to analyze and assess the effectiveness of parallel IO software, and accordingly, tune and optimize the MPI-IO implementation. A proprietary parallel IO code base relinquishes such flexibilities. On the other hand, a generic UFS-based MPI-IO implementation is typically used on many Linux-based platforms. We have developed an open-source MPI-IOmore » package over Lustre, referred to as OPAL (OPportunistic and Adaptive MPI-IO Library over Lustre). OPAL provides a single source-code base for MPI-IO over Lustre on Cray XT and Linux platforms. Compared to Cray implementation, OPAL provides a number of good features, including arbitrary specification of striping patterns and Lustre-stripe aligned file domain partitioning. This paper presents the performance comparisons between OPAL and Cray's proprietary implementation. Our evaluation demonstrates that OPAL achieves the performance comparable to the Cray implementation. We also exemplify the benefits of an open source package in revealing the underpinning of the parallel IO performance.« less
NASA Astrophysics Data System (ADS)
Pablico-Lansigan, Michele H.; Situ, Shu F.; Samia, Anna Cristina S.
2013-05-01
Magnetic particle imaging (MPI) is an emerging biomedical imaging technology that allows the direct quantitative mapping of the spatial distribution of superparamagnetic iron oxide nanoparticles. MPI's increased sensitivity and short image acquisition times foster the creation of tomographic images with high temporal and spatial resolution. The contrast and sensitivity of MPI is envisioned to transcend those of other medical imaging modalities presently used, such as magnetic resonance imaging (MRI), X-ray scans, ultrasound, computed tomography (CT), positron emission tomography (PET) and single photon emission computed tomography (SPECT). In this review, we present an overview of the recent advances in the rapidly developing field of MPI. We begin with a basic introduction of the fundamentals of MPI, followed by some highlights over the past decade of the evolution of strategies and approaches used to improve this new imaging technique. We also examine the optimization of iron oxide nanoparticle tracers used for imaging, underscoring the importance of size homogeneity and surface engineering. Finally, we present some future research directions for MPI, emphasizing the novel and exciting opportunities that it offers as an important tool for real-time in vivo monitoring. All these opportunities and capabilities that MPI presents are now seen as potential breakthrough innovations in timely disease diagnosis, implant monitoring, and image-guided therapeutics.
Comparative accuracy of supine-only and combined supine-prone myocardial perfusion imaging in men.
Taasan, Vicente; Wokhlu, Anita; Taasan, Michael V; Dusaj, Raman S; Mehta, Ajay; Kraft, Steven; Winchester, David; Wymer, David
2016-12-01
Combined supine-prone myocardial perfusion imaging (CSP MPI) has been shown to reduce attenuation artifact in comparison to supine-only (SU) MPI in mixed-gender populations with varying risk for coronary artery disease (CAD), often where patients served as their own controls. However, there is limited direct comparison of these imaging strategies in men. 934 male patients underwent CSP or SU MPI. Diagnostic certainty of interpretation was compared. Within the cohort, 116 were referred for left heart catheterization (LHC) to assess for CAD. Sensitivity, specificity, and area under the curve (AUC) were compared with additional analysis based on body mass index (BMI). 597 patients completed the SU protocol and 337 patients completed the CSP protocol. Equivocal studies were seen more frequently in the SU group (13%) than in the CSP group (4%, P < .001). At catheterization, the specificity for CSP MPI of 70% was higher than 40% for SU MPI (P = .032). The CSP AUC (0.80 ± 0.06) was significantly larger than SU AUC (0.57 ± 0.05, P = .004). CSP specificity was significantly higher in obese patients. CSP MPI increases diagnostic certainty and improves test accuracy for CAD detection in men with CAD risk factors, especially obese patients, compared to SU MPI.
Direct parametric reconstruction in dynamic PET myocardial perfusion imaging: in vivo studies.
Petibon, Yoann; Rakvongthai, Yothin; El Fakhri, Georges; Ouyang, Jinsong
2017-05-07
Dynamic PET myocardial perfusion imaging (MPI) used in conjunction with tracer kinetic modeling enables the quantification of absolute myocardial blood flow (MBF). However, MBF maps computed using the traditional indirect method (i.e. post-reconstruction voxel-wise fitting of kinetic model to PET time-activity-curves-TACs) suffer from poor signal-to-noise ratio (SNR). Direct reconstruction of kinetic parameters from raw PET projection data has been shown to offer parametric images with higher SNR compared to the indirect method. The aim of this study was to extend and evaluate the performance of a direct parametric reconstruction method using in vivo dynamic PET MPI data for the purpose of quantifying MBF. Dynamic PET MPI studies were performed on two healthy pigs using a Siemens Biograph mMR scanner. List-mode PET data for each animal were acquired following a bolus injection of ~7-8 mCi of 18 F-flurpiridaz, a myocardial perfusion agent. Fully-3D dynamic PET sinograms were obtained by sorting the coincidence events into 16 temporal frames covering ~5 min after radiotracer administration. Additionally, eight independent noise realizations of both scans-each containing 1/8th of the total number of events-were generated from the original list-mode data. Dynamic sinograms were then used to compute parametric maps using the conventional indirect method and the proposed direct method. For both methods, a one-tissue compartment model accounting for spillover from the left and right ventricle blood-pools was used to describe the kinetics of 18 F-flurpiridaz. An image-derived arterial input function obtained from a TAC taken in the left ventricle cavity was used for tracer kinetic analysis. For the indirect method, frame-by-frame images were estimated using two fully-3D reconstruction techniques: the standard ordered subset expectation maximization (OSEM) reconstruction algorithm on one side, and the one-step late maximum a posteriori (OSL-MAP) algorithm on the other side, which incorporates a quadratic penalty function. The parametric images were then calculated using voxel-wise weighted least-square fitting of the reconstructed myocardial PET TACs. For the direct method, parametric images were estimated directly from the dynamic PET sinograms using a maximum a posteriori (MAP) parametric reconstruction algorithm which optimizes an objective function comprised of the Poisson log-likelihood term, the kinetic model and a quadratic penalty function. Maximization of the objective function with respect to each set of parameters was achieved using a preconditioned conjugate gradient algorithm with a specifically developed pre-conditioner. The performance of the direct method was evaluated by comparing voxel- and segment-wise estimates of [Formula: see text], the tracer transport rate (ml · min -1 · ml -1 ), to those obtained using the indirect method applied to both OSEM and OSL-MAP dynamic reconstructions. The proposed direct reconstruction method produced [Formula: see text] maps with visibly lower noise than the indirect method based on OSEM and OSL-MAP reconstructions. At normal count levels, the direct method was shown to outperform the indirect method based on OSL-MAP in the sense that at matched level of bias, reduced regional noise levels were obtained. At lower count levels, the direct method produced [Formula: see text] estimates with significantly lower standard deviation across noise realizations than the indirect method based on OSL-MAP at matched bias level. In all cases, the direct method yielded lower noise and standard deviation than the indirect method based on OSEM. Overall, the proposed direct reconstruction offered a better bias-variance tradeoff than the indirect method applied to either OSEM and OSL-MAP. Direct parametric reconstruction as applied to in vivo dynamic PET MPI data is therefore a promising method for producing MBF maps with lower variance.
Direct parametric reconstruction in dynamic PET myocardial perfusion imaging: in-vivo studies
Petibon, Yoann; Rakvongthai, Yothin; Fakhri, Georges El; Ouyang, Jinsong
2017-01-01
Dynamic PET myocardial perfusion imaging (MPI) used in conjunction with tracer kinetic modeling enables the quantification of absolute myocardial blood flow (MBF). However, MBF maps computed using the traditional indirect method (i.e. post-reconstruction voxel-wise fitting of kinetic model to PET time-activity-curves -TACs) suffer from poor signal-to-noise ratio (SNR). Direct reconstruction of kinetic parameters from raw PET projection data has been shown to offer parametric images with higher SNR compared to the indirect method. The aim of this study was to extend and evaluate the performance of a direct parametric reconstruction method using in-vivo dynamic PET MPI data for the purpose of quantifying MBF. Dynamic PET MPI studies were performed on two healthy pigs using a Siemens Biograph mMR scanner. List-mode PET data for each animal were acquired following a bolus injection of ~7-8 mCi of 18F-flurpiridaz, a myocardial perfusion agent. Fully-3D dynamic PET sinograms were obtained by sorting the coincidence events into 16 temporal frames covering ~5 min after radiotracer administration. Additionally, eight independent noise realizations of both scans - each containing 1/8th of the total number of events - were generated from the original list-mode data. Dynamic sinograms were then used to compute parametric maps using the conventional indirect method and the proposed direct method. For both methods, a one-tissue compartment model accounting for spillover from the left and right ventricle blood-pools was used to describe the kinetics of 18F-flurpiridaz. An image-derived arterial input function obtained from a TAC taken in the left ventricle cavity was used for tracer kinetic analysis. For the indirect method, frame-by-frame images were estimated using two fully-3D reconstruction techniques: the standard Ordered Subset Expectation Maximization (OSEM) reconstruction algorithm on one side, and the One-Step Late Maximum a Posteriori (OSL-MAP) algorithm on the other side, which incorporates a quadratic penalty function. The parametric images were then calculated using voxel-wise weighted least-square fitting of the reconstructed myocardial PET TACs. For the direct method, parametric images were estimated directly from the dynamic PET sinograms using a maximum a posteriori (MAP) parametric reconstruction algorithm which optimizes an objective function comprised of the Poisson log-likelihood term, the kinetic model and a quadratic penalty function. Maximization of the objective function with respect to each set of parameters was achieved using a preconditioned conjugate gradient algorithm with a specifically developed pre-conditioner. The performance of the direct method was evaluated by comparing voxel- and segment-wise estimates of K1, the tracer transport rate (mL.min−1.mL−1), to those obtained using the indirect method applied to both OSEM and OSL-MAP dynamic reconstructions. The proposed direct reconstruction method produced K1 maps with visibly lower noise than the indirect method based on OSEM and OSL-MAP reconstructions. At normal count levels, the direct method was shown to outperform the indirect method based on OSL-MAP in the sense that at matched level of bias, reduced regional noise levels were obtained. At lower count levels, the direct method produced K1 estimates with significantly lower standard deviation across noise realizations than the indirect method based on OSL-MAP at matched bias level. In all cases, the direct method yielded lower noise and standard deviation than the indirect method based on OSEM. Overall, the proposed direct reconstruction offered a better bias-variance tradeoff than the indirect method applied to either OSEM and OSL-MAP. Direct parametric reconstruction as applied to in-vivo dynamic PET MPI data is therefore a promising method for producing MBF maps with lower variance. PMID:28379843
Direct parametric reconstruction in dynamic PET myocardial perfusion imaging: in vivo studies
NASA Astrophysics Data System (ADS)
Petibon, Yoann; Rakvongthai, Yothin; El Fakhri, Georges; Ouyang, Jinsong
2017-05-01
Dynamic PET myocardial perfusion imaging (MPI) used in conjunction with tracer kinetic modeling enables the quantification of absolute myocardial blood flow (MBF). However, MBF maps computed using the traditional indirect method (i.e. post-reconstruction voxel-wise fitting of kinetic model to PET time-activity-curves-TACs) suffer from poor signal-to-noise ratio (SNR). Direct reconstruction of kinetic parameters from raw PET projection data has been shown to offer parametric images with higher SNR compared to the indirect method. The aim of this study was to extend and evaluate the performance of a direct parametric reconstruction method using in vivo dynamic PET MPI data for the purpose of quantifying MBF. Dynamic PET MPI studies were performed on two healthy pigs using a Siemens Biograph mMR scanner. List-mode PET data for each animal were acquired following a bolus injection of ~7-8 mCi of 18F-flurpiridaz, a myocardial perfusion agent. Fully-3D dynamic PET sinograms were obtained by sorting the coincidence events into 16 temporal frames covering ~5 min after radiotracer administration. Additionally, eight independent noise realizations of both scans—each containing 1/8th of the total number of events—were generated from the original list-mode data. Dynamic sinograms were then used to compute parametric maps using the conventional indirect method and the proposed direct method. For both methods, a one-tissue compartment model accounting for spillover from the left and right ventricle blood-pools was used to describe the kinetics of 18F-flurpiridaz. An image-derived arterial input function obtained from a TAC taken in the left ventricle cavity was used for tracer kinetic analysis. For the indirect method, frame-by-frame images were estimated using two fully-3D reconstruction techniques: the standard ordered subset expectation maximization (OSEM) reconstruction algorithm on one side, and the one-step late maximum a posteriori (OSL-MAP) algorithm on the other side, which incorporates a quadratic penalty function. The parametric images were then calculated using voxel-wise weighted least-square fitting of the reconstructed myocardial PET TACs. For the direct method, parametric images were estimated directly from the dynamic PET sinograms using a maximum a posteriori (MAP) parametric reconstruction algorithm which optimizes an objective function comprised of the Poisson log-likelihood term, the kinetic model and a quadratic penalty function. Maximization of the objective function with respect to each set of parameters was achieved using a preconditioned conjugate gradient algorithm with a specifically developed pre-conditioner. The performance of the direct method was evaluated by comparing voxel- and segment-wise estimates of {{K}1} , the tracer transport rate (ml · min-1 · ml-1), to those obtained using the indirect method applied to both OSEM and OSL-MAP dynamic reconstructions. The proposed direct reconstruction method produced {{K}1} maps with visibly lower noise than the indirect method based on OSEM and OSL-MAP reconstructions. At normal count levels, the direct method was shown to outperform the indirect method based on OSL-MAP in the sense that at matched level of bias, reduced regional noise levels were obtained. At lower count levels, the direct method produced {{K}1} estimates with significantly lower standard deviation across noise realizations than the indirect method based on OSL-MAP at matched bias level. In all cases, the direct method yielded lower noise and standard deviation than the indirect method based on OSEM. Overall, the proposed direct reconstruction offered a better bias-variance tradeoff than the indirect method applied to either OSEM and OSL-MAP. Direct parametric reconstruction as applied to in vivo dynamic PET MPI data is therefore a promising method for producing MBF maps with lower variance.
A climate analysis using CORDEX simulations in a cooperation framework: the case of Paraguay
NASA Astrophysics Data System (ADS)
Mercogliano, Paola; Bucchignani, Edoardo; Ciervo, Fabio; Montesarchio, Myriam; Zollo, Alessandra Lucia; Villani, Veronica; Barbato, Giuliana; Vendemia, Rosalba; Polato, Raul; Baez, Julian; Pasten, Max
2017-04-01
In recent years, changes in climate have entailed variations in surface temperature and precipitation patterns in various countries of the South America, among which Paraguay. Climate change-attributed effects on weather impacts, such as river and urban floods, droughts and heat waves could severely affect the actual conditions of the country. In fact, Paraguay exhibits significant vulnerabilities to climate changes, especially because of its dependence on commodities production (e.g. agriculture, livestock, etc.) and its infrastructural and logistic asset not yet fully formed. In this context, climate change analysis can be an important technical support for practitioners to assist - under uncertainty - national/regional planning, financial resources managing and development (e.g. land-use practices, population growth, economic and community behavior, health, etc.). Moreover, actions in adaptation, disaster risk reduction (DRR), social protection and impacts mitigation may involve high costs if not properly contextualized. The assessment of 21st century climate change and development of whatever response strategies requires climate scenarios at high resolution, including an accurate evaluation of projection uncertainties (i.e. robustness of the analysis). This should ensure adequate insights into the potential impacts of climate change and allow practitioners, usually ill equipped to consider uncertain climate outputs into a broader context (e.g. planning, designing, managing), to make appropriate choices. In the framework of CORDEX initiative, Paraguay is included into the SOUTH-AMERICA-CORDEX one. Three climate simulations over this area are available at the spatial resolution of 0.44° (about 50km), obtained with RCM SMHI-RCA4 (forced by GCMs ICHEC-EC-EARTH and MPI-M-MPI-ESM-LR) and RCM MPI-CSC-REMO2009 (forced by MPI-M-MPI-ESM-LR). Simulations over the 21st century have been performed according with IPCC RCP2.6, RCP4.5 and RCP8.5 scenarios. The plausibility of the acquired climate simulations has been determined by comparison with different observational datasets over the baseline period. Three future periods have been selected for the analysis: 2011-2040, 2041-2070 and 2071-2100. The analysis is carried out in order to address the mean changes in seasonal mean temperature and total precipitation, and of some indicators suitable to quantify the impact of climate extreme events. The analysis is performed in the framework of the Chake Ou project "Strengthening of institutional and community preparedness and coordination capacities for disaster risk reduction in Paraguay" funded by the European Commission's Humanitarian Aid and Civil Protection Department (ECHO), in the context of the Disaster Preparedness Action Plan (DIPECHO) (code ECHO/-SM/BUD/2015/91028). The partners of the project are COOPI (a humanitarian, no-confessional and independent organization that works to support civil, economic and social development of populations struck by emergencies (disasters and conflicts), PLAN International (a child-centered community development organization) and CMCC Foundation (Euro-Mediterranean Center on Climate Change). The consortium works in close collaboration with the local institutions such as the Secretaria de Emergencia Nacional (SEN) and the Dirección de Meteorología e Hidrología (DMH - DINAC).
Local search to improve coordinate-based task mapping
Balzuweit, Evan; Bunde, David P.; Leung, Vitus J.; ...
2015-10-31
We present a local search strategy to improve the coordinate-based mapping of a parallel job’s tasks to the MPI ranks of its parallel allocation in order to reduce network congestion and the job’s communication time. The goal is to reduce the number of network hops between communicating pairs of ranks. Our target is applications with a nearest-neighbor stencil communication pattern running on mesh systems with non-contiguous processor allocation, such as Cray XE and XK Systems. Utilizing the miniGhost mini-app, which models the shock physics application CTH, we demonstrate that our strategy reduces application running time while also reducing the runtimemore » variability. Furthermore, we further show that mapping quality can vary based on the selected allocation algorithm, even between allocation algorithms of similar apparent quality.« less
Determination of the optimal atrioventricular interval in sick sinus syndrome during DDD pacing.
Kato, Masaya; Dote, Keigo; Sasaki, Shota; Goto, Kenji; Takemoto, Hiroaki; Habara, Seiji; Hasegawa, Daiji; Matsuda, Osamu
2005-09-01
Although the AAI pacing mode has been shown to be electromechanically superior to the DDD pacing mode in sick sinus syndrome (SSS), there is evidence suggesting that during AAI pacing the presence of natural ventricular activation pattern is not enough for hemodynamic benefit to occur. Myocardial performance index (MPI) is a simply measurable Doppler-derived index of combined systolic and diastolic myocardial performance. The aim of this study was to investigate whether AAI pacing mode is electromechanically superior to the DDD mode in patients with SSS by using Doppler-derived MPI. Thirty-nine SSS patients with dual-chamber pacing devices were evaluated by using Doppler echocardiography in AAI mode and DDD mode. The optimal atrioventricular (AV) interval in DDD mode was determined and atrial stimulus-R interval was measured in AAI mode. The ratio of the atrial stimulus-R interval to the optimal AV interval was defined as relative AV interval (rAVI) and the ratio of MPI in AAI mode to that in DDD mode was defined as relative MPI (rMPI). The rMPI was significantly correlated with atrial stimulus-R interval and rAVI (r = 0.57, P = 0.0002, and r = 0.67, P < 0.0001, respectively). A cutoff point of 1.73 for rAVI provided optimum sensitivity and specificity for rMPI >1 based on the receiver operator curves. Even though the intrinsic AV conduction is moderately prolonged, some SSS patients with dual-chamber pacing devices benefit from the ventricular pacing with optimal AV interval. MPI is useful to determine the optimal pacing mode in acute experiment.
Aida, Kazuo; Sugie, Toshihiko
2011-12-12
We propose a method of testing transmission fiber lines and distributed amplifiers. Multipath interference (MPI) is detected as a beat spectrum between a multipath signal and a direct signal using a synthesized chirped test signal with lightwave frequencies of f(1) and f(2) periodically emitted from a distributed feedback laser diode (DFB-LD). This chirped test pulse is generated using a directly modulated DFB-LD with a drive signal calculated using a digital signal processing technique (DSP). A receiver consisting of a photodiode and an electrical spectrum analyzer (ESA) detects a baseband power spectrum peak appearing at the frequency of the test signal frequency deviation (f(1)-f(2)) as a beat spectrum of self-heterodyne detection. Multipath interference is converted from the spectrum peak power. This method improved the minimum detectable MPI to as low as -78 dB. We discuss the detailed design and performance of the proposed test method, including a DFB-LD drive signal calculation algorithm with DSP for synthesis of the chirped test signal and experiments on single-mode fibers with discrete reflections. © 2011 Optical Society of America
NASA Astrophysics Data System (ADS)
Bonelli, Francesco; Tuttafesta, Michele; Colonna, Gianpiero; Cutrone, Luigi; Pascazio, Giuseppe
2017-10-01
This paper describes the most advanced results obtained in the context of fluid dynamic simulations of high-enthalpy flows using detailed state-to-state air kinetics. Thermochemical non-equilibrium, typical of supersonic and hypersonic flows, was modeled by using both the accurate state-to-state approach and the multi-temperature model proposed by Park. The accuracy of the two thermochemical non-equilibrium models was assessed by comparing the results with experimental findings, showing better predictions provided by the state-to-state approach. To overcome the huge computational cost of the state-to-state model, a multiple-nodes GPU implementation, based on an MPI-CUDA approach, was employed and a comprehensive code performance analysis is presented. Both the pure MPI-CPU and the MPI-CUDA implementations exhibit excellent scalability performance. GPUs outperform CPUs computing especially when the state-to-state approach is employed, showing speed-ups, of the single GPU with respect to the single-core CPU, larger than 100 in both the case of one MPI process and multiple MPI process.
Kao, Ling-Jing; Chiu, Shu-Yu; Ko, Hsien-Tang
2014-01-01
The purpose of this study is to evaluate the training institution performance and to improve the management of the Manpower Training Project (MTP) administered by the Semiconductor Institute in Taiwan. Much literature assesses the efficiency of an internal training program initiated by a firm, but only little literature studies the efficiency of an external training program led by government. In the study, a hybrid solution of ICA-DEA and ICA-MPI is developed for measuring the efficiency and the productivity growth of each training institution over the period. The technical efficiency change, the technological change, pure technical efficiency change, scale efficiency change, and the total factor productivity change were evaluated according to five inputs and two outputs. According to the results of the study, the training institutions can be classified by their efficiency successfully and the guidelines for the optimal level of input resources can be obtained for each inefficient training institution. The Semiconductor Institute in Taiwan can allocate budget more appropriately and establish withdrawal mechanisms for inefficient training institutions.
Kao, Ling-Jing; Chiu, Shu-Yu; Ko, Hsien-Tang
2014-01-01
The purpose of this study is to evaluate the training institution performance and to improve the management of the Manpower Training Project (MTP) administered by the Semiconductor Institute in Taiwan. Much literature assesses the efficiency of an internal training program initiated by a firm, but only little literature studies the efficiency of an external training program led by government. In the study, a hybrid solution of ICA-DEA and ICA-MPI is developed for measuring the efficiency and the productivity growth of each training institution over the period. The technical efficiency change, the technological change, pure technical efficiency change, scale efficiency change, and the total factor productivity change were evaluated according to five inputs and two outputs. According to the results of the study, the training institutions can be classified by their efficiency successfully and the guidelines for the optimal level of input resources can be obtained for each inefficient training institution. The Semiconductor Institute in Taiwan can allocate budget more appropriately and establish withdrawal mechanisms for inefficient training institutions. PMID:24977192
A real-time MPEG software decoder using a portable message-passing library
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kwong, Man Kam; Tang, P.T. Peter; Lin, Biquan
1995-12-31
We present a real-time MPEG software decoder that uses message-passing libraries such as MPL, p4 and MPI. The parallel MPEG decoder currently runs on the IBM SP system but can be easil ported to other parallel machines. This paper discusses our parallel MPEG decoding algorithm as well as the parallel programming environment under which it uses. Several technical issues are discussed, including balancing of decoding speed, memory limitation, 1/0 capacities, and optimization of MPEG decoding components. This project shows that a real-time portable software MPEG decoder is feasible in a general-purpose parallel machine.
Challenges at Petascale for Pseudo-Spectral Methods on Spheres (A Last Hurrah?)
NASA Technical Reports Server (NTRS)
Clune, Thomas
2011-01-01
Conclusions: a) Proper software abstractions should enable rapid-exploration of platform-specific optimizations/ tradeoffs. b) Pseudo-spectra! methods are marginally viable for at least some classes of petascaie problems. i.e., GPU based machine with good bisection would be best. c) Scalability at exascale is possible, but the necessary resolution will make algorithm prohibitively expensive. Efficient implementations of realistic global transposes are mtricate and tedious in MPI. PS at petascaie requires exploration of a variety of strategies for spreading local and remote communic3tions. PGAS allows far simpler implementation and thus rapid exploration of variants.
Starmans, Lucas W. E.; Burdinski, Dirk; Haex, Nicole P. M.; Moonen, Rik P. M.; Strijkers, Gustav J.; Nicolay, Klaas; Grüll, Holger
2013-01-01
Background Iron oxide nanoparticles (IONs) are a promising nanoplatform for contrast-enhanced MRI. Recently, magnetic particle imaging (MPI) was introduced as a new imaging modality, which is able to directly visualize magnetic particles and could serve as a more sensitive and quantitative alternative to MRI. However, MPI requires magnetic particles with specific magnetic properties for optimal use. Current commercially available iron oxide formulations perform suboptimal in MPI, which is triggering research into optimized synthesis strategies. Most synthesis procedures aim at size control of iron oxide nanoparticles rather than control over the magnetic properties. In this study, we report on the synthesis, characterization and application of a novel ION platform for sensitive MPI and MRI. Methods and Results IONs were synthesized using a thermal-decomposition method and subsequently phase-transferred by encapsulation into lipidic micelles (ION-Micelles). Next, the material and magnetic properties of the ION-Micelles were analyzed. Most notably, vibrating sample magnetometry measurements showed that the effective magnetic core size of the IONs is 16 nm. In addition, magnetic particle spectrometry (MPS) measurements were performed. MPS is essentially zero-dimensional MPI and therefore allows to probe the potential of iron oxide formulations for MPI. ION-Micelles induced up to 200 times higher signal in MPS measurements than commercially available iron oxide formulations (Endorem, Resovist and Sinerem) and thus likely allow for significantly more sensitive MPI. In addition, the potential of the ION-Micelle platform for molecular MPI and MRI was showcased by MPS and MRI measurements of fibrin-binding peptide functionalized ION-Micelles (FibPep-ION-Micelles) bound to blood clots. Conclusions The presented data underlines the potential of the ION-Micelle nanoplatform for sensitive (molecular) MPI and warrants further investigation of the FibPep-ION-Micelle platform for in vivo, non-invasive imaging of fibrin in preclinical disease models of thrombus-related pathologies and atherosclerosis. PMID:23437371
Starmans, Lucas W E; Burdinski, Dirk; Haex, Nicole P M; Moonen, Rik P M; Strijkers, Gustav J; Nicolay, Klaas; Grüll, Holger
2013-01-01
Iron oxide nanoparticles (IONs) are a promising nanoplatform for contrast-enhanced MRI. Recently, magnetic particle imaging (MPI) was introduced as a new imaging modality, which is able to directly visualize magnetic particles and could serve as a more sensitive and quantitative alternative to MRI. However, MPI requires magnetic particles with specific magnetic properties for optimal use. Current commercially available iron oxide formulations perform suboptimal in MPI, which is triggering research into optimized synthesis strategies. Most synthesis procedures aim at size control of iron oxide nanoparticles rather than control over the magnetic properties. In this study, we report on the synthesis, characterization and application of a novel ION platform for sensitive MPI and MRI. IONs were synthesized using a thermal-decomposition method and subsequently phase-transferred by encapsulation into lipidic micelles (ION-Micelles). Next, the material and magnetic properties of the ION-Micelles were analyzed. Most notably, vibrating sample magnetometry measurements showed that the effective magnetic core size of the IONs is 16 nm. In addition, magnetic particle spectrometry (MPS) measurements were performed. MPS is essentially zero-dimensional MPI and therefore allows to probe the potential of iron oxide formulations for MPI. ION-Micelles induced up to 200 times higher signal in MPS measurements than commercially available iron oxide formulations (Endorem, Resovist and Sinerem) and thus likely allow for significantly more sensitive MPI. In addition, the potential of the ION-Micelle platform for molecular MPI and MRI was showcased by MPS and MRI measurements of fibrin-binding peptide functionalized ION-Micelles (FibPep-ION-Micelles) bound to blood clots. The presented data underlines the potential of the ION-Micelle nanoplatform for sensitive (molecular) MPI and warrants further investigation of the FibPep-ION-Micelle platform for in vivo, non-invasive imaging of fibrin in preclinical disease models of thrombus-related pathologies and atherosclerosis.
NASA Astrophysics Data System (ADS)
Clay, M. P.; Buaria, D.; Gotoh, T.; Yeung, P. K.
2017-10-01
A new dual-communicator algorithm with very favorable performance characteristics has been developed for direct numerical simulation (DNS) of turbulent mixing of a passive scalar governed by an advection-diffusion equation. We focus on the regime of high Schmidt number (S c), where because of low molecular diffusivity the grid-resolution requirements for the scalar field are stricter than those for the velocity field by a factor √{ S c }. Computational throughput is improved by simulating the velocity field on a coarse grid of Nv3 points with a Fourier pseudo-spectral (FPS) method, while the passive scalar is simulated on a fine grid of Nθ3 points with a combined compact finite difference (CCD) scheme which computes first and second derivatives at eighth-order accuracy. A static three-dimensional domain decomposition and a parallel solution algorithm for the CCD scheme are used to avoid the heavy communication cost of memory transposes. A kernel is used to evaluate several approaches to optimize the performance of the CCD routines, which account for 60% of the overall simulation cost. On the petascale supercomputer Blue Waters at the University of Illinois, Urbana-Champaign, scalability is improved substantially with a hybrid MPI-OpenMP approach in which a dedicated thread per NUMA domain overlaps communication calls with computational tasks performed by a separate team of threads spawned using OpenMP nested parallelism. At a target production problem size of 81923 (0.5 trillion) grid points on 262,144 cores, CCD timings are reduced by 34% compared to a pure-MPI implementation. Timings for 163843 (4 trillion) grid points on 524,288 cores encouragingly maintain scalability greater than 90%, although the wall clock time is too high for production runs at this size. Performance monitoring with CrayPat for problem sizes up to 40963 shows that the CCD routines can achieve nearly 6% of the peak flop rate. The new DNS code is built upon two existing FPS and CCD codes. With the grid ratio Nθ /Nv = 8, the disparity in the computational requirements for the velocity and scalar problems is addressed by splitting the global communicator MPI_COMM_WORLD into disjoint communicators for the velocity and scalar fields, respectively. Inter-communicator transfer of the velocity field from the velocity communicator to the scalar communicator is handled with discrete send and non-blocking receive calls, which are overlapped with other operations on the scalar communicator. For production simulations at Nθ = 8192 and Nv = 1024 on 262,144 cores for the scalar field, the DNS code achieves 94% strong scaling relative to 65,536 cores and 92% weak scaling relative to Nθ = 1024 and Nv = 128 on 512 cores.
''Towards a High-Performance and Robust Implementation of MPI-IO on Top of GPFS''
DOE Office of Scientific and Technical Information (OSTI.GOV)
Prost, J.P.; Tremann, R.; Blackwore, R.
2000-01-11
MPI-IO/GPFS is a prototype implementation of the I/O chapter of the Message Passing Interface (MPI) 2 standard. It uses the IBM General Parallel File System (GPFS), with prototyped extensions, as the underlying file system. this paper describes the features of this prototype which support its high performance and robustness. The use of hints at the file system level and at the MPI-IO level allows tailoring the use of the file system to the application needs. Error handling in collective operations provides robust error reporting and deadlock prevention in case of returning errors.
Impact of Forecast and Model Error Correlations In 4dvar Data Assimilation
NASA Astrophysics Data System (ADS)
Zupanski, M.; Zupanski, D.; Vukicevic, T.; Greenwald, T.; Eis, K.; Vonder Haar, T.
A weak-constraint 4DVAR data assimilation system has been developed at Cooper- ative Institute for Research in the Atmosphere (CIRA), Colorado State University. It is based on the NCEP's ETA 4DVAR system, and it is fully parallel (MPI coding). The CIRA's 4DVAR system is aimed for satellite data assimilation research, with cur- rent focus on assimilation of cloudy radiances and microwave satellite measurements. Most important improvement over the previous 4DVAR system is a degree of gener- ality introduced into the new algorithm, namely for applications with different NWP models (e.g., RAMS, WRF, ETA, etc.), and for the choice of control variable. In cur- rent applications, the non-hydrostatic RAMS model and its adjoint are used, including all microphysical processess. The control variable includes potential temperature, ve- locity potential and stream function, vertical velocity, and seven mixing ratios with respect to all water phases. Since the statistics of the microphysical components of the control variable is not well known, a special attention will be paid to the impact of the forecast and model (prior) error correlations on the 4DVAR analysis. In particular, the sensitivity of the analysis with respect to decorrelation length will be examined. The prior error covariances are modelled using the compactly-supported, space-limited correlations developed at NASA DAO.
A novel artificial neural network method for biomedical prediction based on matrix pseudo-inversion.
Cai, Binghuang; Jiang, Xia
2014-04-01
Biomedical prediction based on clinical and genome-wide data has become increasingly important in disease diagnosis and classification. To solve the prediction problem in an effective manner for the improvement of clinical care, we develop a novel Artificial Neural Network (ANN) method based on Matrix Pseudo-Inversion (MPI) for use in biomedical applications. The MPI-ANN is constructed as a three-layer (i.e., input, hidden, and output layers) feed-forward neural network, and the weights connecting the hidden and output layers are directly determined based on MPI without a lengthy learning iteration. The LASSO (Least Absolute Shrinkage and Selection Operator) method is also presented for comparative purposes. Single Nucleotide Polymorphism (SNP) simulated data and real breast cancer data are employed to validate the performance of the MPI-ANN method via 5-fold cross validation. Experimental results demonstrate the efficacy of the developed MPI-ANN for disease classification and prediction, in view of the significantly superior accuracy (i.e., the rate of correct predictions), as compared with LASSO. The results based on the real breast cancer data also show that the MPI-ANN has better performance than other machine learning methods (including support vector machine (SVM), logistic regression (LR), and an iterative ANN). In addition, experiments demonstrate that our MPI-ANN could be used for bio-marker selection as well. Copyright © 2013 Elsevier Inc. All rights reserved.
Iskandar, Aline; Limone, Brendan; Parker, Matthew W; Perugini, Andrew; Kim, Hyejin; Jones, Charles; Calamari, Brian; Coleman, Craig I; Heller, Gary V
2013-02-01
It remains controversial whether the diagnostic accuracy of single-photon emission computed tomography myocardial perfusion imaging (SPECT MPI) is different in men as compared to women. We performed a meta-analysis to investigate gender differences of SPECT MPI for the diagnosis of CAD (≥50% stenosis). Two investigators independently performed a systematic review of the MEDLINE and EMBASE databases from inception through January 2012 for English-language studies determining the diagnostic accuracy of SPECT MPI. We included prospective studies that compared SPECT MPI with conventional coronary angiography which provided sufficient data to calculate gender-specific true and false positives and negatives. Data from studies evaluating <20 patients of one gender were excluded. Bivariate meta-analysis was used to create summary receiver operating curves. Twenty-six studies met inclusion criteria, representing 1,148 women and 1,142 men. Bivariate meta-analysis yielded a mean sensitivity and specificity of 84.2% (95% confidence interval [CI] 78.7%-88.6%) and 78.7% (CI 70.0%-85.3%) for SPECT MPI in women and 89.1% (CI 84.0%-92.7%) and 71.2% (CI 60.8%-79.8%) for SPECT MPI in men. There was no significant difference in the sensitivity (P = .15) or specificity (P = .23) between male and female subjects. In a bivariate meta-analysis of the available literature, the diagnostic accuracy of SPECT MPI is similar for both men and women.
Direct comparison of rest and adenosine stress myocardial perfusion CT with rest and stress SPECT
Okada, David R.; Ghoshhajra, Brian B.; Blankstein, Ron; Rocha-Filho, Jose A.; Shturman, Leonid D.; Rogers, Ian S.; Bezerra, Hiram G.; Sarwar, Ammar; Gewirtz, Henry; Hoffmann, Udo; Mamuya, Wilfred S.; Brady, Thomas J.; Cury, Ricardo C.
2010-01-01
Introduction We have recently described a technique for assessing myocardial perfusion using adenosine-mediated stress imaging (CTP) with dual source computed tomography. SPECT myocardial perfusion imaging (SPECT-MPI) is a widely utilized and extensively validated method for assessing myocardial perfusion. The aim of this study was to determine the level of agreement between CTP and SPECT-MPI at rest and under stress on a per-segment, per-vessel, and per-patient basis. Methods Forty-seven consecutive patients underwent CTP and SPECT-MPI. Perfusion images were interpreted using the 17 segment AHA model and were scored on a 0 (normal) to 3 (abnormal) scale. Summed rest and stress scores were calculated for each vascular territory and patient by adding corresponding segmental scores. Results On a per-segment basis (n = 799), CTP and SPECT-MPI demonstrated excellent correlation: Goodman-Kruskall γ = .59 (P < .0001) for stress and .75 (P < .0001) for rest. On a per-vessel basis (n = 141), CTP and SPECT-MPI summed scores demonstrated good correlation: Pearson r = .56 (P < .0001) for stress and .66 (P < .0001) for rest. On a per-patient basis (n = 47), CTP and SPECT-MPI demonstrated good correlation: Pearson r = .60 (P < .0001) for stress and .76 (P < .0001) for rest. Conclusions CTP compares favorably with SPECT-MPI for detection, extent, and severity of myocardial perfusion defects at rest and stress. PMID:19936863
NASA Astrophysics Data System (ADS)
Guilyardi, E.
2003-04-01
The European Union's PRISM infrastructure project (PRogram for Integrated earth System Modelling) aims at designing a flexible environment to easily assemble and run Earth System Models (http://prism.enes.org). Europe's widely distributed modelling expertise is both a strength and a challenge. Recognizing this, the PRISM project aims at developing an efficient shared modelling software infrastructure for climate scientists, providing them with an opportunity for greater focus on scientific issues, including the necessary scientific diversity (models and approaches). The proposed PRISM system includes 1) the use - or definition - and promotion of scientific and technical standards to increase component modularity, 2) an end-to-end software environment (coupler, user interface, diagnostics) to launch, monitor and analyze complex Earth System Models built around the existing and future community models, 3) testing and quality standards to ensure HPC performance on a variety of platforms and 4) community wide inputs and requirements capture in all stages of system specifications and design through user/developers meetings, workshops and thematic schools. This science driven project, led by 22 institutes* and started December 1st 2001, benefits from a unique gathering of scientific and technical expertise. More than 30 models (both global and regional) have expressed interest to be part of the PRISM system and 6 types of components have been identified: atmosphere, atmosphere chemistry, land surface, ocean, sea ice and ocean biochemistry. Progress and overall architecture design will be presented. * MPI-Met (Coordinator), KNMI (co-coordinator), MPI-M&D, Met Office, University of Reading, IPSL, Meteo-France, CERFACS, DMI, SMHI, NERSC, ETH Zurich, INGV, MPI-BGC, PIK, ECMWF, UCL-ASTR, NEC, FECIT, SGI, SUN, CCRLE
Shelat, Vishal G; Ahmed, Saleem; Chia, Clement L K; Cheah, Yee Lee
2015-02-01
Application of minimal access surgery in acute care surgery is limited due to various reasons. Laparoscopic omental patch repair (LOPR) for perforated peptic ulcer (PPU) surgery is safe and feasible but not widely implemented. We report our early experience of LOPR with emphasis on strict selection criteria. This is a descriptive study of all patients operated on for PPU at academic university-affiliated institutes from December 2010 to February 2012. All the patients who were operated on for LOPR were included as the study population and their records were studied. Perioperative outcomes, Boey score, Mannheim Peritonitis Index (MPI), and physiologic and operative severity scores for enumeration of mortality and morbidity (POSSUM) scores were calculated. All the data were tabulated in a Microsoft Excel spreadsheet and analyzed using Stata Version 8.x. (StataCorp, College Station, TX, USA). Fourteen patients had LOPR out of a total of 45 patients operated for the PPU. Mean age was 46 years (range 22-87 years). Twelve patients (86%) had a Boey score of 0 and all patients had MPI < 21 (mean MPI = 14). The predicted POSSUM morbidity and mortality were 36% and 7%, respectively. Mean ulcer size was 5 mm (range 2-10 mm), mean operating time was 100 minutes (range 70-123 minutes) and mean length of hospital stay was 4 days (range 3-6 days). There was no morbidity or mortality pertaining to LOPR. LOPR should be offered by acute care surgical teams when local expertise is available. This can optimize patient outcomes when strict selection criteria are applied.
Pilotto, Alberto; Polidori, Maria Cristina; Veronese, Nicola; Panza, Francesco; Arboretti Giancristofaro, Rosa; Pilotto, Andrea; Daragjati, Julia; Carrozzo, Eleonora; Prete, Camilla; Gallina, Pietro; Padovani, Alessandro; Maggi, Stefania
2018-02-01
To evaluate whether treatment with antidementia drugs is associated with reduced mortality in older patients with different mortality risk at baseline. Retrospective. Community-dwelling. A total of 6818 older people who underwent a Standardized Multidimensional Assessment Schedule for Adults and Aged Persons (SVaMA) evaluation to determine accessibility to homecare services or nursing home admission from 2005 to 2013 in the Padova Health District, Italy were included. Mortality risk at baseline was calculated by the Multidimensional Prognostic Index (MPI), based on information collected with the SVaMA. Participants were categorized to have mild (MPI-SVaMA-1), moderate (MPI-SVaMA-2), and high (MPI-SVaMA-3) mortality risk. Propensity score-adjusted hazard ratios (HR) of 2-year mortality were calculated according to antidementia drug treatment. Patients treated with antidementia drugs had a significant lower risk of death than untreated patients (HR 0.82; 95% confidence interval [CI] 0.73-0.92 and 0.56; 95% CI 0.49-0.65 for patients treated less than 2 years and more than 2 years treatment, respectively). After dividing patients according to their MPI-SVaMA grade, antidementia treatment was significantly associated with reduced mortality in the MPI-SVaMA-1 mild (HR 0.71; 95% CI 0.54-0.92) and MPI-SVaMA-2 moderate risk (HR 0.61; 95% CI 0.40-0.91, matched sample), but not in the MPI-SVaMA-3 high risk of death. This large community-dwelling patient study suggests that antidementia drugs might contribute to increased survival in older adults with dementia with lower mortality risk. Copyright © 2017 AMDA – The Society for Post-Acute and Long-Term Care Medicine. Published by Elsevier Inc. All rights reserved.
Cardiovascular outcomes after pharmacologic stress myocardial perfusion imaging.
Lee, Douglas S; Husain, Mansoor; Wang, Xuesong; Austin, Peter C; Iwanochko, Robert M
2016-04-01
While pharmacologic stress single photon emission computed tomography myocardial perfusion imaging (SPECT-MPI) is used for noninvasive evaluation of patients who are unable to perform treadmill exercise, its impact on net reclassification improvement (NRI) of prognosis is unknown. We evaluated the prognostic value of pharmacologic stress MPI for prediction of cardiovascular death or non-fatal myocardial infarction (MI) within 1 year at a single-center, university-based laboratory. We examined continuous and categorical NRI of pharmacologic SPECT-MPI for prediction of outcomes beyond clinical factors alone. Six thousand two hundred forty patients (median age 66 years [IQR 56-74], 3466 men) were studied and followed for 5963 person-years. SPECT-MPI variables associated with increased risk of cardiovascular death or non-fatal MI included summed stress score, stress ST-shift, and post-stress resting left ventricular ejection fraction ≤50%. Compared to a clinical model which included age, sex, cardiovascular disease, risk factors, and medications, model χ(2) (210.5 vs. 281.9, P < .001) and c-statistic (0.74 vs. 0.78, P < .001) were significantly increased by addition of SPECT-MPI predictors (summed stress score, stress ST-shift and stress resting left ventricular ejection fraction). SPECT-MPI predictors increased continuous NRI by 49.4% (P < .001), reclassifying 66.5% of patients as lower risk and 32.8% as higher risk of cardiovascular death or non-fatal MI. Addition of MPI predictors to clinical factors using risk categories, defined as <1%, 1% to 3%, and >3% annualized risk of cardiovascular death or non-fatal MI, yielded a 15.0% improvement in NRI (95% CI 7.6%-27.6%, P < .001). Pharmacologic stress MPI substantially improved net reclassification of cardiovascular death or MI risk beyond that afforded by clinical factors. Copyright © 2016 Elsevier Inc. All rights reserved.
Paz, Yehuda; Morgenstern, Rachelle; Weinberg, Richard; Chiles, Mariana; Bhatti, Navdeep; Ali, Ziad; Mohan, Sumit; Bokhari, Sabahat
2017-12-01
Cardiovascular disease is the leading cause of death in patients with end-stage renal disease (ESRD) and often goes undetected. Abnormal coronary flow reserve (CFR), which predicts increased risk of cardiac death, may be present in patients with ESRD without other evidence of coronary artery disease (CAD). We prospectively studied 131 patients who had rest and dipyridamole pharmacologic stress N 13 -ammonia positron emission tomography myocardial perfusion imaging (PET MPI) for kidney transplant evaluation. Thirty-four patients also had left heart catheterization. Abnormal PET MPI was defined as qualitative ischemia or infarct, stress electrocardiogram ischemia, or transient ischemic dilation. CFR was calculated as the ratio of stress to rest coronary blood flow. Global CFR < 2 was defined as abnormal. Of 131 patients who had PET MPI (66% male, 55.6 ± 12.1 years), 30% (39 of 131) had abnormal PET MPI and 59% (77 of 131) had abnormal CFR. In a subset of 34 patients who had left heart catheterization (66% male, 61.0 ± 12.1 years), 68% (23 of 34) had abnormal CFR on PET MPI, and 68% (23 of 34) had ≥70% obstruction on left heart catheterization. Abnormal CFR was not significantly associated with abnormal PET MPI (p = 0.13) or obstructive CAD on left heart catheterization (p = 0.26). In conclusion, in the first prospective study of PET MPI in patients with ESRD, abnormal CFR is highly prevalent and is independent of abnormal findings on PET MPI or obstructive CAD on left heart catheterization. Copyright © 2017 Elsevier Inc. All rights reserved.
Dobutamine stress myocardial perfusion imaging: 8-year outcomes in patients with diabetes mellitus.
Boiten, Hendrik J; van Domburg, Ron T; Valkema, Roelf; Zijlstra, Felix; Schinkel, Arend F L
2016-08-01
Many studies have examined the prognostic value of myocardial perfusion imaging (MPI) using single-photon emission computed tomography (SPECT) for the prediction of short- to medium-term outcomes. However, the long-term prognostic value of MPI in patients with diabetes mellitus remains unclear. Therefore, this study assessed the long-term prognostic value of MPI in a high-risk cohort of patients with diabetes mellitus. A high-risk cohort of 207 patients with diabetes mellitus who were unable to undergo exercise testing underwent dobutamine stress MPI. Follow-up was successful in 206 patients; 12 patients were excluded due to early revascularization. The current data are based on the remaining 194 patients. Follow-up end points were all-cause mortality, cardiac mortality, and nonfatal myocardial infarction. The Kaplan-Meier survival curves were constructed, and univariable and multivariable analyses were performed to identify predictors of long-term outcome. During a mean follow-up of 8.1 ± 5.9 years, 134 (69%) patients died of which 68 (35%) died due to cardiac causes. Nonfatal myocardial infarction occurred in 24 patients (12%), and late (>60 days) coronary revascularization was performed in 61 (13%) patients. Survival analysis showed that MPI provided optimal risk stratification up to 4 years after testing. After that period, the outcome was comparable in patients with normal and abnormal MPI. Multivariable analyses showed that MPI provided incremental prognostic value up to 4 years after testing. In high-risk patients with diabetes mellitus, dobutamine MPI provides incremental prognostic information in addition to clinical data for a 4-year period after testing. Published on behalf of the European Society of Cardiology. All rights reserved. © The Author 2016. For permissions please email: journals.permissions@oup.com.
Pilotto, Alberto; Addante, Filomena; D'Onofrio, Grazia; Sancarlo, Daniele; Ferrucci, Luigi
2009-01-01
The Comprehensive Geriatric Assessment (CGA) is a multidimensional, usually interdisciplinary, diagnostic process intended to determine an elderly person's medical, psychosocial, and functional capacity and problems with the objective of developing an overall plan for treatment and short- and long-term follow-up. The potential usefulness of the CGA in evaluating treatment and follow-up of older patients with gastroenterological disorders is unknown. In the paper we reported the efficacy of a Multidimensional-Prognostic Index (MPI), calculated from information collected by a standardized CGA, in predicting mortality risk in older patients hospitalized with upper gastrointestinal bleeding and liver cirrhosis. Patients underwent a CGA that included six standardized scales, i.e. Activities of Daily Living (ADL), Instrumental Activities of Daily Living (IADL), Short-Portable Mental Status Questionnaire (SPMSQ), Mini-Nutritional Assessment (MNA), Exton-Smith Score (ESS) and Comorbity Index Rating Scale (CIRS), as well as information on medication history and cohabitation, for a total of 63 items. The MPI was calculated from the integrated total scores and expressed as MPI 1=low risk, MPI 2=moderate risk and MPI 3=severe risk of mortality. Higher MPI values were significantly associated with higher short- and long-term mortality in older patients with both upper gastrointestinal bleeding and liver cirrhosis. A close agreement was found between the estimated mortality by MPI and the observed mortality. Moreover, MPI seems to have a greater discriminatory power than organ-specific prognostic indices such as Rockall and Blatchford scores (in upper gastrointestinal bleeding patients) and Child-Plugh score (in liver cirrhosis patients). All these findings support the concept that a multidimensional approach may be appropriate for the evaluation of older patients with gastroenterological disorders, like it has been reported for patients with other pathological conditions.
NASA Astrophysics Data System (ADS)
Rastogi, Richa; Srivastava, Abhishek; Khonde, Kiran; Sirasala, Kirannmayi M.; Londhe, Ashutosh; Chavhan, Hitesh
2015-07-01
This paper presents an efficient parallel 3D Kirchhoff depth migration algorithm suitable for current class of multicore architecture. The fundamental Kirchhoff depth migration algorithm exhibits inherent parallelism however, when it comes to 3D data migration, as the data size increases the resource requirement of the algorithm also increases. This challenges its practical implementation even on current generation high performance computing systems. Therefore a smart parallelization approach is essential to handle 3D data for migration. The most compute intensive part of Kirchhoff depth migration algorithm is the calculation of traveltime tables due to its resource requirements such as memory/storage and I/O. In the current research work, we target this area and develop a competent parallel algorithm for post and prestack 3D Kirchhoff depth migration, using hybrid MPI+OpenMP programming techniques. We introduce a concept of flexi-depth iterations while depth migrating data in parallel imaging space, using optimized traveltime table computations. This concept provides flexibility to the algorithm by migrating data in a number of depth iterations, which depends upon the available node memory and the size of data to be migrated during runtime. Furthermore, it minimizes the requirements of storage, I/O and inter-node communication, thus making it advantageous over the conventional parallelization approaches. The developed parallel algorithm is demonstrated and analysed on Yuva II, a PARAM series of supercomputers. Optimization, performance and scalability experiment results along with the migration outcome show the effectiveness of the parallel algorithm.
Myocardial perfusion imaging with PET
Nakazato, Ryo; Berman, Daniel S; Alexanderson, Erick; Slomka, Piotr
2013-01-01
PET-myocardial perfusion imaging (MPI) allows accurate measurement of myocardial perfusion, absolute myocardial blood flow and function at stress and rest in a single study session performed in approximately 30 min. Various PET tracers are available for MPI, and rubidium-82 or nitrogen-13-ammonia is most commonly used. In addition, a new fluorine-18-based PET-MPI tracer is currently being evaluated. Relative quantification of PET perfusion images shows very high diagnostic accuracy for detection of obstructive coronary artery disease. Dynamic myocardial blood flow analysis has demonstrated additional prognostic value beyond relative perfusion imaging. Patient radiation dose can be reduced and image quality can be improved with latest advances in PET/CT equipment. Simultaneous assessment of both anatomy and perfusion by hybrid PET/CT can result in improved diagnostic accuracy. Compared with SPECT-MPI, PET-MPI provides higher diagnostic accuracy, using lower radiation doses during a shorter examination time period for the detection of coronary artery disease. PMID:23671459
How Formal Dynamic Verification Tools Facilitate Novel Concurrency Visualizations
NASA Astrophysics Data System (ADS)
Aananthakrishnan, Sriram; Delisi, Michael; Vakkalanka, Sarvani; Vo, Anh; Gopalakrishnan, Ganesh; Kirby, Robert M.; Thakur, Rajeev
With the exploding scale of concurrency, presenting valuable pieces of information collected by formal verification tools intuitively and graphically can greatly enhance concurrent system debugging. Traditional MPI program debuggers present trace views of MPI program executions. Such views are redundant, often containing equivalent traces that permute independent MPI calls. In our ISP formal dynamic verifier for MPI programs, we present a collection of alternate views made possible by the use of formal dynamic verification. Some of ISP’s views help pinpoint errors, some facilitate discerning errors by eliminating redundancy, while others help understand the program better by displaying concurrent even orderings that must be respected by all MPI implementations, in the form of completes-before graphs. In this paper, we describe ISP’s graphical user interface (GUI) capabilities in all these areas which are currently supported by a portable Java based GUI, a Microsoft Visual Studio GUI, and an Eclipse based GUI whose development is in progress.
Multiphoton imaging with high peak power VECSELs
NASA Astrophysics Data System (ADS)
Mirkhanov, Shamil; Quarterman, Adrian H.; Swift, Samuel; Praveen, Bavishna B.; Smyth, Conor J. C.; Wilcox, Keith G.
2016-03-01
Multiphoton imaging (MMPI) has become one of thee key non-invasive light microscopy techniques. This technique allows deep tissue imaging with high resolution and less photo-damage than conventional confocal microscopy. MPI is type of laser-scanning microscopy that employs localized nonlinear excitation, so that fluorescence is excited only with is scanned focal volume. For many years, Ti: sapphire femtosecond lasers have been the leading light sources for MPI applications. However, recent developments in laser sources and new types of fluorophores indicate that longer wavelength excitation could be a good alternative for these applications. Mode-locked VECSEELs have the potential to be low cost, compact light sources for MPI systems, with the additional advantage of broad wavelength coverage through use of different semiconductor material systems. Here, we use a femtosecond fibber laser to investigate the effect average power and repetition rate has on MPI image quality, to allow us to optimize our mode-locked VVECSELs for MPI.
Memory Compression Techniques for Network Address Management in MPI
DOE Office of Scientific and Technical Information (OSTI.GOV)
Guo, Yanfei; Archer, Charles J.; Blocksome, Michael
MPI allows applications to treat processes as a logical collection of integer ranks for each MPI communicator, while internally translating these logical ranks into actual network addresses. In current MPI implementations the management and lookup of such network addresses use memory sizes that are proportional to the number of processes in each communicator. In this paper, we propose a new mechanism, called AV-Rankmap, for managing such translation. AV-Rankmap takes advantage of logical patterns in rank-address mapping that most applications naturally tend to have, and it exploits the fact that some parts of network address structures are naturally more performance criticalmore » than others. It uses this information to compress the memory used for network address management. We demonstrate that AV-Rankmap can achieve performance similar to or better than that of other MPI implementations while using significantly less memory.« less
Evaluating and extending user-level fault tolerance in MPI applications
Laguna, Ignacio; Richards, David F.; Gamblin, Todd; ...
2016-01-11
The user-level failure mitigation (ULFM) interface has been proposed to provide fault-tolerant semantics in the Message Passing Interface (MPI). Previous work presented performance evaluations of ULFM; yet questions related to its programability and applicability, especially to non-trivial, bulk synchronous applications, remain unanswered. In this article, we present our experiences on using ULFM in a case study with a large, highly scalable, bulk synchronous molecular dynamics application to shed light on the advantages and difficulties of this interface to program fault-tolerant MPI applications. We found that, although ULFM is suitable for master–worker applications, it provides few benefits for more common bulkmore » synchronous MPI applications. Furthermore, to address these limitations, we introduce a new, simpler fault-tolerant interface for complex, bulk synchronous MPI programs with better applicability and support than ULFM for application-level recovery mechanisms, such as global rollback.« less
Accelerating Climate Simulations Through Hybrid Computing
NASA Technical Reports Server (NTRS)
Zhou, Shujia; Sinno, Scott; Cruz, Carlos; Purcell, Mark
2009-01-01
Unconventional multi-core processors (e.g., IBM Cell B/E and NYIDIDA GPU) have emerged as accelerators in climate simulation. However, climate models typically run on parallel computers with conventional processors (e.g., Intel and AMD) using MPI. Connecting accelerators to this architecture efficiently and easily becomes a critical issue. When using MPI for connection, we identified two challenges: (1) identical MPI implementation is required in both systems, and; (2) existing MPI code must be modified to accommodate the accelerators. In response, we have extended and deployed IBM Dynamic Application Virtualization (DAV) in a hybrid computing prototype system (one blade with two Intel quad-core processors, two IBM QS22 Cell blades, connected with Infiniband), allowing for seamlessly offloading compute-intensive functions to remote, heterogeneous accelerators in a scalable, load-balanced manner. Currently, a climate solar radiation model running with multiple MPI processes has been offloaded to multiple Cell blades with approx.10% network overhead.
NASA Astrophysics Data System (ADS)
Keselman, Paul; Yu, Elaine Y.; Zhou, Xinyi Y.; Goodwill, Patrick W.; Chandrasekharan, Prashant; Ferguson, R. Matthew; Khandhar, Amit P.; Kemp, Scott J.; Krishnan, Kannan M.; Zheng, Bo; Conolly, Steven M.
2017-05-01
Magnetic particle imaging (MPI) is an emerging tracer-based medical imaging modality that images non-radioactive, kidney-safe superparamagnetic iron oxide (SPIO) tracers. MPI offers quantitative, high-contrast and high-SNR images, so MPI has exceptional promise for applications such as cell tracking, angiography, brain perfusion, cancer detection, traumatic brain injury and pulmonary imaging. In assessing MPI’s utility for applications mentioned above, it is important to be able to assess tracer short-term biodistribution as well as long-term clearance from the body. Here, we describe the biodistribution and clearance for two commonly used tracers in MPI: Ferucarbotran (Meito Sangyo Co., Japan) and LS-oo8 (LodeSpin Labs, Seattle, WA). We successfully demonstrate that 3D MPI is able to quantitatively assess short-term biodistribution, as well as long-term tracking and clearance of these tracers in vivo.
NASA Astrophysics Data System (ADS)
Fuchs, A.; Androsov, A.; Harig, S.; Hiller, W.; Rakowsky, N.
2012-04-01
Based on the jeopardy of devastating tsunamis and the unpredictability of such events, tsunami modelling as part of warning systems is still a contemporary topic. The tsunami group of Alfred Wegener Institute developed the simulation tool TsunAWI as contribution to the Early Warning System in Indonesia. Although the precomputed scenarios for this purpose qualify for satisfying deliverables, the study of further improvements continues. While TsunAWI is governed by the Shallow Water Equations, an extension of the model is based on a nonhydrostatic approach. At the arrival of a tsunami wave in coastal regions with rough bathymetry, the term containing the nonhydrostatic part of pressure, that is neglected in the original hydrostatic model, gains in importance. In consideration of this term, a better approximation of the wave is expected. Differences of hydrostatic and nonhydrostatic model results are contrasted in the standard benchmark problem of a solitary wave runup on a plane beach. The observation data provided by Titov and Synolakis (1995) serves as reference. The nonhydrostatic approach implies a set of equations that are similar to the Shallow Water Equations, so the variation of the code can be implemented on top. However, this additional routines cause a lot of issues you have to cope with. So far the computations of the model were purely explicit. In the nonhydrostatic version the determination of an additional unknown and the solution of a large sparse system of linear equations is necessary. The latter constitutes the lion's share of computing time and memory requirement. Since the corresponding matrix is only symmetric in structure and not in values, an iterative Krylov Subspace Method is used, in particular the restarted Generalized Minimal Residual Algorithm GMRES(m). With regard to optimization, we present a comparison of several combinations of sequential and parallel preconditioning techniques respective number of iterations and setup/application time. Since the used software package pARMS 3.2, that provides solving and preconditioning techniques, works via MPI parallelism, in an auxiliary branch we adapted TsunAWI and switched from OpenMP to MPI with attached importance to internal partition management.
Using Modules with MPICH-G2 (and "Loose Ends")
NASA Technical Reports Server (NTRS)
Chang, Johnny; Thigpen, William W. (Technical Monitor)
2002-01-01
A new approach to running complex, distributed MPI jobs using the MPICH-G2 library is described. This approach allows the user to switch between different versions of compilers, system libraries, MPI libraries, etc. via the "module" command. The key idea is a departure from the prescribed "(jobtype=mpi)" approach to running distributed MPI jobs. The new method requires the user to provide a script that will be run as the "executable" with the "(jobtype=single)" RSL attribute. The major advantage of the proposed method is to enable users to decide in their own script what modules, environment, etc. they would like to have in running their job.
New Trends in Radionuclide Myocardial Perfusion Imaging
Hung, Guang-Uei; Wang, Yuh-Feng; Su, Hung-Yi; Hsieh, Te-Chun; Ko, Chi-Lun; Yen, Ruoh-Fang
2016-01-01
Radionuclide myocardial perfusion imaging (MPI) with single photon emission computed tomography (SPECT) has been widely used clinically as one of the major functional imaging modalities for patients with coronary artery disease (CAD) for decades. Ample evidence has supported the use of MPI as a useful and important tool in the diagnosis, risk stratification and treatment planning for CAD. Although popular in the United States, MPI has become the most frequently used imaging modality among all nuclear medicine tests in Taiwan. However, it should be acknowledged that MPI SPECT does have its limitations. These include false-positive results due to certain artifacts, false-negative due to balanced ischemia, complexity and adverse reaction arising from current pharmacological stressors, time consuming nature of the imaging procedure, no blood flow quantitation and relatively high radiation exposure. The purpose of this article was to review the recent trends in nuclear cardiology, including the utilization of positron emission tomography (PET) for MPI, new stressor, new SPECT camera with higher resolution and higher sensitivity, dynamic SPECT protocol for blood flow quantitation, new software of phase analysis for evaluation of LV dyssynchrony, and measures utilized for reducing radiation exposure of MPI. PMID:27122946
Relaxation-based viscosity mapping for magnetic particle imaging.
Utkur, M; Muslu, Y; Saritas, E U
2017-05-07
Magnetic particle imaging (MPI) has been shown to provide remarkable contrast for imaging applications such as angiography, stem cell tracking, and cancer imaging. Recently, there is growing interest in the functional imaging capabilities of MPI, where 'color MPI' techniques have explored separating different nanoparticles, which could potentially be used to distinguish nanoparticles in different states or environments. Viscosity mapping is a promising functional imaging application for MPI, as increased viscosity levels in vivo have been associated with numerous diseases such as hypertension, atherosclerosis, and cancer. In this work, we propose a viscosity mapping technique for MPI through the estimation of the relaxation time constant of the nanoparticles. Importantly, the proposed time constant estimation scheme does not require any prior information regarding the nanoparticles. We validate this method with extensive experiments in an in-house magnetic particle spectroscopy (MPS) setup at four different frequencies (between 250 Hz and 10.8 kHz) and at three different field strengths (between 5 mT and 15 mT) for viscosities ranging between 0.89 mPa · s-15.33 mPa · s. Our results demonstrate the viscosity mapping ability of MPI in the biologically relevant viscosity range.
Development of training modules for magnetic particle inspection
NASA Astrophysics Data System (ADS)
Kosaka, Daigo; Eisenmann, David J.; Enyart, Darrel; Nakagawa, Norio; Lo, Chester; Orman, David
2015-03-01
Magnetic particle inspection (MPI) is a nondestructive evaluation technique used with ferromagnetic materials. Although the application of this method may appear straightforward, MPI combines the complicated nature of electromagnetics, metallurgical material effects, fluid-particle motion dynamics, and physiological human factors into a single inspection. To fully appreciate industry specifications such as ASTM E-1444, users should develop a basic understanding of the many factors that are involved in MPI. We have developed a series of MPI training modules that are aimed at addressing this requirement. The modules not only offer qualitative explanations, but also show quantitative explanations in terms of measurement and numerical simulation data in many instances. There are five modules in all. Module ♯1 shows characteristics of waveforms and magnetizing methods. This allows MPI practitioners to make optimum choice of waveform and magnetizing method. Module ♯2 explains how material properties relate to the magnetic characteristics. Module ♯3 shows the strength of the excitation field or the flux leakage from a crack and how it compares to the detectability of a crack by MPI. Module ♯4 shows how specimen status may influence defect detection. Module ♯5 shows the effects of particle properties on defect detection.
Parallel hyperbolic PDE simulation on clusters: Cell versus GPU
NASA Astrophysics Data System (ADS)
Rostrup, Scott; De Sterck, Hans
2010-12-01
Increasingly, high-performance computing is looking towards data-parallel computational devices to enhance computational performance. Two technologies that have received significant attention are IBM's Cell Processor and NVIDIA's CUDA programming model for graphics processing unit (GPU) computing. In this paper we investigate the acceleration of parallel hyperbolic partial differential equation simulation on structured grids with explicit time integration on clusters with Cell and GPU backends. The message passing interface (MPI) is used for communication between nodes at the coarsest level of parallelism. Optimizations of the simulation code at the several finer levels of parallelism that the data-parallel devices provide are described in terms of data layout, data flow and data-parallel instructions. Optimized Cell and GPU performance are compared with reference code performance on a single x86 central processing unit (CPU) core in single and double precision. We further compare the CPU, Cell and GPU platforms on a chip-to-chip basis, and compare performance on single cluster nodes with two CPUs, two Cell processors or two GPUs in a shared memory configuration (without MPI). We finally compare performance on clusters with 32 CPUs, 32 Cell processors, and 32 GPUs using MPI. Our GPU cluster results use NVIDIA Tesla GPUs with GT200 architecture, but some preliminary results on recently introduced NVIDIA GPUs with the next-generation Fermi architecture are also included. This paper provides computational scientists and engineers who are considering porting their codes to accelerator environments with insight into how structured grid based explicit algorithms can be optimized for clusters with Cell and GPU accelerators. It also provides insight into the speed-up that may be gained on current and future accelerator architectures for this class of applications. Program summaryProgram title: SWsolver Catalogue identifier: AEGY_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEGY_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GPL v3 No. of lines in distributed program, including test data, etc.: 59 168 No. of bytes in distributed program, including test data, etc.: 453 409 Distribution format: tar.gz Programming language: C, CUDA Computer: Parallel Computing Clusters. Individual compute nodes may consist of x86 CPU, Cell processor, or x86 CPU with attached NVIDIA GPU accelerator. Operating system: Linux Has the code been vectorised or parallelized?: Yes. Tested on 1-128 x86 CPU cores, 1-32 Cell Processors, and 1-32 NVIDIA GPUs. RAM: Tested on Problems requiring up to 4 GB per compute node. Classification: 12 External routines: MPI, CUDA, IBM Cell SDK Nature of problem: MPI-parallel simulation of Shallow Water equations using high-resolution 2D hyperbolic equation solver on regular Cartesian grids for x86 CPU, Cell Processor, and NVIDIA GPU using CUDA. Solution method: SWsolver provides 3 implementations of a high-resolution 2D Shallow Water equation solver on regular Cartesian grids, for CPU, Cell Processor, and NVIDIA GPU. Each implementation uses MPI to divide work across a parallel computing cluster. Additional comments: Sub-program numdiff is used for the test run.
González-Domínguez, Jorge; Remeseiro, Beatriz; Martín, María J
2017-02-01
The analysis of the interference patterns on the tear film lipid layer is a useful clinical test to diagnose dry eye syndrome. This task can be automated with a high degree of accuracy by means of the use of tear film maps. However, the time required by the existing applications to generate them prevents a wider acceptance of this method by medical experts. Multithreading has been previously successfully employed by the authors to accelerate the tear film map definition on multicore single-node machines. In this work, we propose a hybrid message-passing and multithreading parallel approach that further accelerates the generation of tear film maps by exploiting the computational capabilities of distributed-memory systems such as multicore clusters and supercomputers. The algorithm for drawing tear film maps is parallelized using Message Passing Interface (MPI) for inter-node communications and the multithreading support available in the C++11 standard for intra-node parallelization. The original algorithm is modified to reduce the communications and increase the scalability. The hybrid method has been tested on 32 nodes of an Intel cluster (with two 12-core Haswell 2680v3 processors per node) using 50 representative images. Results show that maximum runtime is reduced from almost two minutes using the previous only-multithreaded approach to less than ten seconds using the hybrid method. The hybrid MPI/multithreaded implementation can be used by medical experts to obtain tear film maps in only a few seconds, which will significantly accelerate and facilitate the diagnosis of the dry eye syndrome. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Cazet, Aurélie; Charest, Jonathan; Bennett, Daniel C; Sambrooks, Cecilia Lopez; Contessa, Joseph N
2014-01-01
Asparagine-linked glycosylation is an endoplasmic reticulum co- and post-translational modification that enables the transit and function of receptor tyrosine kinase (RTK) glycoproteins. To gain insight into the regulatory role of glycosylation enzymes on RTK function, we investigated shRNA and siRNA knockdown of mannose phosphate isomerase (MPI), an enzyme required for mature glycan precursor biosynthesis. Loss of MPI activity reduced phosphorylation of FGFR family receptors in U-251 and SKMG-3 malignant glioma cell lines and also resulted in significant decreases in FRS2, Akt, and MAPK signaling. However, MPI knockdown did not affect ligand-induced activation or signaling of EGFR or MET RTKs, suggesting that FGFRs are more susceptible to MPI inhibition. The reductions in FGFR signaling were not caused by loss of FGF ligands or receptors, but instead were caused by interference with receptor dimerization. Investigations into the cellular consequences of MPI knockdown showed that cellular programs driven by FGFR signaling, and integral to the clinical progression of malignant glioma, were impaired. In addition to a blockade of cellular migration, MPI knockdown also significantly reduced glioma cell clonogenic survival following ionizing radiation. Therefore our results suggest that targeted inhibition of enzymes required for cell surface receptor glycosylation can be manipulated to produce discrete and limited consequences for critical client glycoproteins expressed by tumor cells. Furthermore, this work identifies MPI as a potential enzymatic target for disrupting cell surface receptor-dependent survival signaling and as a novel approach for therapeutic radiosensitization.
ImgLib2--generic image processing in Java.
Pietzsch, Tobias; Preibisch, Stephan; Tomancák, Pavel; Saalfeld, Stephan
2012-11-15
ImgLib2 is an open-source Java library for n-dimensional data representation and manipulation with focus on image processing. It aims at minimizing code duplication by cleanly separating pixel-algebra, data access and data representation in memory. Algorithms can be implemented for classes of pixel types and generic access patterns by which they become independent of the specific dimensionality, pixel type and data representation. ImgLib2 illustrates that an elegant high-level programming interface can be achieved without sacrificing performance. It provides efficient implementations of common data types, storage layouts and algorithms. It is the data model underlying ImageJ2, the KNIME Image Processing toolbox and an increasing number of Fiji-Plugins. ImgLib2 is licensed under BSD. Documentation and source code are available at http://imglib2.net and in a public repository at https://github.com/imagej/imglib. Supplementary data are available at Bioinformatics Online. saalfeld@mpi-cbg.de
NASA Technical Reports Server (NTRS)
Ramaswamy, Shankar; Banerjee, Prithviraj
1994-01-01
Appropriate data distribution has been found to be critical for obtaining good performance on Distributed Memory Multicomputers like the CM-5, Intel Paragon and IBM SP-1. It has also been found that some programs need to change their distributions during execution for better performance (redistribution). This work focuses on automatically generating efficient routines for redistribution. We present a new mathematical representation for regular distributions called PITFALLS and then discuss algorithms for redistribution based on this representation. One of the significant contributions of this work is being able to handle arbitrary source and target processor sets while performing redistribution. Another important contribution is the ability to handle an arbitrary number of dimensions for the array involved in the redistribution in a scalable manner. Our implementation of these techniques is based on an MPI-like communication library. The results presented show the low overheads for our redistribution algorithm as compared to naive runtime methods.
Parallel optimization algorithm for drone inspection in the building industry
NASA Astrophysics Data System (ADS)
Walczyński, Maciej; BoŻejko, Wojciech; Skorupka, Dariusz
2017-07-01
In this paper we present an approach for Vehicle Routing Problem with Drones (VRPD) in case of building inspection from the air. In autonomic inspection process there is a need to determine of the optimal route for inspection drone. This is especially important issue because of the very limited flight time of modern multicopters. The method of determining solutions for Traveling Salesman Problem(TSP), described in this paper bases on Parallel Evolutionary Algorithm (ParEA)with cooperative and independent approach for communication between threads. This method described first by Bożejko and Wodecki [1] bases on the observation that if exists some number of elements on certain positions in a number of permutations which are local minima, then those elements will be in the same position in the optimal solution for TSP problem. Numerical experiments were made on BEM computational cluster with using MPI library.
Grebner, Christoph; Becker, Johannes; Weber, Daniel; Bellinger, Daniel; Tafipolski, Maxim; Brückner, Charlotte; Engels, Bernd
2014-09-15
The presented program package, Conformational Analysis and Search Tool (CAST) allows the accurate treatment of large and flexible (macro) molecular systems. For the determination of thermally accessible minima CAST offers the newly developed TabuSearch algorithm, but algorithms such as Monte Carlo (MC), MC with minimization, and molecular dynamics are implemented as well. For the determination of reaction paths, CAST provides the PathOpt, the Nudge Elastic band, and the umbrella sampling approach. Access to free energies is possible through the free energy perturbation approach. Along with a number of standard force fields, a newly developed symmetry-adapted perturbation theory-based force field is included. Semiempirical computations are possible through DFTB+ and MOPAC interfaces. For calculations based on density functional theory, a Message Passing Interface (MPI) interface to the Graphics Processing Unit (GPU)-accelerated TeraChem program is available. The program is available on request. Copyright © 2014 Wiley Periodicals, Inc.
Eddy current-shielded x-space relaxometer for sensitive magnetic nanoparticle characterization
Bauer, L. M.; Hensley, D. W.; Zheng, B.; Tay, Z. W.; Goodwill, P. W.; Griswold, M. A.; Conolly, S. M.
2016-01-01
The development of magnetic particle imaging (MPI) has created a need for optimized magnetic nanoparticles. Magnetic particle relaxometry is an excellent tool for characterizing potential tracers for MPI. In this paper, we describe the design and construction of a high-throughput tabletop relaxometer that is able to make sensitive measurements of MPI tracers without the need for a dedicated shield room. PMID:27250472
Eddy current-shielded x-space relaxometer for sensitive magnetic nanoparticle characterization.
Bauer, L M; Hensley, D W; Zheng, B; Tay, Z W; Goodwill, P W; Griswold, M A; Conolly, S M
2016-05-01
The development of magnetic particle imaging (MPI) has created a need for optimized magnetic nanoparticles. Magnetic particle relaxometry is an excellent tool for characterizing potential tracers for MPI. In this paper, we describe the design and construction of a high-throughput tabletop relaxometer that is able to make sensitive measurements of MPI tracers without the need for a dedicated shield room.
[Peritonitis in diverticulitis: the Bern concept].
Seiler, C A; Brügger, L; Maurer, C A; Renzulli, P; Büchler, M W
1998-01-01
The colon is the most frequent origine for a diffuse peritonitis and diverticular perforation is again the most common source of a spontaneous secondary peritonitis. This paper first focuses on the treatment of peritonitis and secondly on the strategies of source control in peritonitis with special emphasis on the tactics (primary anastomosis vs. Hartmann procedure with colostomy) for surgical source control. Prospective analysis of 404 patients suffering from peritonitis (11/93-2/98), treated with an uniform treatment concept including early operation, source control and extensive intraoperative lavage (20 to 30 liters) as a standard procedure. Other treatment measures were added in special indications "on demand" only. Peritonitis was graded with the Mannheim Peritonitis Index (MPI). Tactics of source control in peritonitis due to diverticulitis were performed according to "general condition" respectively the MPI of the patient. The 404 patients averaged a MPI of 19 (0-35) in "local" peritonitis and a MPI of 26 (11-43) in "diffuse" peritonitis. The colon as a source of peritonitis resulted in MPI of 16 (0-33) in the case of "local" respectively 27 (11-43) in "diffuse" peritonitis. From 181 patients suffering from diverticulitis 144 needed an operation and in 78 (54%) peritonitis was present. Fourty-six percent (36) of the patients suffered from "local", 54% (42) from "diffuse" peritonitis. Resection with primary anastomosis was performed in 26% (20/78) whereas in 74% (58/78) of the patients a Hartmann procedure with colostomy was performed. The correlating MPI was 16 (0-28) vs. 23 (16-27) respectively. The analysis of complications and mortality based on the MPI showed a decent discrimination potential for primary anastomosis vs Hartmann procedure: morbidity 35% vs. 41%; reoperation 5% vs. 5%; mortality 0% vs. 14%. In case of peritonitis due to diverticulitis the treatment of peritonitis comes first. Thanks to advances in intensive care and improved anti-inflammatory care, a more conservative surgical concept nowadays is accepted. In the case of diverticulitis the MPI is helpful to choose between primary anastomosis vs. Hartmann procedure with colostomy as source control. The MPI includes the "general condition" of the patient into the tactical decision how to attain source control.
[A failed experiment - Carl Friedrich von Weizsäcker, Jürgen Habermas and the Max-Planck Society].
Leendertz, Ariane
2014-01-01
From 1970 to 1980 Carl Friedrich von Weizsäcker headed the Max-Planck-lnstitut zur Erforschung der Lebensbedingungen der wissenschaftlich-technischen Welt (MPI for the study of the living conditions of the world of science and technology) in Starnberg, jointly with Jürgen Habermas since 1971. From the start, the Max Planck Society regarded the new institute as an experiment that might perhaps be aborted a few years later. This is exactly what happened. With the retirement of Weizsäcker, his section was closed and the whole institute was renamed. In 1981. Habermas resigned, and then the institute was closed. This paper focusses on some of the problem constellations within the institute that partly explain its development and eventual closure: its birth out of the idea of scientific policy advice, the debates within the Max Planck Society and the complex relationship between Weizsäcker and Jürgen Habermas.
Improved field free line magnetic particle imaging using saddle coils.
Erbe, Marlitt; Sattel, Timo F; Buzug, Thorsten M
2013-12-01
Magnetic particle imaging (MPI) is a novel tracer-based imaging method detecting the distribution of superparamagnetic iron oxide (SPIO) nanoparticles in vivo in three dimensions and in real time. Conventionally, MPI uses the signal emitted by SPIO tracer material located at a field free point (FFP). To increase the sensitivity of MPI, however, an alternative encoding scheme collecting the particle signal along a field free line (FFL) was proposed. To provide the magnetic fields needed for line imaging in MPI, a very efficient scanner setup regarding electrical power consumption is needed. At the same time, the scanner needs to provide a high magnetic field homogeneity along the FFL as well as parallel to its alignment to prevent the appearance of artifacts, using efficient radon-based reconstruction methods arising for a line encoding scheme. This work presents a dynamic FFL scanner setup for MPI that outperforms all previously presented setups in electrical power consumption as well as magnetic field quality.
MPI implementation of PHOENICS: A general purpose computational fluid dynamics code
NASA Astrophysics Data System (ADS)
Simunovic, S.; Zacharia, T.; Baltas, N.; Spalding, D. B.
1995-03-01
PHOENICS is a suite of computational analysis programs that are used for simulation of fluid flow, heat transfer, and dynamical reaction processes. The parallel version of the solver EARTH for the Computational Fluid Dynamics (CFD) program PHOENICS has been implemented using Message Passing Interface (MPI) standard. Implementation of MPI version of PHOENICS makes this computational tool portable to a wide range of parallel machines and enables the use of high performance computing for large scale computational simulations. MPI libraries are available on several parallel architectures making the program usable across different architectures as well as on heterogeneous computer networks. The Intel Paragon NX and MPI versions of the program have been developed and tested on massively parallel supercomputers Intel Paragon XP/S 5, XP/S 35, and Kendall Square Research, and on the multiprocessor SGI Onyx computer at Oak Ridge National Laboratory. The preliminary testing results of the developed program have shown scalable performance for reasonably sized computational domains.
MPI implementation of PHOENICS: A general purpose computational fluid dynamics code
DOE Office of Scientific and Technical Information (OSTI.GOV)
Simunovic, S.; Zacharia, T.; Baltas, N.
1995-04-01
PHOENICS is a suite of computational analysis programs that are used for simulation of fluid flow, heat transfer, and dynamical reaction processes. The parallel version of the solver EARTH for the Computational Fluid Dynamics (CFD) program PHOENICS has been implemented using Message Passing Interface (MPI) standard. Implementation of MPI version of PHOENICS makes this computational tool portable to a wide range of parallel machines and enables the use of high performance computing for large scale computational simulations. MPI libraries are available on several parallel architectures making the program usable across different architectures as well as on heterogeneous computer networks. Themore » Intel Paragon NX and MPI versions of the program have been developed and tested on massively parallel supercomputers Intel Paragon XP/S 5, XP/S 35, and Kendall Square Research, and on the multiprocessor SGI Onyx computer at Oak Ridge National Laboratory. The preliminary testing results of the developed program have shown scalable performance for reasonably sized computational domains.« less
NASA Technical Reports Server (NTRS)
Lawson, Gary; Sosonkina, Masha; Baurle, Robert; Hammond, Dana
2017-01-01
In many fields, real-world applications for High Performance Computing have already been developed. For these applications to stay up-to-date, new parallel strategies must be explored to yield the best performance; however, restructuring or modifying a real-world application may be daunting depending on the size of the code. In this case, a mini-app may be employed to quickly explore such options without modifying the entire code. In this work, several mini-apps have been created to enhance a real-world application performance, namely the VULCAN code for complex flow analysis developed at the NASA Langley Research Center. These mini-apps explore hybrid parallel programming paradigms with Message Passing Interface (MPI) for distributed memory access and either Shared MPI (SMPI) or OpenMP for shared memory accesses. Performance testing shows that MPI+SMPI yields the best execution performance, while requiring the largest number of code changes. A maximum speedup of 23 was measured for MPI+SMPI, but only 11 was measured for MPI+OpenMP.
Performance of OVERFLOW-D Applications based on Hybrid and MPI Paradigms on IBM Power4 System
NASA Technical Reports Server (NTRS)
Djomehri, M. Jahed; Biegel, Bryan (Technical Monitor)
2002-01-01
This report briefly discusses our preliminary performance experiments with parallel versions of OVERFLOW-D applications. These applications are based on MPI and hybrid paradigms on the IBM Power4 system here at the NAS Division. This work is part of an effort to determine the suitability of the system and its parallel libraries (MPI/OpenMP) for specific scientific computing objectives.
: A Scalable and Transparent System for Simulating MPI Programs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Perumalla, Kalyan S
2010-01-01
is a scalable, transparent system for experimenting with the execution of parallel programs on simulated computing platforms. The level of simulated detail can be varied for application behavior as well as for machine characteristics. Unique features of are repeatability of execution, scalability to millions of simulated (virtual) MPI ranks, scalability to hundreds of thousands of host (real) MPI ranks, portability of the system to a variety of host supercomputing platforms, and the ability to experiment with scientific applications whose source-code is available. The set of source-code interfaces supported by is being expanded to support a wider set of applications, andmore » MPI-based scientific computing benchmarks are being ported. In proof-of-concept experiments, has been successfully exercised to spawn and sustain very large-scale executions of an MPI test program given in source code form. Low slowdowns are observed, due to its use of purely discrete event style of execution, and due to the scalability and efficiency of the underlying parallel discrete event simulation engine, sik. In the largest runs, has been executed on up to 216,000 cores of a Cray XT5 supercomputer, successfully simulating over 27 million virtual MPI ranks, each virtual rank containing its own thread context, and all ranks fully synchronized by virtual time.« less
Assessment of Fetal Myocardial Performance Index in Women with Placenta Previa.
Zhang, Na; Sun, Lijuan; Zhang, Lina; Li, Zhen; Han, Jijing; Wu, Qingqing
2017-12-15
BACKGROUND This study investigated whether fetuses of placenta previa pregnancies have cardiac dysfunction by use of a modified myocardial performance index (Mod-MPI). MATERIAL AND METHODS A prospective cross-sectional study was conducted including 178 fetuses at 28-40 weeks of gestation. Eighty-nine fetuses of mothers with placenta previa and without pregnancy complications were recruited (placenta previa group) and matched with 89 fetuses of mothers with normal pregnancies (control group). Fetal cardiac function parameters and perinatal outcomes as well as the Mod-MPI were compared between the 2 groups. RESULTS The median Mod-MPI was significantly increased in fetuses of mothers with placenta previa compared with controls (0.47±0.05 vs. 0.45±0.05; P<0.01). Among fetuses of mothers with or without placenta previa, the Mod-MPI was significantly higher in the incomplete placenta previa group compared with the complete placenta previa group and control group (P<0.01). An increased Mod-MPI in placenta previa pregnancies was independently associated with fetal cord pH <7.2 (odds ratio, 4.8; 95% confidence interval, 0.98-23.54; P=0.003). CONCLUSIONS There is impairment of fetal cardiac function in pregnancies with placenta previa. An increased MPI was independently associated with adverse perinatal outcomes to some extent in the placenta previa pregnancies.
Qinghua, Zhao; Jipeng, Li; Yongxing, Zhang; He, Liang; Xuepeng, Wang; Peng, Yan; Xiaofeng, Wu
2015-04-07
To employ three-dimensional finite element modeling and biomechanical simulation for evaluating the stability and stress conduction of two postoperative internal fixed modeling-multilevel posterior instrumentation ( MPI) and MPI with anterior instrumentation (MPAI) with neck-thoracic vertebral tumor en bloc resection. Mimics software and computed tomography (CT) images were used to establish the three-dimensional (3D) model of vertebrae C5-T2 and simulated the C7 en bloc vertebral resection for MPI and MPAI modeling. Then the statistics and images were transmitted into the ANSYS finite element system and 20N distribution load (simulating body weight) and applied 1 N · m torque on neutral point for simulating vertebral displacement and stress conduction and distribution of motion mode, i. e. flexion, extension, bending and rotating. With a better stability, the displacement of two adjacent vertebral bodies of MPI and MPAI modeling was less than that of complete vertebral modeling. No significant differences existed between each other. But as for stress shielding effect reduction, MPI was slightly better than MPAI. From biomechanical point of view, two internal instrumentations with neck-thoracic tumor en bloc resection may achieve an excellent stability with no significant differences. But with better stress conduction, MPI is more advantageous in postoperative reconstruction.
Zheng, Bo; von See, Marc P.; Yu, Elaine; Gunel, Beliz; Lu, Kuan; Vazin, Tandis; Schaffer, David V.; Goodwill, Patrick W.; Conolly, Steven M.
2016-01-01
Stem cell therapies have enormous potential for treating many debilitating diseases, including heart failure, stroke and traumatic brain injury. For maximal efficacy, these therapies require targeted cell delivery to specific tissues followed by successful cell engraftment. However, targeted delivery remains an open challenge. As one example, it is common for intravenous deliveries of mesenchymal stem cells (MSCs) to become entrapped in lung microvasculature instead of the target tissue. Hence, a robust, quantitative imaging method would be essential for developing efficacious cell therapies. Here we show that Magnetic Particle Imaging (MPI), a novel technique that directly images iron-oxide nanoparticle-tagged cells, can longitudinally monitor and quantify MSC administration in vivo. MPI offers near-ideal image contrast, depth penetration, and robustness; these properties make MPI both ultra-sensitive and linearly quantitative. Here, we imaged, for the first time, the dynamic trafficking of intravenous MSC administrations using MPI. Our results indicate that labeled MSC injections are immediately entrapped in lung tissue and then clear to the liver within one day, whereas standard iron oxide particle (Resovist) injections are immediately taken up by liver and spleen. Longitudinal MPI-CT imaging also indicated a clearance half-life of MSC iron oxide labels in the liver at 4.6 days. Finally, our ex vivo MPI biodistribution measurements of iron in liver, spleen, heart, and lungs after injection showed excellent agreement (R2 = 0.943) with measurements from induction coupled plasma spectrometry. These results demonstrate that MPI offers strong utility for noninvasively imaging and quantifying the systemic distribution of cell therapies and other therapeutic agents. PMID:26909106
Min, James K; Hasegawa, James T; Machacz, Susanne F; O'Day, Ken
2016-02-01
This study compared costs and clinical outcomes of invasive versus non-invasive diagnostic evaluations for patients with suspected in-stent restenosis (ISR) after percutaneous coronary intervention. We developed a decision model to compare 2 year diagnosis-related costs for patients who presented with suspected ISR and were evaluated by: (1) invasive coronary angiography (ICA); (2) non-invasive stress testing strategy of myocardial perfusion imaging (MPI) with referral to ICA based on MPI; (3) coronary CT angiography-based testing strategy with referral to ICA based on CCTA. Costs were modeled from the payer's perspective using 2014 Medicare rates. 56 % of patients underwent follow-up diagnostic testing over 2 years. Compared to ICA, MPI (98.6 %) and CCTA (98.1 %) exhibited lower rates of correct diagnoses. Non-invasive strategies were associated with reduced referrals to ICA and costs compared to an ICA-based strategy, with diagnostic costs lower for CCTA than MPI. Overall 2-year costs were highest for ICA for both metallic as well as BVS stents ($1656 and $1656, respectively) when compared to MPI ($1444 and $1411) and CCTA. CCTA costs differed based upon stent size and type, and were highest for metallic stents >3.0 mm followed by metallic stents <3.0 mm, BVS < 3.0 mm and BVS > 3.0 mm ($1466 vs. $1242 vs. $855 vs. $490, respectively). MPI for suspected ISR results in lower costs and rates of complications than invasive strategies using ICA while maintaining high diagnostic performance. Depending upon stent size and type, CCTA results in lower costs than MPI.
Meybeck, Michel; Horowitz, A.J.; Grosbois, C.
2004-01-01
Spatial analysis (1994-2001) and temporal trends (1980-2000) for particulate-associated metals at key stations in the Seine River Basin have been determined using a new metal pollution index (MPI). The MPI is based on the concentrations of Cd, Cu, Hg, Pb and Zn, normalized to calculated background levels estimated for each particulate matter samples for four fractions (clays and other aluminosilicates, carbonates, organic matter, and quartz). Background levels ascribed to each fraction were determined from a specific set of samples collected from relatively pristine areas in the upper Seine basin and validated on prehistoric samples. The unitless MPI is designed to vary between 0 for pristine samples to 100 for the ones extremely impacted by human activities and to assess the trends of general metal contamination and its mapping. Throughout the Seine basin, MPI currently range from 1 to 40, but values exceeding 100 have been found in periurban streams and the Eure tributary. Based on the MPI spatial distribution, the Seine River Basin displays a wide range of anthropogenic impacts linked to variations in population density, stream order, wastewater discharges and industrial activities. Correlations between the MPI and other trace elements indicate that anthropogenic impacts also strongly affect the concentrations of Ag, Sb, and P, marginally affect the concentrations of Ba, Ni, and Cr, and appear to have little effect on the concentrations of Li, Be, V, Co, and the major elements. Temporal MPI trends can also be reconstituted from past regulatory surveys. In the early 1980s, MPI were 2-5 times higher than nowadays at most locations, particularly downstream of Greater Paris where it reached levels as high as 250 (now 40), a value characteristic of present Paris urban sewage. The exceptional contamination of the Seine basin is gradually improving over the last 20 years but remains very high. ?? 2004 Elsevier B.V. All rights reserved.
Yıldırım Poyraz, Nilüfer; Özdemir, Elif; Poyraz, Barış Mustafa; Kandemir, Zuhal; Keskin, Mutlay; Türkölmez, Şeyda
2014-01-01
Objective: The aim of this study was to investigate the relationship between patient characteristics and adenosine-related side-effects during stress myocard perfusion imaging (MPI). The effect of presence of adenosine-related side-effects on the diagnostic value of MPI with integrated SPECT/CT system for coronary artery disease (CAD), was also assessed in this study. Methods: Total of 281 patients (109 M, 172 F; mean age:62.6±10) who underwent standard adenosine stress protocol for MPI, were included in this study. All symptoms during adenosine infusion were scored according to the severity and duration. For the estimation of diagnostic value of adenosine MPI with integrated SPECT/CT system, coronary angiography (CAG) or clinical follow-up were used as gold standard. Results: Total of 173 patients (61.6%) experienced adenosine-related side-effects (group 1); flushing, dyspnea, and chest pain were the most common. Other 108 patients completed pharmacologic stress (PS) test without any side-effects (group 2). Test tolerability were similar in the patients with cardiovascular or airway disease to others, however dyspnea were observed significantly more common in patients with mild airway disease. Body mass index (BMI) ≥30 kg/m2 and age ≤45 years were independent predictors of side-effects. The diagnostic value of MPI was similar in both groups. Sensitivity of adenosine MPI SPECT/CT was calculated to be 86%, specificity was 94% and diagnostic accuracy was 92% for diagnosis of CAD. Conclusion: Adenosine MPI is a feasible and well tolerated method in patients who are not suitable for exercise stress test as well as patients with cardiopulmonary disease. However age ≤45 years and BMI ≥30 kg/m2 are the positive predictors of adenosine-related side-effects, the diagnostic value of adenosine MPI SPECT/CT is not affected by the presence of adenosine related side-effects. PMID:25541932
NASA Technical Reports Server (NTRS)
Jost, Gabriele; Labarta, Jesus; Gimenez, Judit
2004-01-01
With the current trend in parallel computer architectures towards clusters of shared memory symmetric multi-processors, parallel programming techniques have evolved that support parallelism beyond a single level. When comparing the performance of applications based on different programming paradigms, it is important to differentiate between the influence of the programming model itself and other factors, such as implementation specific behavior of the operating system (OS) or architectural issues. Rewriting-a large scientific application in order to employ a new programming paradigms is usually a time consuming and error prone task. Before embarking on such an endeavor it is important to determine that there is really a gain that would not be possible with the current implementation. A detailed performance analysis is crucial to clarify these issues. The multilevel programming paradigms considered in this study are hybrid MPI/OpenMP, MLP, and nested OpenMP. The hybrid MPI/OpenMP approach is based on using MPI [7] for the coarse grained parallelization and OpenMP [9] for fine grained loop level parallelism. The MPI programming paradigm assumes a private address space for each process. Data is transferred by explicitly exchanging messages via calls to the MPI library. This model was originally designed for distributed memory architectures but is also suitable for shared memory systems. The second paradigm under consideration is MLP which was developed by Taft. The approach is similar to MPi/OpenMP, using a mix of coarse grain process level parallelization and loop level OpenMP parallelization. As it is the case with MPI, a private address space is assumed for each process. The MLP approach was developed for ccNUMA architectures and explicitly takes advantage of the availability of shared memory. A shared memory arena which is accessible by all processes is required. Communication is done by reading from and writing to the shared memory.
A portable MPI-based parallel vector template library
NASA Technical Reports Server (NTRS)
Sheffler, Thomas J.
1995-01-01
This paper discusses the design and implementation of a polymorphic collection library for distributed address-space parallel computers. The library provides a data-parallel programming model for C++ by providing three main components: a single generic collection class, generic algorithms over collections, and generic algebraic combining functions. Collection elements are the fourth component of a program written using the library and may be either of the built-in types of C or of user-defined types. Many ideas are borrowed from the Standard Template Library (STL) of C++, although a restricted programming model is proposed because of the distributed address-space memory model assumed. Whereas the STL provides standard collections and implementations of algorithms for uniprocessors, this paper advocates standardizing interfaces that may be customized for different parallel computers. Just as the STL attempts to increase programmer productivity through code reuse, a similar standard for parallel computers could provide programmers with a standard set of algorithms portable across many different architectures. The efficacy of this approach is verified by examining performance data collected from an initial implementation of the library running on an IBM SP-2 and an Intel Paragon.
A Portable MPI-Based Parallel Vector Template Library
NASA Technical Reports Server (NTRS)
Sheffler, Thomas J.
1995-01-01
This paper discusses the design and implementation of a polymorphic collection library for distributed address-space parallel computers. The library provides a data-parallel programming model for C + + by providing three main components: a single generic collection class, generic algorithms over collections, and generic algebraic combining functions. Collection elements are the fourth component of a program written using the library and may be either of the built-in types of c or of user-defined types. Many ideas are borrowed from the Standard Template Library (STL) of C++, although a restricted programming model is proposed because of the distributed address-space memory model assumed. Whereas the STL provides standard collections and implementations of algorithms for uniprocessors, this paper advocates standardizing interfaces that may be customized for different parallel computers. Just as the STL attempts to increase programmer productivity through code reuse, a similar standard for parallel computers could provide programmers with a standard set of algorithms portable across many different architectures. The efficacy of this approach is verified by examining performance data collected from an initial implementation of the library running on an IBM SP-2 and an Intel Paragon.
DOE Office of Scientific and Technical Information (OSTI.GOV)
O'Brien, M. J.; Brantley, P. S.
2015-01-20
In order to run Monte Carlo particle transport calculations on new supercomputers with hundreds of thousands or millions of processors, care must be taken to implement scalable algorithms. This means that the algorithms must continue to perform well as the processor count increases. In this paper, we examine the scalability of:(1) globally resolving the particle locations on the correct processor, (2) deciding that particle streaming communication has finished, and (3) efficiently coupling neighbor domains together with different replication levels. We have run domain decomposed Monte Carlo particle transport on up to 2 21 = 2,097,152 MPI processes on the IBMmore » BG/Q Sequoia supercomputer and observed scalable results that agree with our theoretical predictions. These calculations were carefully constructed to have the same amount of work on every processor, i.e. the calculation is already load balanced. We also examine load imbalanced calculations where each domain’s replication level is proportional to its particle workload. In this case we show how to efficiently couple together adjacent domains to maintain within workgroup load balance and minimize memory usage.« less
PICsar: Particle in cell pulsar magnetosphere simulator
NASA Astrophysics Data System (ADS)
Belyaev, Mikhail A.
2016-07-01
PICsar simulates the magnetosphere of an aligned axisymmetric pulsar and can be used to simulate other arbitrary electromagnetics problems in axisymmetry. Written in Fortran, this special relativistic, electromagnetic, charge conservative particle in cell code features stretchable body-fitted coordinates that follow the surface of a sphere, simplifying the application of boundary conditions in the case of the aligned pulsar; a radiation absorbing outer boundary, which allows a steady state to be set up dynamically and maintained indefinitely from transient initial conditions; and algorithms for injection of charged particles into the simulation domain. PICsar is parallelized using MPI and has been used on research problems with 1000 CPUs.
EMPIRE and pyenda: Two ensemble-based data assimilation systems written in Fortran and Python
NASA Astrophysics Data System (ADS)
Geppert, Gernot; Browne, Phil; van Leeuwen, Peter Jan; Merker, Claire
2017-04-01
We present and compare the features of two ensemble-based data assimilation frameworks, EMPIRE and pyenda. Both frameworks allow to couple models to the assimilation codes using the Message Passing Interface (MPI), leading to extremely efficient and fast coupling between models and the data-assimilation codes. The Fortran-based system EMPIRE (Employing Message Passing Interface for Researching Ensembles) is optimized for parallel, high-performance computing. It currently includes a suite of data assimilation algorithms including variants of the ensemble Kalman and several the particle filters. EMPIRE is targeted at models of all kinds of complexity and has been coupled to several geoscience models, eg. the Lorenz-63 model, a barotropic vorticity model, the general circulation model HadCM3, the ocean model NEMO, and the land-surface model JULES. The Python-based system pyenda (Python Ensemble Data Assimilation) allows Fortran- and Python-based models to be used for data assimilation. Models can be coupled either using MPI or by using a Python interface. Using Python allows quick prototyping and pyenda is aimed at small to medium scale models. pyenda currently includes variants of the ensemble Kalman filter and has been coupled to the Lorenz-63 model, an advection-based precipitation nowcasting scheme, and the dynamic global vegetation model JSBACH.
Shelat, Vishal G.; Ahmed, Saleem; Chia, Clement L. K.; Cheah, Yee Lee
2015-01-01
Application of minimal access surgery in acute care surgery is limited due to various reasons. Laparoscopic omental patch repair (LOPR) for perforated peptic ulcer (PPU) surgery is safe and feasible but not widely implemented. We report our early experience of LOPR with emphasis on strict selection criteria. This is a descriptive study of all patients operated on for PPU at academic university-affiliated institutes from December 2010 to February 2012. All the patients who were operated on for LOPR were included as the study population and their records were studied. Perioperative outcomes, Boey score, Mannheim Peritonitis Index (MPI), and physiologic and operative severity scores for enumeration of mortality and morbidity (POSSUM) scores were calculated. All the data were tabulated in a Microsoft Excel spreadsheet and analyzed using Stata Version 8.x. (StataCorp, College Station, TX, USA). Fourteen patients had LOPR out of a total of 45 patients operated for the PPU. Mean age was 46 years (range 22−87 years). Twelve patients (86%) had a Boey score of 0 and all patients had MPI < 21 (mean MPI = 14). The predicted POSSUM morbidity and mortality were 36% and 7%, respectively. Mean ulcer size was 5 mm (range 2−10 mm), mean operating time was 100 minutes (range 70−123 minutes) and mean length of hospital stay was 4 days (range 3−6 days). There was no morbidity or mortality pertaining to LOPR. LOPR should be offered by acute care surgical teams when local expertise is available. This can optimize patient outcomes when strict selection criteria are applied. PMID:25692444
An evaluation of MPI message rate on hybrid-core processors
Barrett, Brian W.; Brightwell, Ron; Grant, Ryan; ...
2014-11-01
Power and energy concerns are motivating chip manufacturers to consider future hybrid-core processor designs that may combine a small number of traditional cores optimized for single-thread performance with a large number of simpler cores optimized for throughput performance. This trend is likely to impact the way in which compute resources for network protocol processing functions are allocated and managed. In particular, the performance of MPI match processing is critical to achieving high message throughput. In this paper, we analyze the ability of simple and more complex cores to perform MPI matching operations for various scenarios in order to gain insightmore » into how MPI implementations for future hybrid-core processors should be designed.« less
MPI investigation for 40G NRZ link with low-RL cable assemblies
NASA Astrophysics Data System (ADS)
Satake, Toshiaki; Berdinskikh, Tatiana; Thongdaeng, Rutsuda; Faysanyo, Pitak; Gurreri, Michael
2017-01-01
Bit Error Ratio (BER) dependence on received power was studied for 40Gb/s NRZ short optical fiber transmission, including a series of four low return loss (RL 21dB) and low insertion loss (IL 0.1dB) connections. The calculated power penalty (PP) was 0.15dB for BER 10-11. Although the fiber length was within DFB laser's coherent length of 100m and the multi path interference (MPI) value was 34.3dB, no PP of BER was observed. There was no PP due to low MPI probably because the polarization of the signal pulses were not aligned for optical interference, indicating that NRZ systems have a high resistance to MPI.
Exploiting Efficient Transpacking for One-Sided Communication and MPI-IO
NASA Astrophysics Data System (ADS)
Mir, Faisal Ghias; Träff, Jesper Larsson
Based on a construction of socalled input-output datatypes that define a mapping between non-consecutive input and output buffers, we outline an efficient method for copying of structured data. We term this operation transpacking, and show how transpacking can be applied for the MPI implementation of one-sided communication and MPI-IO. For one-sided communication via shared-memory, we demonstrate the expected performance improvements by up to a factor of two. For individual MPI-IO, the time to read or write from file dominates the overall time, but even here efficient transpacking can in some scenarios reduce file I/O time considerably. The reported results have been achieved on a single NEC SX-8 vector node.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Slattery, Stuart R.
In this study we analyze and extend mesh-free algorithms for three-dimensional data transfer problems in partitioned multiphysics simulations. We first provide a direct comparison between a mesh-based weighted residual method using the common-refinement scheme and two mesh-free algorithms leveraging compactly supported radial basis functions: one using a spline interpolation and one using a moving least square reconstruction. Through the comparison we assess both the conservation and accuracy of the data transfer obtained from each of the methods. We do so for a varying set of geometries with and without curvature and sharp features and for functions with and without smoothnessmore » and with varying gradients. Our results show that the mesh-based and mesh-free algorithms are complementary with cases where each was demonstrated to perform better than the other. We then focus on the mesh-free methods by developing a set of algorithms to parallelize them based on sparse linear algebra techniques. This includes a discussion of fast parallel radius searching in point clouds and restructuring the interpolation algorithms to leverage data structures and linear algebra services designed for large distributed computing environments. The scalability of our new algorithms is demonstrated on a leadership class computing facility using a set of basic scaling studies. Finally, these scaling studies show that for problems with reasonable load balance, our new algorithms for both spline interpolation and moving least square reconstruction demonstrate both strong and weak scalability using more than 100,000 MPI processes with billions of degrees of freedom in the data transfer operation.« less
Compiled MPI: Cost-Effective Exascale Applications Development
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bronevetsky, G; Quinlan, D; Lumsdaine, A
2012-04-10
The complexity of petascale and exascale machines makes it increasingly difficult to develop applications that can take advantage of them. Future systems are expected to feature billion-way parallelism, complex heterogeneous compute nodes and poor availability of memory (Peter Kogge, 2008). This new challenge for application development is motivating a significant amount of research and development on new programming models and runtime systems designed to simplify large-scale application development. Unfortunately, DoE has significant multi-decadal investment in a large family of mission-critical scientific applications. Scaling these applications to exascale machines will require a significant investment that will dwarf the costs of hardwaremore » procurement. A key reason for the difficulty in transitioning today's applications to exascale hardware is their reliance on explicit programming techniques, such as the Message Passing Interface (MPI) programming model to enable parallelism. MPI provides a portable and high performance message-passing system that enables scalable performance on a wide variety of platforms. However, it also forces developers to lock the details of parallelization together with application logic, making it very difficult to adapt the application to significant changes in the underlying system. Further, MPI's explicit interface makes it difficult to separate the application's synchronization and communication structure, reducing the amount of support that can be provided by compiler and run-time tools. This is in contrast to the recent research on more implicit parallel programming models such as Chapel, OpenMP and OpenCL, which promise to provide significantly more flexibility at the cost of reimplementing significant portions of the application. We are developing CoMPI, a novel compiler-driven approach to enable existing MPI applications to scale to exascale systems with minimal modifications that can be made incrementally over the application's lifetime. It includes: (1) New set of source code annotations, inserted either manually or automatically, that will clarify the application's use of MPI to the compiler infrastructure, enabling greater accuracy where needed; (2) A compiler transformation framework that leverages these annotations to transform the original MPI source code to improve its performance and scalability; (3) Novel MPI runtime implementation techniques that will provide a rich set of functionality extensions to be used by applications that have been transformed by our compiler; and (4) A novel compiler analysis that leverages simple user annotations to automatically extract the application's communication structure and synthesize most complex code annotations.« less
Cury, Alexandre Ferreira; Bonilha, Andre; Saraiva, Roberto; Campos, Orlando; Carvalho, Antonio Carlos C; De Paola, Angelo Amato V; Fischer, Claudio; Tucci, Paulo Ferreira; Moises, Valdir Ambrosio
2005-05-01
The aim of the study was to analyze the myocardial performance index (MPI), its relationship with the standard variables of systolic and diastolic functions, and the influence of time intervals in an experimental model of female rats with myocardial infarction (MI). Forty-one Wistar female rats were submitted to surgery to induce MI. Six weeks later, Doppler echocardiography was performed to assess infarct size (IS,%), fractional area change (FAC,%), ejection fraction biplane Simpson (EF), E/A ratio of mitral inflow, MPI and its time intervals: isovolumetric contraction (IVCT, ms) and relaxation (IVRT, ms) times, and ejection time (ET, ms); MPI = IVCT + IVRT/ET. EF and FAC were progressively lower in rats with small, medium and large-size MI ( P < .001). E/A ratio was higher only in rats with large-size MI (6.25 +/- 2.69; P < .001). MPI was not different between control rats and small-size MI (0.37 +/- 0.03 vs 0.34 +/- 0.06, P = .87), but different between large and medium-size MI (0.69 +/- 0.08 vs 0.47 +/- 0.07; P < .001) and between these two compared to small-size MI. MPI correlated with IS (r = 0.85; P < .001), EF (r = -0.86; P < .001), FAC (r = -0.77; P < .001) and E/A ratio (r = 0.77; P < .001, non-linear). IVCT was longer in large size MI compared to medium-size MI (31.87 +/- 7.99 vs 15.92 +/- 5.88; P < .001) and correlated with IS (r = 0.85; P < .001) and MPI (r = 0.92; P < .001). ET was shorter only in large-size MI (81.07 +/- 7.23; P < .001), and correlated with IS (r = -0.70; P < .001) and MPI (r = -0.85; P < .001). IVRT was shorter only in large-size compared to medium-size MI (24.40 +/- 5.38 vs 29.69 +/- 5.92; P < .037), had borderline correlation with MPI (r = 0.34; P = .0534) and no correlation with IS (r = 0.26; p = 0.144). The MPI increased with IS, correlated inversely with systolic function parameters and had a non-linear relationship with diastolic function. These changes were due to the increase of IVCT and a decrease of ET, without significant influence of IVRT.
A Case for Application Oblivious Energy-Efficient MPI Runtime
DOE Office of Scientific and Technical Information (OSTI.GOV)
Venkatesh, Akshay; Vishnu, Abhinav; Hamidouche, Khaled
Power has become the major impediment in designing large scale high-end systems. Message Passing Interface (MPI) is the {\\em de facto} communication interface used as the back-end for designing applications, programming models and runtime for these systems. Slack --- the time spent by an MPI process in a single MPI call --- provides a potential for energy and power savings, if an appropriate power reduction technique such as core-idling/Dynamic Voltage and Frequency Scaling (DVFS) can be applied without perturbing application's execution time. Existing techniques that exploit slack for power savings assume that application behavior repeats across iterations/executions. However, an increasingmore » use of adaptive, data-dependent workloads combined with system factors (OS noise, congestion) makes this assumption invalid. This paper proposes and implements Energy Aware MPI (EAM) --- an application-oblivious energy-efficient MPI runtime. EAM uses a combination of communication models of common MPI primitives (point-to-point, collective, progress, blocking/non-blocking) and an online observation of slack for maximizing energy efficiency. Each power lever incurs time overhead, which must be amortized over slack to minimize degradation. When predicted communication time exceeds a lever overhead, the lever is used {\\em as soon as possible} --- to maximize energy efficiency. When mis-prediction occurs, the lever(s) are used automatically at specific intervals for amortization. We implement EAM using MVAPICH2 and evaluate it on ten applications using up to 4096 processes. Our performance evaluation on an InfiniBand cluster indicates that EAM can reduce energy consumption by 5--41\\% in comparison to the default approach, with negligible (less than 4\\% in all cases) performance loss.« less
2011-01-01
Background Rapeseed is an emerging and promising source of dietary protein for human nutrition and health. We previously found that rapeseed protein displayed atypical nutritional properties in humans, characterized by low bioavailability and a high postprandial biological value. The objective of the present study was to investigate the metabolic fate of rapeseed protein isolate (RPI) and its effect on protein fractional synthesis rates (FSR) in various tissues when compared to a milk protein isolate (MPI). Methods Rats (n = 48) were given a RPI or MPI meal, either for the first time or after 2-week adaptation to a MPI or RPI-based diet. They were divided in two groups for measuring the fed-state tissue FSR 2 h after the meal (using a flooding dose of 13C-valine) and the dietary N postprandial distribution at 5 h (using 15N-labeled meals). Results RPI and MPI led to similar FSR and dietary nitrogen (N) losses (ileal and deamination losses of 4% and 12% of the meal, respectively). By contrast, the dietary N incorporation was significantly higher in the intestinal mucosa and liver (+36% and +16%, respectively) and lower in skin (-24%) after RPI than MPI. Conclusions Although RPI and MPI led to the same overall level of postprandial dietary N retention in rats (in line with our findings in humans), this global response conceals marked qualitative differences at the tissue level regarding dietary N accretion. The fact that FSR did not however differed between groups suggest a differential modulation of proteolysis after RPI or MPI ingestion, or other mechanisms that warrant further study. PMID:21787407
Abdelwahab, N A; Morsy, E M H
2018-03-01
TiO 2 /Fe 3 O 4 , TiO 2 /Fe 3 O 4 /chitosan and Methylpyrazolone functionalized TiO 2 /Fe 3 O 4 /chitosan (MPyTMChi) were successfully prepared. The chemical structure of the prepared materials was confirmed by FT-IR spectra, XRD, SEM and TEM. BET surface area increased from 2.4 to 3.1m 2 /g, E g decreased from 2.58 to 2.25eV and more quenching of PL emission spectra was observed upon functionalization of TMChi by MPy. Moreover, high Ti and oxygen percentages were detected by EDX. Magnetization value (M s ) reached 21 emu.g -1 for MPyTMChi. MPyTMChi showed enhanced photocatalytic degradation rate of methylene blue (MB) dye under visibe light irradiation (99.8% after 40min) as compared with that for TiO 2 /Fe 3 O 4 (96.7% after 100min) and TMChi (98.9% after 60min), respectively. It was regarded that the photocatalytic degradation of MB dye on MPyTMChi follows apparent pseudo-first-order according to the Langmuir-Hinshelwood (L-H) model and k app value was 0.089min -1 . Active species trapping experiment revealed that h + and O 2 - played the main role in the photodegradation of MB dye while OH quenching did not greatly affect photodegradation rate. Additionally, MPyTMChi can be efficiently reused for six repetitive cycles. MPyTMChi showed higher antimicrobial activity against gram-positive, gram- negative bacterial and fungal strains while large inhibition zone was observed for gram-positive bacteria. Copyright © 2017 Elsevier B.V. All rights reserved.
Goyal, Parag; Kim, Jiwon; Feher, Attila; Ma, Claudia L.; Gurevich, Sergey; Veal, David R.; Szulc, Massimiliano; Wong, Franklin J.; Ratcliffe, Mark B.; Levine, Robert A.; Devereux, Richard B.; Weinsaft, Jonathan W.
2015-01-01
Objective Ischemic mitral regurgitation (MR) is common, but its response to percutaneous coronary intervention (PCI) is poorly understood. This study tested utility of myocardial perfusion imaging (MPI) for stratification of MR response to PCI. Methods MPI and echo were performed among patients undergoing PCI. MPI was used to assess stress/rest myocardial perfusion. MR was assessed via echo (performed pre- and post-PCI). Results 317 patients with abnormal myocardial perfusion on MPI underwent echo 25±39 days prior to PCI. MR was present in 52%, among whom 24% had advanced (≥moderate) MR. MR was associated with LV chamber dilation on MPI and echo (both p<0.001). Magnitude of global LV perfusion deficits increased in relation to MR severity (p<0.01). Perfusion differences were greatest for global summed rest scores, which were 1.6-fold higher among patients with advanced MR vs. those with mild MR (p=0.004), and 2.4-fold higher vs. those without MR (p<0.001). In multivariate analysis, advanced MR was associated with fixed perfusion defect size on MPI (OR 1.16 per segment [CI 1.002–1.34], p=0.046) independent of LV volume (OR 1.10 per 10ml [CI 1.04–1.17], p=0.002). Follow-up via echo (1.0±0.6 years) demonstrated MR to decrease (≥1 grade) in 31% of patients, and increase in 12%. Patients with increased MR after PCI had more severe inferior perfusion defects on baseline MPI (p=0.028), whereas defects in other distributions and LV volumes were similar (p=NS). Conclusions Extent and distribution of SPECT-evidenced myocardial perfusion defects impacts MR response to revascularization. Increased magnitude of inferior fixed perfusion defects predicts post-PCI progression of MR. PMID:26049923
Cationic ionene as an n-dopant agent of poly(3,4-ethylenedioxythiophene).
Saborío, Maricruz G; Bertran, Oscar; Lanzalaco, Sonia; Häring, Marleen; Díaz Díaz, David; Estrany, Francesc; Alemán, Carlos
2018-04-18
We report the reduction of poly(3,4-ethylenedioxythiophene) (PEDOT) films with a cationic 1,4-diazabicyclo[2.2.2]octane-based ionene bearing N,N'-(meta-phenylene)dibenzamide linkages (mPI). Our main goal is to obtain n-doped PEDOT using a polymeric dopant agent rather than small conventional tetramethylammonium (TMA), as is usual. This has been achieved using a three-step process, which has been individually optimized: (1) preparation of p-doped (oxidized) PEDOT at a constant potential of +1.40 V in acetonitrile with LiClO4 as the electrolyte; (2) dedoping of oxidized PEDOT using a fixed potential of -1.30 V in water; and (3) redoping of dedoped PEDOT applying a reduction potential of -1.10 V in water with mPI. The resulting films display the globular appearance typically observed for PEDOT, with mPI being structured in separated phases forming nanospheres or ultrathin sheets. This organization, which has been supported by atomistic molecular dynamics simulations, resembles the nanosegregated phase distribution observed for PEDOT p-doped with poly(styrenesulfonate). Furthermore, the doping level achieved using mPI as the doping agent is comparable to that achieved using TMA, even though ionene provides distinctive properties to the conducting polymer. For example, films redoped with mPI exhibit much more hydrophilicity than the oxidized ones, whereas films redoped with TMA are hydrophobic. Similarly, films redoped with mPI exhibit the highest thermal stability, while those redoped with TMA show thermal stability that is intermediate between those of the latter and the dedoped PEDOT. Overall, the incorporation of an mPI polycation as the n-dopant into PEDOT has important advantages for modulating the properties of this emblematic conducting polymer.
Smit, Jeff M; Koning, Gerhard; van Rosendael, Alexander R; Dibbets-Schneider, Petra; Mertens, Bart J; Jukema, J Wouter; Delgado, Victoria; Reiber, Johan H C; Bax, Jeroen J; Scholte, Arthur J
2017-10-01
A new method has been developed to calculate fractional flow reserve (FFR) from invasive coronary angiography, the so-called "contrast-flow quantitative flow ratio (cQFR)". Recently, cQFR was compared to invasive FFR in intermediate coronary lesions showing an overall diagnostic accuracy of 85%. The purpose of this study was to investigate the relationship between cQFR and myocardial ischemia assessed by single-photon emission computed tomography myocardial perfusion imaging (SPECT MPI). Patients who underwent SPECT MPI and coronary angiography within 3 months were included. The cQFR computation was performed offline, using dedicated software. The cQFR computation was based on 3-dimensional quantitative coronary angiography (QCA) and computational fluid dynamics. The standard 17-segment model was used to determine the vascular territories. Myocardial ischemia was defined as a summed difference score ≥2 in a vascular territory. A cQFR of ≤0.80 was considered abnormal. Two hundred and twenty-four coronary arteries were analysed in 85 patients. Overall accuracy of cQFR to detect ischemia on SPECT MPI was 90%. In multivariable analysis, cQFR was independently associated with ischemia on SPECT MPI (OR per 0.01 decrease of cQFR: 1.10; 95% CI 1.04-1.18, p = 0.002), whereas clinical and QCA parameters were not. Furthermore, cQFR showed incremental value for the detection of ischemia compared to clinical and QCA parameters (global chi square 48.7 to 62.6; p <0.001). A good relationship between cQFR and SPECT MPI was found. cQFR was independently associated with ischemia on SPECT MPI and showed incremental value to detect ischemia compared to clinical and QCA parameters.
GCView: the genomic context viewer for protein homology searches
Grin, Iwan; Linke, Dirk
2011-01-01
Genomic neighborhood can provide important insights into evolution and function of a protein or gene. When looking at operons, changes in operon structure and composition can only be revealed by looking at the operon as a whole. To facilitate the analysis of the genomic context of a query in multiple organisms we have developed Genomic Context Viewer (GCView). GCView accepts results from one or multiple protein homology searches such as BLASTp as input. For each hit, the neighboring protein-coding genes are extracted, the regions of homology are labeled for each input and the results are presented as a clear, interactive graphical output. It is also possible to add more searches to iteratively refine the output. GCView groups outputs by the hits for different proteins. This allows for easy comparison of different operon compositions and structures. The tool is embedded in the framework of the Bioinformatics Toolkit of the Max-Planck Institute for Developmental Biology (MPI Toolkit). Job results from the homology search tools inside the MPI Toolkit can be forwarded to GCView and results can be subsequently analyzed by sequence analysis tools. Results are stored online, allowing for later reinspection. GCView is freely available at http://toolkit.tuebingen.mpg.de/gcview. PMID:21609955
X-Ray Spectrometer For ROSAT II (SPECTROSAT)
NASA Astrophysics Data System (ADS)
Predehl, Peter; Brauninger, Heinrich
1986-01-01
The objective transmission grating was one of the earliest inventions in the field of X-ray astronomy and has been incorporated into Skylab, HERO-P, and EXOTAT. In recent years there have been advances in grating technology and spectrometer design. A high precision mechanical ruling and replication process for manufacturing large self-supporting transmission gratings has been developed by an industrial manufacturer in cooperation with the Max-Planck-Institute (MPI). Theoretical analyses have determined the optimum configuration of the grating facets and the grating surface in order to correct third order aberations and obtain maximum resolving power. We have verified experimentally that the predicted efficiencies may be achieved. In addition, an experimental study of large grating assemblies for space telescopes was made in industry with scientific guidance by MPI. Main objectives of this study were the determination of mechanical loads during launch, as well as the design, construction and fabrication of a representative model of a ROSAT grating ring. Performancy studies including instrument pro-perties as well as the simulated radiation from hot plasmas have shown the ability of SPECTROSAT to perform high efficiency, high resolution line-spectroscopy on a wide variety of cosmic X-ray sources.
Safety and efficacy of Regadenoson in myocardial perfusion imaging (MPI) stress tests: A review
NASA Astrophysics Data System (ADS)
Ahmed, Ambereen
2018-02-01
Myocardial perfusion imaging (MPI) tests are often used to help diagnose coronary heart disease (CAD). The tests usually involve applying stress, such as hard physical exercise together with administration of vasodilators, to the patients. To date, many of these tests use non-selective A2A adenosine receptor agonists which, however, can be associated with highly undesirable and life-threatening side effects such as chest pain, dyspnea, severe bronchoconstriction and atrioventricular conduction anomalies. Regadenoson is a relatively new, highly selective A2A adenosine receptor agonist, suitable for use in MPI tests which exhibits far fewer adverse side effects and, unlike others testing agents, can be used without the necessity of excessive concomitant exercise. Also, the dose of regadenoson required is not dependent upon patient weight or renal impairment, and it can be rapidly administered by i.v. Injection. Regadenoson use in MPI testing thus has the potential as a simplified, relatively safe, time-saving and cost-effective method for helping diagnose CAD. The present study was designed to review several articles on the safety, efficacy, and suitability of regadenoson in MPI testing for CAD. Overall, the combined studies demonstrated that use of regadenoson in conjunction with low-level exercise in MPI is a highly efficient and relatively safe test for CAD, especially for more severe health-compromised patients.
Assessment of Fetal Myocardial Performance Index in Women with Placenta Previa
Zhang, Na; Sun, Lijuan; Zhang, Lina; Li, Zhen; Han, Jijing; Wu, Qingqing
2017-01-01
Background This study investigated whether fetuses of placenta previa pregnancies have cardiac dysfunction by use of a modified myocardial performance index (Mod-MPI). Material/Methods A prospective cross-sectional study was conducted including 178 fetuses at 28–40 weeks of gestation. Eighty-nine fetuses of mothers with placenta previa and without pregnancy complications were recruited (placenta previa group) and matched with 89 fetuses of mothers with normal pregnancies (control group). Fetal cardiac function parameters and perinatal outcomes as well as the Mod-MPI were compared between the 2 groups. Results The median Mod-MPI was significantly increased in fetuses of mothers with placenta previa compared with controls (0.47±0.05 vs. 0.45±0.05; P<0.01). Among fetuses of mothers with or without placenta previa, the Mod-MPI was significantly higher in the incomplete placenta previa group compared with the complete placenta previa group and control group (P<0.01). An increased Mod-MPI in placenta previa pregnancies was independently associated with fetal cord pH <7.2 (odds ratio, 4.8; 95% confidence interval, 0.98–23.54; P=0.003). Conclusions There is impairment of fetal cardiac function in pregnancies with placenta previa. An increased MPI was independently associated with adverse perinatal outcomes to some extent in the placenta previa pregnancies. PMID:29242496
NASA Astrophysics Data System (ADS)
Jungclaus, J. H.; Fischer, N.; Haak, H.; Lohmann, K.; Marotzke, J.; Matei, D.; Mikolajewicz, U.; Notz, D.; von Storch, J. S.
2013-06-01
MPI-ESM is a new version of the global Earth system model developed at the Max Planck Institute for Meteorology. This paper describes the ocean state and circulation as well as basic aspects of variability in simulations contributing to the fifth phase of the Coupled Model Intercomparison Project (CMIP5). The performance of the ocean/sea-ice model MPIOM, coupled to a new version of the atmosphere model ECHAM6 and modules for land surface and ocean biogeochemistry, is assessed for two model versions with different grid resolution in the ocean. The low-resolution configuration has a nominal resolution of 1.5°, whereas the higher resolution version features a quasiuniform, eddy-permitting global resolution of 0.4°. The paper focuses on important oceanic features, such as surface temperature and salinity, water mass distribution, large-scale circulation, and heat and freshwater transports. In general, these integral quantities are simulated well in comparison with observational estimates, and improvements in comparison with the predecessor system are documented; for example, for tropical variability and sea ice representation. Introducing an eddy-permitting grid configuration in the ocean leads to improvements, in particular, in the representation of interior water mass properties in the Atlantic and in the representation of important ocean currents, such as the Agulhas and Equatorial current systems. In general, however, there are more similarities than differences between the two grid configurations, and several shortcomings, known from earlier versions of the coupled model, prevail.
Ito, Mikio; Noguchi, Hidenori; Ikeda, Katsuyoshi; Uosaki, Kohei
2010-04-07
Effects of metal substrate on the bonding nature of isocyanide group of two aryl isocyanides, 1,4-phenylene diisocyanide (PDI) and 4-methylphenyl isocyanide (MPI), and tilt angle of MPI were examined by measuring sum frequency generation (SFG) spectra of the self-assembled monolayers (SAMs) of these molecules on Au, Pt, Ag, and Pd surfaces. The SFG peaks due to "metal bonded" and "free"-NC groups were resolved by comparing the SFG spectra of PDI with IR spectra obtained by DFT calculations and previous results of vibrational spectroscopy. Based on the peak positions of the "metal bonded"-NC, it is clarified that while PDI and MPI were adsorbed at top sites on Au, Ag, and Pt surfaces, they adsorbed at bridge sites on the Pd surface. The tilt angles of MPI were determined from the intensity ratio between the SFG peaks of C-H symmetric and asymmetric stretching vibrational modes of the CH(3) group. The tilt angles of the MPI SAMs were in the order of Pt < Pd < Ag < Au, reflecting the bonding nature between the -NC group and the substrate atoms.
First in vivo magnetic particle imaging of lung perfusion in rats
NASA Astrophysics Data System (ADS)
Zhou, Xinyi Y.; Jeffris, Kenneth E.; Yu, Elaine Y.; Zheng, Bo; Goodwill, Patrick W.; Nahid, Payam; Conolly, Steven M.
2017-05-01
Pulmonary embolism (PE), along with the closely related condition of deep vein thrombosis, affect an estimated 600 000 patients in the US per year. Untreated, PE carries a mortality rate of 30%. Because many patients experience mild or non-specific symptoms, imaging studies are necessary for definitive diagnosis of PE. Iodinated CT pulmonary angiography is recommended for most patients, while nuclear medicine-based ventilation/perfusion (V/Q) scans are reserved for patients in whom the use of iodine is contraindicated. Magnetic particle imaging (MPI) is an emerging tracer imaging modality with high image contrast (no tissue background signal) and sensitivity to superparamagnetic iron oxide (SPIO) tracer. Importantly, unlike CT or nuclear medicine, MPI uses no ionizing radiation. Further, MPI is not derived from magnetic resonance imaging (MRI); MPI directly images SPIO tracers via their strong electronic magnetization, enabling deep imaging of anatomy including within the lungs, which is very challenging with MRI. Here, the first high-contrast in vivo MPI lung perfusion images of rats are shown using a novel lung perfusion agent, MAA-SPIOs.
Kim, Jeonghyo; Lee, Kil-Soo; Kim, Eun Bee; Paik, Seungwha; Chang, Chulhun L; Park, Tae Jung; Kim, Hwa-Jung; Lee, Jaebeom
2017-10-15
Tuberculosis (TB) is an often neglected, epidemic disease that remains to be controlled by contemporary techniques of medicine and biotechnology. In this study, a nanoscale sensing system, referred to as magnetophoretic immunoassay (MPI) was designed to capture culture filtrate protein (CFP)-10 antigens effectively using two different types of nanoparticles (NPs). Two specific monoclonal antibodies against CFP-10 antigen were used, including gold NPs for signaling and magnetic particles for separation. These results were carefully compared with those obtained using the commercial mycobacteria growth indicator tube (MGIT) test via 2 sequential clinical tests (with ca. 260 clinical samples). The sensing linearity of MPI was shown in the range of pico- to micromoles and the detection limit was 0.3pM. MPI using clinical samples shows robust and reliable sensing while monitoring Mycobacterium tuberculosis (MTB) growth with monitoring time 3-10 days) comparable to that with the MGIT test. Furthermore, MPI distinguished false-positive samples from MGIT-positive samples, probably containing non-tuberculous mycobacteria. Thus, MPI shows promise in early TB diagnosis. Copyright © 2017 Elsevier B.V. All rights reserved.
Model-based phase-shifting interferometer
NASA Astrophysics Data System (ADS)
Liu, Dong; Zhang, Lei; Shi, Tu; Yang, Yongying; Chong, Shiyao; Miao, Liang; Huang, Wei; Shen, Yibing; Bai, Jian
2015-10-01
A model-based phase-shifting interferometer (MPI) is developed, in which a novel calculation technique is proposed instead of the traditional complicated system structure, to achieve versatile, high precision and quantitative surface tests. In the MPI, the partial null lens (PNL) is employed to implement the non-null test. With some alternative PNLs, similar as the transmission spheres in ZYGO interferometers, the MPI provides a flexible test for general spherical and aspherical surfaces. Based on modern computer modeling technique, a reverse iterative optimizing construction (ROR) method is employed for the retrace error correction of non-null test, as well as figure error reconstruction. A self-compiled ray-tracing program is set up for the accurate system modeling and reverse ray tracing. The surface figure error then can be easily extracted from the wavefront data in forms of Zernike polynomials by the ROR method. Experiments of the spherical and aspherical tests are presented to validate the flexibility and accuracy. The test results are compared with those of Zygo interferometer (null tests), which demonstrates the high accuracy of the MPI. With such accuracy and flexibility, the MPI would possess large potential in modern optical shop testing.
Shaikh, Ayaz Hussain; Hanif, Bashir; Siddiqui, Adeel M; Shahab, Hunaina; Qazi, Hammad Ali; Mujtaba, Iqbal
2010-04-01
To determine the association of prolonged ST segment depression after an exercise test with severity of coronary artery disease. A cross sectional study of 100 consecutive patients referred to the cardiology laboratory for stress myocardial perfusion imaging (MPI) conducted between April-August 2008. All selected patients were monitored until their ST segment depression was recovered to baseline. ST segment recovery time was categorized into less and more than 5 minutes. Subsequent gated SPECT-MPI was performed and stratified according to severity of perfusion defect. Association was determined between post exercise ST segment depression recovery time (<5 minutes and >5 minutes) and severity of perfusion defect on MPI. The mean age of the patients was 57.12 +/- 9.0 years. The results showed statistically insignificant association (p > 0.05) between ST segment recovery time of <5 minutes and >5 minutes with low, intermediate or high risk MPI. Our findings suggest that the commonly used cut-off levels used in literature for prolonged, post exercise ST segment depression (>5 minutes into recovery phase) does not correlate with severity of ischaemia based on MPI results.
NASA Technical Reports Server (NTRS)
Katz, Daniel
2004-01-01
PVM Wrapper is a software library that makes it possible for code that utilizes the Parallel Virtual Machine (PVM) software library to run using the message-passing interface (MPI) software library, without needing to rewrite the entire code. PVM and MPI are the two most common software libraries used for applications that involve passing of messages among parallel computers. Since about 1996, MPI has been the de facto standard. Codes written when PVM was popular often feature patterns of {"initsend," "pack," "send"} and {"receive," "unpack"} calls. In many cases, these calls are not contiguous and one set of calls may even exist over multiple subroutines. These characteristics make it difficult to obtain equivalent functionality via a single MPI "send" call. Because PVM Wrapper is written to run with MPI- 1.2, some PVM functions are not permitted and must be replaced - a task that requires some programming expertise. The "pvm_spawn" and "pvm_parent" function calls are not replaced, but a programmer can use "mpirun" and knowledge of the ranks of parent and child tasks with supplied macroinstructions to enable execution of codes that use "pvm_spawn" and "pvm_parent."
Message Passing and Shared Address Space Parallelism on an SMP Cluster
NASA Technical Reports Server (NTRS)
Shan, Hongzhang; Singh, Jaswinder P.; Oliker, Leonid; Biswas, Rupak; Biegel, Bryan (Technical Monitor)
2002-01-01
Currently, message passing (MP) and shared address space (SAS) are the two leading parallel programming paradigms. MP has been standardized with MPI, and is the more common and mature approach; however, code development can be extremely difficult, especially for irregularly structured computations. SAS offers substantial ease of programming, but may suffer from performance limitations due to poor spatial locality and high protocol overhead. In this paper, we compare the performance of and the programming effort required for six applications under both programming models on a 32-processor PC-SMP cluster, a platform that is becoming increasingly attractive for high-end scientific computing. Our application suite consists of codes that typically do not exhibit scalable performance under shared-memory programming due to their high communication-to-computation ratios and/or complex communication patterns. Results indicate that SAS can achieve about half the parallel efficiency of MPI for most of our applications, while being competitive for the others. A hybrid MPI+SAS strategy shows only a small performance advantage over pure MPI in some cases. Finally, improved implementations of two MPI collective operations on PC-SMP clusters are presented.
A study of patients with spinal disease using Maudsley Personality Inventory.
Kasai, Yuichi; Takegami, Kenji; Uchida, Atsumasa
2004-02-01
We administered the Maudsley Personality Inventory (MPI) preoperatively to 303 patients with spinal diseases about to undergo surgery. Patients younger than 20 years, patients previously treated in the Department of Psychiatry, and patients with poor postoperative results were excluded. Patients with N-scores (neuroticism scale) of 39 points or greater or L-scores (lie scale) of 26 points or greater were regarded as "abnormal." Based on clinical definitions we identified 24 "problem patients" during the course and categorized them as "Unsatisfied," "Indecisive," "Doctor shoppers," or "Distrustful." Preoperative MPI categorized 26 patients as abnormal; 22 patients categorized as abnormal became problem patients ( p<0.001). MPI sensitivity and specificity was 84.6% and 99.3%, respectively. Preoperative MPI to patients with spinal disease was found to be useful in detecting problem patients.
Slattery, Stuart R.
2015-12-02
In this study we analyze and extend mesh-free algorithms for three-dimensional data transfer problems in partitioned multiphysics simulations. We first provide a direct comparison between a mesh-based weighted residual method using the common-refinement scheme and two mesh-free algorithms leveraging compactly supported radial basis functions: one using a spline interpolation and one using a moving least square reconstruction. Through the comparison we assess both the conservation and accuracy of the data transfer obtained from each of the methods. We do so for a varying set of geometries with and without curvature and sharp features and for functions with and without smoothnessmore » and with varying gradients. Our results show that the mesh-based and mesh-free algorithms are complementary with cases where each was demonstrated to perform better than the other. We then focus on the mesh-free methods by developing a set of algorithms to parallelize them based on sparse linear algebra techniques. This includes a discussion of fast parallel radius searching in point clouds and restructuring the interpolation algorithms to leverage data structures and linear algebra services designed for large distributed computing environments. The scalability of our new algorithms is demonstrated on a leadership class computing facility using a set of basic scaling studies. Finally, these scaling studies show that for problems with reasonable load balance, our new algorithms for both spline interpolation and moving least square reconstruction demonstrate both strong and weak scalability using more than 100,000 MPI processes with billions of degrees of freedom in the data transfer operation.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tang, Guoping; D'Azevedo, Ed F; Zhang, Fan
2010-01-01
Calibration of groundwater models involves hundreds to thousands of forward solutions, each of which may solve many transient coupled nonlinear partial differential equations, resulting in a computationally intensive problem. We describe a hybrid MPI/OpenMP approach to exploit two levels of parallelisms in software and hardware to reduce calibration time on multi-core computers. HydroGeoChem 5.0 (HGC5) is parallelized using OpenMP for direct solutions for a reactive transport model application, and a field-scale coupled flow and transport model application. In the reactive transport model, a single parallelizable loop is identified to account for over 97% of the total computational time using GPROF.more » Addition of a few lines of OpenMP compiler directives to the loop yields a speedup of about 10 on a 16-core compute node. For the field-scale model, parallelizable loops in 14 of 174 HGC5 subroutines that require 99% of the execution time are identified. As these loops are parallelized incrementally, the scalability is found to be limited by a loop where Cray PAT detects over 90% cache missing rates. With this loop rewritten, similar speedup as the first application is achieved. The OpenMP-parallelized code can be run efficiently on multiple workstations in a network or multiple compute nodes on a cluster as slaves using parallel PEST to speedup model calibration. To run calibration on clusters as a single task, the Levenberg Marquardt algorithm is added to HGC5 with the Jacobian calculation and lambda search parallelized using MPI. With this hybrid approach, 100 200 compute cores are used to reduce the calibration time from weeks to a few hours for these two applications. This approach is applicable to most of the existing groundwater model codes for many applications.« less
A Hybrid MPI/OpenMP Approach for Parallel Groundwater Model Calibration on Multicore Computers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tang, Guoping; D'Azevedo, Ed F; Zhang, Fan
2010-01-01
Groundwater model calibration is becoming increasingly computationally time intensive. We describe a hybrid MPI/OpenMP approach to exploit two levels of parallelism in software and hardware to reduce calibration time on multicore computers with minimal parallelization effort. At first, HydroGeoChem 5.0 (HGC5) is parallelized using OpenMP for a uranium transport model with over a hundred species involving nearly a hundred reactions, and a field scale coupled flow and transport model. In the first application, a single parallelizable loop is identified to consume over 97% of the total computational time. With a few lines of OpenMP compiler directives inserted into the code,more » the computational time reduces about ten times on a compute node with 16 cores. The performance is further improved by selectively parallelizing a few more loops. For the field scale application, parallelizable loops in 15 of the 174 subroutines in HGC5 are identified to take more than 99% of the execution time. By adding the preconditioned conjugate gradient solver and BICGSTAB, and using a coloring scheme to separate the elements, nodes, and boundary sides, the subroutines for finite element assembly, soil property update, and boundary condition application are parallelized, resulting in a speedup of about 10 on a 16-core compute node. The Levenberg-Marquardt (LM) algorithm is added into HGC5 with the Jacobian calculation and lambda search parallelized using MPI. With this hybrid approach, compute nodes at the number of adjustable parameters (when the forward difference is used for Jacobian approximation), or twice that number (if the center difference is used), are used to reduce the calibration time from days and weeks to a few hours for the two applications. This approach can be extended to global optimization scheme and Monte Carol analysis where thousands of compute nodes can be efficiently utilized.« less
32 CFR 637.8 - Identification of MPI.
Code of Federal Regulations, 2010 CFR
2010-07-01
... CRIMINAL INVESTIGATIONS MILITARY POLICE INVESTIGATION Investigations § 637.8 Identification of MPI. (a... referring to themselves as “INVESTIGATOR.” When signing military police records the title “Military Police...
Specification of Fenix MPI Fault Tolerance library version 1.0.1
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gamble, Marc; Van Der Wijngaart, Rob; Teranishi, Keita
This document provides a specification of Fenix, a software library compatible with the Message Passing Interface (MPI) to support fault recovery without application shutdown. The library consists of two modules. The first, termed process recovery , restores an application to a consistent state after it has suffered a loss of one or more MPI processes (ranks). The second specifies functions the user can invoke to store application data in Fenix managed redundant storage, and to retrieve it from that storage after process recovery.
Simultaneous monitoring technique for ASE and MPI noises in distributed Raman Amplified Systems.
Choi, H Y; Jun, S B; Shin, S K; Chung, Y C
2007-07-09
We develop a new technique for simultaneously monitoring the amplified spontaneous emission (ASE) and multi-path interference (MPI) noises in distributed Raman amplified (DRA) systems. This technique utilizes the facts that the degree-of polarization (DOP) of the MPI noise is 1/9, while the ASE noise is unpolarized. The results show that the proposed technique can accurately monitor both of these noises regardless of the bit rates, modulation formats, and optical signal-to-noise ratio (OSNR) levels of the signals.
Kaul, Michael Gerhard; Mummert, Tobias; Jung, Caroline; Salamon, Johannes; Khandhar, Amit P; Ferguson, R Matthew; Kemp, Scott J; Ittrich, Harald; Krishnan, Kannan M; Adam, Gerhard; Knopp, Tobias
2017-05-07
Optimizing tracers for individual imaging techniques is an active field of research. The purpose of this study was to perform in vitro and in vivo magnetic particle imaging (MPI) measurements using a new monodisperse and size-optimized tracer, LS-008, and to compare it with the performance of Resovist, the standard MPI tracer. Magnetic particle spectroscopy (MPS) and in vitro MPI measurements were performed in concerns of concentration and amount of tracer in a phantom. In vivo studies were carried out in healthy FVB mice. The first group (n = 3) received 60 µl LS-008 (87 mM) and the second (n = 3) diluted Resovist of the same concentration and volume. Tracer injections were performed with a syringe pump during a dynamic MPI scan. For anatomic referencing MRI was applied beforehand of the MPI measurements. Summing up MPS examinations and in vitro MPI experiments, LS-008 showed better sensitivity and spatial resolution than Resovist. In vivo both tracers can visualize the propagation of the bolus through the inferior vena cava. MPI with LS-008 did show less temporal fluctuation artifacts and the pulsation of blood due to respiratory and cardiac cycle was detectable. With LS-008 the aorta was distinguishable from the caval vein while with Resovist this failed. A liver vessel and a vessel structure leading cranially could only be observed with LS-008 and not with Resovist. Beside these structural advantages both tracers showed very different blood half-life. For LS-008 we found 88 min. Resovist did show a fast liver accumulation and a half-life of 13 min. Only with LS-008 the perfusion fraction in liver and kidney was measureable. MPI for angiography can be significantly improved by applying more effective tracers. LS-008 shows a clear improvement concerning the delineation while resolving a larger number of vessels in comparison to Resovist. Therefore, in aspects of quality and quantity LS-008 is clearly favorable for angiographic and perfusion studies.
Verra, Martin L; Angst, Felix; Staal, J Bart; Brioschi, Roberto; Lehmann, Susanne; Aeschlimann, André; de Bie, Rob A
2011-06-30
Patients with non-specific back pain are not a homogeneous group but heterogeneous with regard to their bio-psycho-social impairments. This study examined a sample of 173 highly disabled patients with chronic back pain to find out how the three subgroups based on the Multidimensional Pain Inventory (MPI) differed in their response to an inpatient pain management program. Subgroup classification was conducted by cluster analysis using MPI subscale scores at entry into the program. At program entry and at discharge after four weeks, participants completed the MPI, the MOS Short Form-36 (SF-36), the Hospital Anxiety and Depression Scale (HADS), and the Coping Strategies Questionnaire (CSQ). Pairwise analyses of the score changes of the mentioned outcomes of the three MPI subgroups were performed using the Mann-Whitney-U-test for significance. Cluster analysis identified three MPI subgroups in this highly disabled sample: a dysfunctional, interpersonally distressed and an adaptive copers subgroup. The dysfunctional subgroup (29% of the sample) showed the highest level of depression in SF-36 mental health (33.4 ± 13.9), the interpersonally distressed subgroup (35% of the sample) a modest level of depression (46.8 ± 20.4), and the adaptive copers subgroup (32% of the sample) the lowest level of depression (57.8 ± 19.1). Significant differences in pain reduction and improvement of mental health and coping were observed across the three MPI subgroups, i.e. the effect sizes for MPI pain reduction were: 0.84 (0.44-1.24) for the dysfunctional subgroup, 1.22 (0.86-1.58) for the adaptive copers subgroup, and 0.53 (0.24-0.81) for the interpersonally distressed subgroup (p = 0.006 for pairwise comparison). Significant score changes between subgroups concerning activities and physical functioning could not be identified. MPI subgroup classification showed significant differences in score changes for pain, mental health and coping. These findings underscore the importance of assessing individual differences to understand how patients adjust to chronic back pain.
Ge, Zhongming; Feng, Yan; Muthupalani, Sureshkumar; Eurell, Laura Lemke; Taylor, Nancy S.; Whary, Mark T.; Fox, James G.
2011-01-01
To investigate how different enterohepatic Helicobacter species (EHS) influence Helicobacter pylori gastric pathology, C57BL/6 mice were infected with Helicobacter hepaticus or Helicobacter muridarum, followed by H. pylori infection 2 weeks later. Compared to H. pylori-infected mice, mice infected with H. muridarum and H. pylori (HmHp mice) developed significantly lower histopathologic activity index (HAI) scores (P < 0.0001) at 6 and 11 months postinoculation (MPI). However, mice infected with H. hepaticus and H. pylori (HhHp mice) developed more severe gastric pathology at 6 MPI (P = 0.01), with a HAI at 11 MPI (P = 0.8) similar to that of H. pylori-infected mice. H. muridarum-mediated attenuation of gastritis in coinfected mice was associated with significant downregulation of proinflammatory Th1 (interlukin-1beta [Il-1β], gamma interferon [Ifn-γ], and tumor necrosis factor-alpha [Tnf-α]) cytokines at both time points and Th17 (Il-17A) cytokine mRNA levels at 6 MPI in murine stomachs compared to those of H. pylori-infected mice (P < 0.01). Coinfection with H. hepaticus also suppressed H. pylori-induced elevation of gastric Th1 cytokines Ifn-γ and Tnf-α (P < 0.0001) but increased Th17 cytokine mRNA levels (P = 0.028) at 6 MPI. Furthermore, mRNA levels of Il-17A were positively correlated with the severity of helicobacter-induced gastric pathology (HhHp>H. pylori>HmHp) (at 6 MPI, r2 = 0.92, P < 0.0001; at 11 MPI, r2 = 0.82, P < 0.002). Despite disparate effects on gastritis, colonization levels of gastric H. pylori were increased in HhHp mice (at 6 MPI) and HmHp mice (at both time points) compared to those in mono-H. pylori-infected mice. These data suggest that despite consistent downregulation of Th1 responses, EHS coinfection either attenuated or promoted the severity of H. pylori-induced gastric pathology in C57BL/6 mice. This modulation was related to the variable effects of EHS on gastric interleukin 17 (IL-17) responses to H. pylori infection. PMID:21788386
NASA Astrophysics Data System (ADS)
Kaul, Michael Gerhard; Mummert, Tobias; Jung, Caroline; Salamon, Johannes; Khandhar, Amit P.; Ferguson, R. Matthew; Kemp, Scott J.; Ittrich, Harald; Krishnan, Kannan M.; Adam, Gerhard; Knopp, Tobias
2017-05-01
Optimizing tracers for individual imaging techniques is an active field of research. The purpose of this study was to perform in vitro and in vivo magnetic particle imaging (MPI) measurements using a new monodisperse and size-optimized tracer, LS-008, and to compare it with the performance of Resovist, the standard MPI tracer. Magnetic particle spectroscopy (MPS) and in vitro MPI measurements were performed in concerns of concentration and amount of tracer in a phantom. In vivo studies were carried out in healthy FVB mice. The first group (n = 3) received 60 µl LS-008 (87 mM) and the second (n = 3) diluted Resovist of the same concentration and volume. Tracer injections were performed with a syringe pump during a dynamic MPI scan. For anatomic referencing MRI was applied beforehand of the MPI measurements. Summing up MPS examinations and in vitro MPI experiments, LS-008 showed better sensitivity and spatial resolution than Resovist. In vivo both tracers can visualize the propagation of the bolus through the inferior vena cava. MPI with LS-008 did show less temporal fluctuation artifacts and the pulsation of blood due to respiratory and cardiac cycle was detectable. With LS-008 the aorta was distinguishable from the caval vein while with Resovist this failed. A liver vessel and a vessel structure leading cranially could only be observed with LS-008 and not with Resovist. Beside these structural advantages both tracers showed very different blood half-life. For LS-008 we found 88 min. Resovist did show a fast liver accumulation and a half-life of 13 min. Only with LS-008 the perfusion fraction in liver and kidney was measureable. MPI for angiography can be significantly improved by applying more effective tracers. LS-008 shows a clear improvement concerning the delineation while resolving a larger number of vessels in comparison to Resovist. Therefore, in aspects of quality and quantity LS-008 is clearly favorable for angiographic and perfusion studies.
Highly Parallel Alternating Directions Algorithm for Time Dependent Problems
NASA Astrophysics Data System (ADS)
Ganzha, M.; Georgiev, K.; Lirkov, I.; Margenov, S.; Paprzycki, M.
2011-11-01
In our work, we consider the time dependent Stokes equation on a finite time interval and on a uniform rectangular mesh, written in terms of velocity and pressure. For this problem, a parallel algorithm based on a novel direction splitting approach is developed. Here, the pressure equation is derived from a perturbed form of the continuity equation, in which the incompressibility constraint is penalized in a negative norm induced by the direction splitting. The scheme used in the algorithm is composed of two parts: (i) velocity prediction, and (ii) pressure correction. This is a Crank-Nicolson-type two-stage time integration scheme for two and three dimensional parabolic problems in which the second-order derivative, with respect to each space variable, is treated implicitly while the other variable is made explicit at each time sub-step. In order to achieve a good parallel performance the solution of the Poison problem for the pressure correction is replaced by solving a sequence of one-dimensional second order elliptic boundary value problems in each spatial direction. The parallel code is implemented using the standard MPI functions and tested on two modern parallel computer systems. The performed numerical tests demonstrate good level of parallel efficiency and scalability of the studied direction-splitting-based algorithm.
NASA Astrophysics Data System (ADS)
Wu, J.; Yang, Y.; Luo, Q.; Wu, J.
2012-12-01
This study presents a new hybrid multi-objective evolutionary algorithm, the niched Pareto tabu search combined with a genetic algorithm (NPTSGA), whereby the global search ability of niched Pareto tabu search (NPTS) is improved by the diversification of candidate solutions arose from the evolving nondominated sorting genetic algorithm II (NSGA-II) population. Also, the NPTSGA coupled with the commonly used groundwater flow and transport codes, MODFLOW and MT3DMS, is developed for multi-objective optimal design of groundwater remediation systems. The proposed methodology is then applied to a large-scale field groundwater remediation system for cleanup of large trichloroethylene (TCE) plume at the Massachusetts Military Reservation (MMR) in Cape Cod, Massachusetts. Furthermore, a master-slave (MS) parallelization scheme based on the Message Passing Interface (MPI) is incorporated into the NPTSGA to implement objective function evaluations in distributed processor environment, which can greatly improve the efficiency of the NPTSGA in finding Pareto-optimal solutions to the real-world application. This study shows that the MS parallel NPTSGA in comparison with the original NPTS and NSGA-II can balance the tradeoff between diversity and optimality of solutions during the search process and is an efficient and effective tool for optimizing the multi-objective design of groundwater remediation systems under complicated hydrogeologic conditions.
NASA Astrophysics Data System (ADS)
Gupta, Sourendu; Majumdar, Pushan
2018-07-01
We present the results of an effort to accelerate a Rational Hybrid Monte Carlo (RHMC) program for lattice quantum chromodynamics (QCD) simulation for 2 flavors of staggered fermions on multiple Kepler K20X GPUs distributed on different nodes of a Cray XC30. We do not use CUDA but adopt a higher level directive based programming approach using the OpenACC platform. The lattice QCD algorithm is known to be bandwidth bound; our timing results illustrate this clearly, and we discuss how this limits the parallelization gains. We achieve more than a factor three speed-up compared to the CPU only MPI program.
Comparing the OpenMP, MPI, and Hybrid Programming Paradigm on an SMP Cluster
NASA Technical Reports Server (NTRS)
Jost, Gabriele; Jin, Hao-Qiang; anMey, Dieter; Hatay, Ferhat F.
2003-01-01
Clusters of SMP (Symmetric Multi-Processors) nodes provide support for a wide range of parallel programming paradigms. The shared address space within each node is suitable for OpenMP parallelization. Message passing can be employed within and across the nodes of a cluster. Multiple levels of parallelism can be achieved by combining message passing and OpenMP parallelization. Which programming paradigm is the best will depend on the nature of the given problem, the hardware components of the cluster, the network, and the available software. In this study we compare the performance of different implementations of the same CFD benchmark application, using the same numerical algorithm but employing different programming paradigms.
Gui, Zhipeng; Yu, Manzhu; Yang, Chaowei; Jiang, Yunfeng; Chen, Songqing; Xia, Jizhe; Huang, Qunying; Liu, Kai; Li, Zhenlong; Hassan, Mohammed Anowarul; Jin, Baoxuan
2016-01-01
Dust storm has serious disastrous impacts on environment, human health, and assets. The developments and applications of dust storm models have contributed significantly to better understand and predict the distribution, intensity and structure of dust storms. However, dust storm simulation is a data and computing intensive process. To improve the computing performance, high performance computing has been widely adopted by dividing the entire study area into multiple subdomains and allocating each subdomain on different computing nodes in a parallel fashion. Inappropriate allocation may introduce imbalanced task loads and unnecessary communications among computing nodes. Therefore, allocation is a key factor that may impact the efficiency of parallel process. An allocation algorithm is expected to consider the computing cost and communication cost for each computing node to minimize total execution time and reduce overall communication cost for the entire simulation. This research introduces three algorithms to optimize the allocation by considering the spatial and communicational constraints: 1) an Integer Linear Programming (ILP) based algorithm from combinational optimization perspective; 2) a K-Means and Kernighan-Lin combined heuristic algorithm (K&K) integrating geometric and coordinate-free methods by merging local and global partitioning; 3) an automatic seeded region growing based geometric and local partitioning algorithm (ASRG). The performance and effectiveness of the three algorithms are compared based on different factors. Further, we adopt the K&K algorithm as the demonstrated algorithm for the experiment of dust model simulation with the non-hydrostatic mesoscale model (NMM-dust) and compared the performance with the MPI default sequential allocation. The results demonstrate that K&K method significantly improves the simulation performance with better subdomain allocation. This method can also be adopted for other relevant atmospheric and numerical modeling. PMID:27044039
MSAProbs-MPI: parallel multiple sequence aligner for distributed-memory systems.
González-Domínguez, Jorge; Liu, Yongchao; Touriño, Juan; Schmidt, Bertil
2016-12-15
MSAProbs is a state-of-the-art protein multiple sequence alignment tool based on hidden Markov models. It can achieve high alignment accuracy at the expense of relatively long runtimes for large-scale input datasets. In this work we present MSAProbs-MPI, a distributed-memory parallel version of the multithreaded MSAProbs tool that is able to reduce runtimes by exploiting the compute capabilities of common multicore CPU clusters. Our performance evaluation on a cluster with 32 nodes (each containing two Intel Haswell processors) shows reductions in execution time of over one order of magnitude for typical input datasets. Furthermore, MSAProbs-MPI using eight nodes is faster than the GPU-accelerated QuickProbs running on a Tesla K20. Another strong point is that MSAProbs-MPI can deal with large datasets for which MSAProbs and QuickProbs might fail due to time and memory constraints, respectively. Source code in C ++ and MPI running on Linux systems as well as a reference manual are available at http://msaprobs.sourceforge.net CONTACT: jgonzalezd@udc.esSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Myocardial Performance Index for Patients with Overt and Subclinical Hypothyroidism.
Karabulut, Aziz; Doğan, Abdullah; Tuzcu, Alpaslan Kemal
2017-05-25
BACKGROUND Hypothyroid has several effects on the cardiovascular system. Global myocardial performance index (MPI) is used in assessment of both left ventricular (LV) systolic and diastolic function. We compared MPI in hypothyroidism patients vs. normal control subjects. MATERIAL AND METHODS Eighty-two hypothyroid patients were divided into 2 groups: a subclinical hypothyroid (SH) group (n=50), and an overt hypothyroid (OH) group (n=32). The healthy control group (CG) constituted of 37 patients. TSH, FT3, and FT4, anti-TPO, anti-TG, insulin, lipid values, and fasting glucose levels were studied. All patients underwent an echocardiographic examination. Myocardial performance indexes were assessed and standard echocardiographic examinations were investigated. RESULTS MPI averages in OH, SH, and control groups were 0.53±0.06, 0.51±0.05, and 0.44±0.75 mm, respectively. MPI was increased in the OH and SH groups in comparison to CG (p<0.001, p<0.001, respectively). CONCLUSIONS MPI value was significantly higher in hypothyroid patients in comparison to the control group, showing that regression in global left ventricular functions is an important echocardiographic finding. Future studies are required to determine the effects of this finding on long-term cardiovascular outcomes.
Relaxation-based viscosity mapping for magnetic particle imaging
NASA Astrophysics Data System (ADS)
Utkur, M.; Muslu, Y.; Saritas, E. U.
2017-05-01
Magnetic particle imaging (MPI) has been shown to provide remarkable contrast for imaging applications such as angiography, stem cell tracking, and cancer imaging. Recently, there is growing interest in the functional imaging capabilities of MPI, where ‘color MPI’ techniques have explored separating different nanoparticles, which could potentially be used to distinguish nanoparticles in different states or environments. Viscosity mapping is a promising functional imaging application for MPI, as increased viscosity levels in vivo have been associated with numerous diseases such as hypertension, atherosclerosis, and cancer. In this work, we propose a viscosity mapping technique for MPI through the estimation of the relaxation time constant of the nanoparticles. Importantly, the proposed time constant estimation scheme does not require any prior information regarding the nanoparticles. We validate this method with extensive experiments in an in-house magnetic particle spectroscopy (MPS) setup at four different frequencies (between 250 Hz and 10.8 kHz) and at three different field strengths (between 5 mT and 15 mT) for viscosities ranging between 0.89 mPa · s-15.33 mPa · s. Our results demonstrate the viscosity mapping ability of MPI in the biologically relevant viscosity range.
32 CFR 637.2 - Use of MPI and DAC Detectives/Investigators.
Code of Federal Regulations, 2010 CFR
2010-07-01
.../investigators may be employed in joint MPI/USACIDC drug suppression teams; however, the conduct of such... and DAC detectives/investigators may also be utilized to make controlled buys of suspected controlled...
76 FR 18865 - Airworthiness Directives; Bell Helicopter Textron, Inc. Model 212 Helicopters
Federal Register 2010, 2011, 2012, 2013, 2014
2011-04-06
... also requires performing a magnetic particle inspection (MPI) on fittings with certain serial numbers... expanding the applicability to require performing a magnetic particle inspection (MPI) for a crack on the...
MPI-IO: A Parallel File I/O Interface for MPI Version 0.3
NASA Technical Reports Server (NTRS)
Corbett, Peter; Feitelson, Dror; Hsu, Yarsun; Prost, Jean-Pierre; Snir, Marc; Fineberg, Sam; Nitzberg, Bill; Traversat, Bernard; Wong, Parkson
1995-01-01
Thanks to MPI [9], writing portable message passing parallel programs is almost a reality. One of the remaining problems is file I/0. Although parallel file systems support similar interfaces, the lack of a standard makes developing a truly portable program impossible. Further, the closest thing to a standard, the UNIX file interface, is ill-suited to parallel computing. Working together, IBM Research and NASA Ames have drafted MPI-I0, a proposal to address the portable parallel I/0 problem. In a nutshell, this proposal is based on the idea that I/0 can be modeled as message passing: writing to a file is like sending a message, and reading from a file is like receiving a message. MPI-IO intends to leverage the relatively wide acceptance of the MPI interface in order to create a similar I/0 interface. The above approach can be materialized in different ways. The current proposal represents the result of extensive discussions (and arguments), but is by no means finished. Many changes can be expected as additional participants join the effort to define an interface for portable I/0. This document is organized as follows. The remainder of this section includes a discussion of some issues that have shaped the style of the interface. Section 2 presents an overview of MPI-IO as it is currently defined. It specifies what the interface currently supports and states what would need to be added to the current proposal to make the interface more complete and robust. The next seven sections contain the interface definition itself. Section 3 presents definitions and conventions. Section 4 contains functions for file control, most notably open. Section 5 includes functions for independent I/O, both blocking and nonblocking. Section 6 includes functions for collective I/O, both blocking and nonblocking. Section 7 presents functions to support system-maintained file pointers, and shared file pointers. Section 8 presents constructors that can be used to define useful filetypes (the role of filetypes is explained in Section 2 below). Section 9 presents how the error handling mechanism of MPI is supported by the MPI-IO interface. All this is followed by a set of appendices, which contain information about issues that have not been totally resolved yet, and about design considerations. The reader can find there the motivation behind some of our design choices. More information on this would definitely be welcome and will be included in a further release of this document. The first appendix contains a description of MPI-I0's 'hints' structure which is used when opening a file. Appendix B is a discussion of various issues in the support for file pointers. Appendix C explains what we mean in talking about atomic access. Appendix D provides detailed examples of filetype constructors, and Appendix E contains a collection of arguments for and against various design decisions.
Nudi, Francesco; Schillaci, Orazio; Di Belardino, Natale; Versaci, Francesco; Tomai, Fabrizio; Pinto, Annamaria; Neri, Giandomenico; Procaccini, Enrica; Nudi, Alessandro; Frati, Giacomo; Biondi-Zoccai, Giuseppe
2017-10-15
The definition, presentation, and management of myocardial infarction (MI) have changed substantially in the last decade. Whether these changes have impacted on the presence, severity, and localization of necrosis at myocardial perfusion imaging (MPI) has not been appraised to date. Subjects undergoing MPI and reporting a history of clinical MI were shortlisted. We focused on the presence, severity, and localization of necrosis at MPI with a retrospective single-center analysis. A total of 10,476 patients were included, distinguishing 5 groups according to the period in which myocardial perfusion scintigraphy had been performed (2004 to 2005, 2006 to 2007, 2008 to 2009, 2010 to 2011, 2012 to 2013). Trend analysis showed over time a significant worsening in baseline features (e.g., age, diabetes mellitus, and Q waves at electrocardiogram), whereas medical therapy and revascularization were offered with increasing frequency. Over the years, there was also a lower prevalence of normal MPI (from 16.8% to 13.6%) and ischemic MPI (from 35.6% to 32.8%), and a higher prevalence of ischemic and necrotic MPI (from 12.0% to 12.7%) or solely necrotic MPI (from 35.7% to 40.9%, p <0.001). Yet the prevalence of severe ischemia decreased over time from 11.4% to 2.0%, with a similar trend for moderate ischemia (from 15.9% to 11.8%, p <0.001). Similarly sobering results were wound for the prevalence of severe necrosis (from 19.8% to 8.2%) and moderate necrosis (from 8.5% to 7.8%, p = 0.028). These trends were largely confirmed at regional level and after propensity score matching. In conclusion, the outlook of stable patients with previous MI has substantially improved in the last decade, with a decrease in the severity of residual myocardial ischemia and necrosis, despite an apparent worsening in baseline features. Copyright © 2017 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Dieckhoff, J.; Kaul, M. G.; Mummert, T.; Jung, C.; Salamon, J.; Adam, G.; Knopp, T.; Ludwig, F.; Balceris, C.; Ittrich, H.
2017-05-01
Magnetic particle imaging (MPI) facilitates the rapid determination of 3D in vivo magnetic nanoparticle distributions. In this work, liver MPI following intravenous injections of ferucarbotran (Resovist®) was studied. The image reconstruction was based on a calibration measurement, the so called system function. The application of an enhanced system function sample reflecting the particle mobility and aggregation status of ferucarbotran resulted in significantly improved image reconstructions. The finding was supported by characterizations of different ferucarbotran compositions with the magnetorelaxometry and magnetic particle spectroscopy technique. For instance, similar results were obtained between ferucarbotran embedded in freeze-dried mannitol sugar and liver tissue harvested after a ferucarbotran injection. In addition, the combination of multiple shifted measurement patches for a joint reconstruction of the MPI data enlarged the field of view and increased the covering of liver MPI on magnetic resonance images noticeably.
Dieckhoff, J; Kaul, M G; Mummert, T; Jung, C; Salamon, J; Adam, G; Knopp, T; Ludwig, F; Balceris, C; Ittrich, H
2017-05-07
Magnetic particle imaging (MPI) facilitates the rapid determination of 3D in vivo magnetic nanoparticle distributions. In this work, liver MPI following intravenous injections of ferucarbotran (Resovist ® ) was studied. The image reconstruction was based on a calibration measurement, the so called system function. The application of an enhanced system function sample reflecting the particle mobility and aggregation status of ferucarbotran resulted in significantly improved image reconstructions. The finding was supported by characterizations of different ferucarbotran compositions with the magnetorelaxometry and magnetic particle spectroscopy technique. For instance, similar results were obtained between ferucarbotran embedded in freeze-dried mannitol sugar and liver tissue harvested after a ferucarbotran injection. In addition, the combination of multiple shifted measurement patches for a joint reconstruction of the MPI data enlarged the field of view and increased the covering of liver MPI on magnetic resonance images noticeably.
Toxoplasma gondii strain-dependent effects on mouse behaviour.
Kannan, Geetha; Moldovan, Krisztina; Xiao, Jian-Chun; Yolken, Robert H; Jones-Brando, Lorraine; Pletnikov, Mikhail V
2010-06-01
Toxoplasma gondii reportedly manipulates rodent behaviour to increase transmission to its definitive feline host. We compared the effects of mouse infection by two Type II strains of T. gondii, Prugniaud (PRU) and ME49, on attraction to cat odour, locomotor activity, anxiety, sensorimotor gating, and spatial working and recognition memory 2 months post-infection (mpi). Attraction to cat odour was reassessed 7 mpi. At 2 mpi, mice infected with either strain exhibited significantly more attraction to cat odour than uninfected animals did, but only PRU-infected mice exhibited this behaviour 7 mpi. PRU-infected mice had significantly greater body weights and hyperactivity, while ME49-infected mice exhibited impaired spatial working memory. No differences in parasite antibody titres were seen between PRU- and ME49-infected mice. The present data suggest the effect of T. gondii infection on mouse behaviour is parasite strain-dependent.
Architecture and method for a burst buffer using flash technology
Tzelnic, Percy; Faibish, Sorin; Gupta, Uday K.; Bent, John; Grider, Gary Alan; Chen, Hsing-bung
2016-03-15
A parallel supercomputing cluster includes compute nodes interconnected in a mesh of data links for executing an MPI job, and solid-state storage nodes each linked to a respective group of the compute nodes for receiving checkpoint data from the respective compute nodes, and magnetic disk storage linked to each of the solid-state storage nodes for asynchronous migration of the checkpoint data from the solid-state storage nodes to the magnetic disk storage. Each solid-state storage node presents a file system interface to the MPI job, and multiple MPI processes of the MPI job write the checkpoint data to a shared file in the solid-state storage in a strided fashion, and the solid-state storage node asynchronously migrates the checkpoint data from the shared file in the solid-state storage to the magnetic disk storage and writes the checkpoint data to the magnetic disk storage in a sequential fashion.
NASA Astrophysics Data System (ADS)
Kjærgaard, Thomas; Baudin, Pablo; Bykov, Dmytro; Eriksen, Janus Juul; Ettenhuber, Patrick; Kristensen, Kasper; Larkin, Jeff; Liakh, Dmitry; Pawłowski, Filip; Vose, Aaron; Wang, Yang Min; Jørgensen, Poul
2017-03-01
We present a scalable cross-platform hybrid MPI/OpenMP/OpenACC implementation of the Divide-Expand-Consolidate (DEC) formalism with portable performance on heterogeneous HPC architectures. The Divide-Expand-Consolidate formalism is designed to reduce the steep computational scaling of conventional many-body methods employed in electronic structure theory to linear scaling, while providing a simple mechanism for controlling the error introduced by this approximation. Our massively parallel implementation of this general scheme has three levels of parallelism, being a hybrid of the loosely coupled task-based parallelization approach and the conventional MPI +X programming model, where X is either OpenMP or OpenACC. We demonstrate strong and weak scalability of this implementation on heterogeneous HPC systems, namely on the GPU-based Cray XK7 Titan supercomputer at the Oak Ridge National Laboratory. Using the "resolution of the identity second-order Møller-Plesset perturbation theory" (RI-MP2) as the physical model for simulating correlated electron motion, the linear-scaling DEC implementation is applied to 1-aza-adamantane-trione (AAT) supramolecular wires containing up to 40 monomers (2440 atoms, 6800 correlated electrons, 24 440 basis functions and 91 280 auxiliary functions). This represents the largest molecular system treated at the MP2 level of theory, demonstrating an efficient removal of the scaling wall pertinent to conventional quantum many-body methods.
Acoustic 3D modeling by the method of integral equations
NASA Astrophysics Data System (ADS)
Malovichko, M.; Khokhlov, N.; Yavich, N.; Zhdanov, M.
2018-02-01
This paper presents a parallel algorithm for frequency-domain acoustic modeling by the method of integral equations (IE). The algorithm is applied to seismic simulation. The IE method reduces the size of the problem but leads to a dense system matrix. A tolerable memory consumption and numerical complexity were achieved by applying an iterative solver, accompanied by an effective matrix-vector multiplication operation, based on the fast Fourier transform (FFT). We demonstrate that, the IE system matrix is better conditioned than that of the finite-difference (FD) method, and discuss its relation to a specially preconditioned FD matrix. We considered several methods of matrix-vector multiplication for the free-space and layered host models. The developed algorithm and computer code were benchmarked against the FD time-domain solution. It was demonstrated that, the method could accurately calculate the seismic field for the models with sharp material boundaries and a point source and receiver located close to the free surface. We used OpenMP to speed up the matrix-vector multiplication, while MPI was used to speed up the solution of the system equations, and also for parallelizing across multiple sources. The practical examples and efficiency tests are presented as well.
MULTINEST: an efficient and robust Bayesian inference tool for cosmology and particle physics
NASA Astrophysics Data System (ADS)
Feroz, F.; Hobson, M. P.; Bridges, M.
2009-10-01
We present further development and the first public release of our multimodal nested sampling algorithm, called MULTINEST. This Bayesian inference tool calculates the evidence, with an associated error estimate, and produces posterior samples from distributions that may contain multiple modes and pronounced (curving) degeneracies in high dimensions. The developments presented here lead to further substantial improvements in sampling efficiency and robustness, as compared to the original algorithm presented in Feroz & Hobson, which itself significantly outperformed existing Markov chain Monte Carlo techniques in a wide range of astrophysical inference problems. The accuracy and economy of the MULTINEST algorithm are demonstrated by application to two toy problems and to a cosmological inference problem focusing on the extension of the vanilla Λ cold dark matter model to include spatial curvature and a varying equation of state for dark energy. The MULTINEST software, which is fully parallelized using MPI and includes an interface to COSMOMC, is available at http://www.mrao.cam.ac.uk/software/multinest/. It will also be released as part of the SUPERBAYES package, for the analysis of supersymmetric theories of particle physics, at http://www.superbayes.org.
Globalized Newton-Krylov-Schwarz Algorithms and Software for Parallel Implicit CFD
NASA Technical Reports Server (NTRS)
Gropp, W. D.; Keyes, D. E.; McInnes, L. C.; Tidriri, M. D.
1998-01-01
Implicit solution methods are important in applications modeled by PDEs with disparate temporal and spatial scales. Because such applications require high resolution with reasonable turnaround, "routine" parallelization is essential. The pseudo-transient matrix-free Newton-Krylov-Schwarz (Psi-NKS) algorithmic framework is presented as an answer. We show that, for the classical problem of three-dimensional transonic Euler flow about an M6 wing, Psi-NKS can simultaneously deliver: globalized, asymptotically rapid convergence through adaptive pseudo- transient continuation and Newton's method-, reasonable parallelizability for an implicit method through deferred synchronization and favorable communication-to-computation scaling in the Krylov linear solver; and high per- processor performance through attention to distributed memory and cache locality, especially through the Schwarz preconditioner. Two discouraging features of Psi-NKS methods are their sensitivity to the coding of the underlying PDE discretization and the large number of parameters that must be selected to govern convergence. We therefore distill several recommendations from our experience and from our reading of the literature on various algorithmic components of Psi-NKS, and we describe a freely available, MPI-based portable parallel software implementation of the solver employed here.
Zhang, Lijun; Song, Xiantao; Dong, Li; Li, Jianan; Dou, Ruiyu; Fan, Zhanming; An, Jing; Li, Debiao
2018-04-30
The purpose of the work was to evaluate the incremental diagnostic value of free-breathing, contrast-enhanced, whole-heart, 3 T cardiovascular magnetic resonance coronary angiography (CE-MRCA) to stress/rest myocardial perfusion imaging (MPI) and late gadolinium enhancement (LGE) imaging for detecting coronary artery disease (CAD). Fifty-one patients with suspected CAD underwent a comprehensive cardiovascular magnetic resonance (CMR) examination (CE-MRCA, MPI, and LGE). The additive diagnostic value of MRCA to MPI and LGE was evaluated using invasive x-ray coronary angiography (XA) as the standard for defining functionally significant CAD (≥ 50% stenosis in vessels > 2 mm in diameter). 90.2% (46/51) patients (54.0 ± 11.5 years; 71.7% men) completed CE-MRCA successfully. On per-patient basis, compared to MPI/LGE alone or MPI alone, the addition of MRCA resulted in higher sensitivity (100% vs. 76.5%, p < 0.01), no change in specificity (58.3% vs. 66.7%, p = 0.6), and higher accuracy (89.1% vs 73.9%, p < 0.01) for CAD detection (prevalence = 73.9%). Compared to LGE alone, the addition of CE-MRCA resulted in higher sensitivity (97.1% vs. 41.2%, p < 0.01), inferior specificity (83.3% vs. 91.7%, p = 0.02), and higher diagnostic accuracy (93.5% vs. 54.3%, p < 0.01). The inclusion of successful free-breathing, whole-heart, 3 T CE-MRCA significantly improved the sensitivity and diagnostic accuracy as compared to MPI and LGE alone for CAD detection.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-07-30
... Leased Workers From Echelon Service Company, Sun Associated Industries, Inc., MPI Consultants LLC...-site leased workers from Echelon Service Company, Sun Associated Industries, Inc., MPI Consultants LLC...
Shojaeifard, Maryam; Ghaedian, Tahereh; Yaghoobi, Nahid; Malek, Hadi; Firoozabadi, Hasan; Bitarafan-Rajabi, Ahmad; Haghjoo, Majid; Amin, Ahmad; Azizian, Nasrin; Rastgou, Feridoon
2015-01-01
Background: Gated single-photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI) is known as a feasible tool for the measurement of left ventricular ejection fraction (EF) and volumes, which are of great importance in the management and follow-up of patients with coronary artery diseases. However, considering the technical shortcomings of SPECT in the presence of perfusion defect, the accuracy of this method in heart failure patients is still controversial. Objectives: The aim of the present study was to compare the results from gated SPECT MPI with those from echocardiography in heart failure patients to compare echocardiographically-derived left ventricular dimension and function data to those from gated SPECT MPI in heart failure patients. Patients and Methods: Forty-one patients with severely reduced left ventricular systolic function (EF ≤ 35%) who were referred for gated SPECT MPI were prospectively enrolled. Quantification of EF, end-diastolic volume (EDV), and end-systolic volume (ESV) was performed by using quantitative gated spect (QGS) (QGS, version 0.4, May 2009) and emory cardiac toolbox (ECTb) (ECTb, revision 1.0, copyright 2007) software packages. EF, EDV, and ESV were also measured with two-dimensional echocardiography within 3 days after MPI. Results: A good correlation was found between echocardiographically-derived EF, EDV, and ESV and the values derived using QGS (r = 0.67, r = 0.78, and r = 0.80 for EF, EDV, and ESV, respectively; P < 0.001) and ECTb (r = 0.68, 0.79, and r = 0.80 for EF, EDV, and ESV, respectively; P < 0.001). However, Bland-Altman plots indicated significantly different mean values for EF, 11.4 and 20.9 using QGS and ECTb, respectively, as compared with echocardiography. ECTb-derived EDV was also significantly higher than the EDV measured with echocardiography and QGS. The highest correlation between echocardiography and gated SPECT MPI was found for mean values of ESV different. Conclusions: Gated SPECT MPI has a good correlation with echocardiography for the measurement of left ventricular EF, EDV, and ESV in patients with severe heart failure. However, the absolute values of these functional parameters from echocardiography and gated SPECT MPI measured with different software packages should not be used interchangeably. PMID:26889455
Application of the Low-dose One-stop-shop Cardiac CT Protocol with Third-generation Dual-source CT.
Lin, Lu; Wang, Yining; Yi, Yan; Cao, Jian; Kong, Lingyan; Qian, Hao; Zhang, Hongzhi; Wu, Wei; Wang, Yun; Jin, Zhengyu
2017-02-20
Objective To evaluate the feasibility of a low-dose one-stop-shop cardiac CT imaging protocol with third-generation dual-source CT (DSCT). Methods Totally 23 coronary artery disease (CAD) patients were prospectively enrolled between March to September in 2016. All patients underwent an ATP stress dynamic myocardial perfusion imaging (MPI) (data acquired prospectively ECG-triggered during end systole by table shuttle mode in 32 seconds) at 70 kV combined with prospectively ECG-triggered high-pitch coronary artery angiography (CCTA) on a third-generation DSCT system. Myocardial blood flow (MBF) was quantified and compared between perfusion normal and abnormal myocardial segments based on AHA-17-segment model. CCTA images were evaluated qualitatively based on SCCT-18-segment model and the effective dose(ED) was calculated. In patients with subsequent catheter coronary angiography (CCA) as reference,the diagnosis performance of MPI (for per-vessel ≥50% and ≥70% stenosis) and CCTA (for≥50% stenosis) were assessed. Results Of 23 patients who had completed the examination of ATP stress MPI plus CCTA,12 patients received follow-up CCA. At ATP stress MPI,77 segments (19.7%) in 13 patients (56.5%) had perfusion abnormalities. The MBF values of hypo-perfused myocardial segments decreased significantly compared with normal segments [(93±22)ml/(100 ml·min) vs. (147±27)ml/(100 ml·min);t=15.978,P=0.000]. At CCTA,93.9% (308/328) of the coronary segments had diagnostic image quality. With CCA as the reference standard,the per-vessel and per-segment sensitivity,specificity,and accuracy of CCTA for stenosis≥50% were 94.1%,93.5%,and 93.7% and 90.9%,97.8%,and 96.8%,and the per-vessel sensitivity,specificity and accuracy of ATP stress MPI for stenosis≥50% and ≥70% were 68.7%,100%,and 89.5% and 91.7%,100%,and 97.9%. The total ED of MPI and CCTA was (3.9±1.3) mSv [MPI:(3.5±1.2) mSv,CCTA:(0.3±0.1) mSv]. Conclusion The third-generation DSCT stress dynamic MPI at 70 kV combined with prospectively ECG-triggered high-pitch CCTA is a feasible and reliable tool for clinical diagnosis,with remarkably reduced radiation dose.
Genome-scale strain designs based on regulatory minimal cut sets.
Mahadevan, Radhakrishnan; von Kamp, Axel; Klamt, Steffen
2015-09-01
Stoichiometric and constraint-based methods of computational strain design have become an important tool for rational metabolic engineering. One of those relies on the concept of constrained minimal cut sets (cMCSs). However, as most other techniques, cMCSs may consider only reaction (or gene) knockouts to achieve a desired phenotype. We generalize the cMCSs approach to constrained regulatory MCSs (cRegMCSs), where up/downregulation of reaction rates can be combined along with reaction deletions. We show that flux up/downregulations can virtually be treated as cuts allowing their direct integration into the algorithmic framework of cMCSs. Because of vastly enlarged search spaces in genome-scale networks, we developed strategies to (optionally) preselect suitable candidates for flux regulation and novel algorithmic techniques to further enhance efficiency and speed of cMCSs calculation. We illustrate the cRegMCSs approach by a simple example network and apply it then by identifying strain designs for ethanol production in a genome-scale metabolic model of Escherichia coli. The results clearly show that cRegMCSs combining reaction deletions and flux regulations provide a much larger number of suitable strain designs, many of which are significantly smaller relative to cMCSs involving only knockouts. Furthermore, with cRegMCSs, one may also enable the fine tuning of desired behaviours in a narrower range. The new cRegMCSs approach may thus accelerate the implementation of model-based strain designs for the bio-based production of fuels and chemicals. MATLAB code and the examples can be downloaded at http://www.mpi-magdeburg.mpg.de/projects/cna/etcdownloads.html. krishna.mahadevan@utoronto.ca or klamt@mpi-magdeburg.mpg.de Supplementary data are available at Bioinformatics online. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Plant functional diversity affects climate-vegetation interaction
NASA Astrophysics Data System (ADS)
Groner, Vivienne P.; Raddatz, Thomas; Reick, Christian H.; Claussen, Martin
2018-04-01
We present how variations in plant functional diversity affect climate-vegetation interaction towards the end of the African Humid Period (AHP) in coupled land-atmosphere simulations using the Max Planck Institute Earth system model (MPI-ESM). In experiments with AHP boundary conditions, the extent of the green
Sahara varies considerably with changes in plant functional diversity. Differences in vegetation cover extent and plant functional type (PFT) composition translate into significantly different land surface parameters, water cycling, and surface energy budgets. These changes have not only regional consequences but considerably alter large-scale atmospheric circulation patterns and the position of the tropical rain belt. Towards the end of the AHP, simulations with the standard PFT set in MPI-ESM depict a gradual decrease of precipitation and vegetation cover over time, while simulations with modified PFT composition show either a sharp decline of both variables or an even slower retreat. Thus, not the quantitative but the qualitative PFT composition determines climate-vegetation interaction and the climate-vegetation system response to external forcing. The sensitivity of simulated system states to changes in PFT composition raises the question how realistically Earth system models can actually represent climate-vegetation interaction, considering the poor representation of plant diversity in the current generation of land surface models.
Lee, Daniel Jin Keat; Ye, MaDong; Sun, Keith Haozhe; Shelat, Vishalkumar G; Koura, Aaryan
2016-01-01
Introduction. The aim of this study was to compare the outcomes between laparoscopic and open omental patch repair (LOPR versus OR) in patients with similar presentation of perforated peptic ulcer (PPU). The secondary aim was to evaluate the outcomes according to the severity of peritonitis. Methods. All patients who underwent omental patch repair at two university-affiliated institutes between January 2010 and December 2014 were reviewed. Matched cohort between LOPR and OR groups was achieved by only including patients that had ulcer perforation <2 cm in size and symptoms occurring <48 hours. Outcome measures were defined in accordance with length of stay (LOS), postoperative complications, and mortality. Results. 148 patients met the predefined inclusion criteria with LOPR performed in 40 patients. Outcome measures consistently support laparoscopic approach but only length of hospital stay (LOS) achieved statistical significance (LOPR 4 days versus OR 5 days, p < 0.01). In a subgroup analysis of patients with MPI score >21, LOPR is also shown to benefit, particularly resulting in significant shorter LOS (4 days versus 11 days, p < 0.01). Conclusion. LOPR offers improved short-term outcomes in patients who present within 48 hours and with perforation size <2 cm. LOPR also proved to be more beneficial in high MPI cases.
Sun, Keith Haozhe; Koura, Aaryan
2016-01-01
Introduction. The aim of this study was to compare the outcomes between laparoscopic and open omental patch repair (LOPR versus OR) in patients with similar presentation of perforated peptic ulcer (PPU). The secondary aim was to evaluate the outcomes according to the severity of peritonitis. Methods. All patients who underwent omental patch repair at two university-affiliated institutes between January 2010 and December 2014 were reviewed. Matched cohort between LOPR and OR groups was achieved by only including patients that had ulcer perforation <2 cm in size and symptoms occurring <48 hours. Outcome measures were defined in accordance with length of stay (LOS), postoperative complications, and mortality. Results. 148 patients met the predefined inclusion criteria with LOPR performed in 40 patients. Outcome measures consistently support laparoscopic approach but only length of hospital stay (LOS) achieved statistical significance (LOPR 4 days versus OR 5 days, p < 0.01). In a subgroup analysis of patients with MPI score >21, LOPR is also shown to benefit, particularly resulting in significant shorter LOS (4 days versus 11 days, p < 0.01). Conclusion. LOPR offers improved short-term outcomes in patients who present within 48 hours and with perforation size <2 cm. LOPR also proved to be more beneficial in high MPI cases. PMID:27722200
Performance Evaluation of Supercomputers using HPCC and IMB Benchmarks
NASA Technical Reports Server (NTRS)
Saini, Subhash; Ciotti, Robert; Gunney, Brian T. N.; Spelce, Thomas E.; Koniges, Alice; Dossa, Don; Adamidis, Panagiotis; Rabenseifner, Rolf; Tiyyagura, Sunil R.; Mueller, Matthias;
2006-01-01
The HPC Challenge (HPCC) benchmark suite and the Intel MPI Benchmark (IMB) are used to compare and evaluate the combined performance of processor, memory subsystem and interconnect fabric of five leading supercomputers - SGI Altix BX2, Cray XI, Cray Opteron Cluster, Dell Xeon cluster, and NEC SX-8. These five systems use five different networks (SGI NUMALINK4, Cray network, Myrinet, InfiniBand, and NEC IXS). The complete set of HPCC benchmarks are run on each of these systems. Additionally, we present Intel MPI Benchmarks (IMB) results to study the performance of 11 MPI communication functions on these systems.
High-Performance Design Patterns for Modern Fortran
Haveraaen, Magne; Morris, Karla; Rouson, Damian; ...
2015-01-01
This paper presents ideas for using coordinate-free numerics in modern Fortran to achieve code flexibility in the partial differential equation (PDE) domain. We also show how Fortran, over the last few decades, has changed to become a language well-suited for state-of-the-art software development. Fortran’s new coarray distributed data structure, the language’s class mechanism, and its side-effect-free, pure procedure capability provide the scaffolding on which we implement HPC software. These features empower compilers to organize parallel computations with efficient communication. We present some programming patterns that support asynchronous evaluation of expressions comprised of parallel operations on distributed data. We implemented thesemore » patterns using coarrays and the message passing interface (MPI). We compared the codes’ complexity and performance. The MPI code is much more complex and depends on external libraries. The MPI code on Cray hardware using the Cray compiler is 1.5–2 times faster than the coarray code on the same hardware. The Intel compiler implements coarrays atop Intel’s MPI library with the result apparently being 2–2.5 times slower than manually coded MPI despite exhibiting nearly linear scaling efficiency. As compilers mature and further improvements to coarrays comes in Fortran 2015, we expect this performance gap to narrow.« less
NASA Astrophysics Data System (ADS)
Raju, S. G.; Hariharan, Krishnan S.; Park, Da-Hye; Kang, HyoRang; Kolake, Subramanya Mayya
2015-10-01
Molecular dynamics (MD) simulations of ternary polymer electrolyte - ionic liquid mixtures are conducted using an all-atom model. N-alkyl-N-methylpyrrolidinium bis(trifluoromethylsulfonyl)imide ([CnMPy][TFSI], n = 1, 3, 6, 9) and polyethylene oxide (PEO) are used. Microscopic structure, energetics and dynamics of ionic liquid (IL) in these ternary mixtures are studied. Properties of these four pure IL are also calculated and compared to that in ternary mixtures. Interaction between pyrrolidinium cation and TFSI is stronger and there is larger propensity of ion-pair formation in ternary mixtures. Unlike the case in imidazolium IL, near neighbor structural correlation between TFSI reduces with increase in chain length on cation in both pure IL and ternary mixtures. Using spatial density maps, regions where PEO and TFSI interact with pyrrolidinium cation are identified. Oxygens of PEO are above and below the pyrrolidinium ring and away from the bulky alkyl groups whereas TFSI is present close to nitrogen atom of CnMPy. In pure IL, diffusion coefficient (D) of C3MPy is larger than of TFSI but D of C9MPy and C6MPy are larger than that of TFSI. The reasons for alkyl chain dependent phenomena are explored.
OpenGeoSys-GEMS: Hybrid parallelization of a reactive transport code with MPI and threads
NASA Astrophysics Data System (ADS)
Kosakowski, G.; Kulik, D. A.; Shao, H.
2012-04-01
OpenGeoSys-GEMS is a generic purpose reactive transport code based on the operator splitting approach. The code couples the Finite-Element groundwater flow and multi-species transport modules of the OpenGeoSys (OGS) project (http://www.ufz.de/index.php?en=18345) with the GEM-Selektor research package to model thermodynamic equilibrium of aquatic (geo)chemical systems utilizing the Gibbs Energy Minimization approach (http://gems.web.psi.ch/). The combination of OGS and the GEM-Selektor kernel (GEMS3K) is highly flexible due to the object-oriented modular code structures and the well defined (memory based) data exchange modules. Like other reactive transport codes, the practical applicability of OGS-GEMS is often hampered by the long calculation time and large memory requirements. • For realistic geochemical systems which might include dozens of mineral phases and several (non-ideal) solid solutions the time needed to solve the chemical system with GEMS3K may increase exceptionally. • The codes are coupled in a sequential non-iterative loop. In order to keep the accuracy, the time step size is restricted. In combination with a fine spatial discretization the time step size may become very small which increases calculation times drastically even for small 1D problems. • The current version of OGS is not optimized for memory use and the MPI version of OGS does not distribute data between nodes. Even for moderately small 2D problems the number of MPI processes that fit into memory of up-to-date workstations or HPC hardware is limited. One strategy to overcome the above mentioned restrictions of OGS-GEMS is to parallelize the coupled code. For OGS a parallelized version already exists. It is based on a domain decomposition method implemented with MPI and provides a parallel solver for fluid and mass transport processes. In the coupled code, after solving fluid flow and solute transport, geochemical calculations are done in form of a central loop over all finite element nodes with calls to GEMS3K and consecutive calculations of changed material parameters. In a first step the existing MPI implementation was utilized to parallelize this loop. Calculations were split between the MPI processes and afterwards data was synchronized by using MPI communication routines. Furthermore, multi-threaded calculation of the loop was implemented with help of the boost thread library (http://www.boost.org). This implementation provides a flexible environment to distribute calculations between several threads. For each MPI process at least one and up to several dozens of worker threads are spawned. These threads do not replicate the complete OGS-GEM data structure and use only a limited amount of memory. Calculation of the central geochemical loop is shared between all threads. Synchronization between the threads is done by barrier commands. The overall number of local threads times MPI processes should match the number of available computing nodes. The combination of multi-threading and MPI provides an effective and flexible environment to speed up OGS-GEMS calculations while limiting the required memory use. Test calculations on different hardware show that for certain types of applications tremendous speedups are possible.
Fenix, A Fault Tolerant Programming Framework for MPI Applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gamel, Marc; Teranihi, Keita; Valenzuela, Eric
2016-10-05
Fenix provides APIs to allow the users to add fault tolerance capability to MPI-based parallel programs in a transparent manner. Fenix-enabled programs can run through process failures during program execution using a pool of spare processes accommodated by Fenix.
Fast and Accurate Support Vector Machines on Large Scale Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vishnu, Abhinav; Narasimhan, Jayenthi; Holder, Larry
Support Vector Machines (SVM) is a supervised Machine Learning and Data Mining (MLDM) algorithm, which has become ubiquitous largely due to its high accuracy and obliviousness to dimensionality. The objective of SVM is to find an optimal boundary --- also known as hyperplane --- which separates the samples (examples in a dataset) of different classes by a maximum margin. Usually, very few samples contribute to the definition of the boundary. However, existing parallel algorithms use the entire dataset for finding the boundary, which is sub-optimal for performance reasons. In this paper, we propose a novel distributed memory algorithm to eliminatemore » the samples which do not contribute to the boundary definition in SVM. We propose several heuristics, which range from early (aggressive) to late (conservative) elimination of the samples, such that the overall time for generating the boundary is reduced considerably. In a few cases, a sample may be eliminated (shrunk) pre-emptively --- potentially resulting in an incorrect boundary. We propose a scalable approach to synchronize the necessary data structures such that the proposed algorithm maintains its accuracy. We consider the necessary trade-offs of single/multiple synchronization using in-depth time-space complexity analysis. We implement the proposed algorithm using MPI and compare it with libsvm--- de facto sequential SVM software --- which we enhance with OpenMP for multi-core/many-core parallelism. Our proposed approach shows excellent efficiency using up to 4096 processes on several large datasets such as UCI HIGGS Boson dataset and Offending URL dataset.« less
Al-Mallah, Mouaz H; Pascual, Thomas N B; Mercuri, Mathew; Vitola, João V; Karthikeyan, Ganesan; Better, Nathan; Dondi, Maurizio; Paez, Diana; Einstein, Andrew J
2018-05-15
There is growing concern about radiation exposure from nuclear myocardial perfusion imaging (MPI), particularly among younger patients who are more prone to develop untoward effects of ionizing radiation, and hence US and European professional society guidelines recommend age as a consideration in weighing radiation risk from MPI. We aimed to determine how patient radiation doses from MPI vary across age groups in a large contemporary international cohort. Data were collected as part of a global cross-sectional study of centers performing MPI coordinated by the International Atomic Energy Agency (IAEA). Sites provided information on each MPI study completed during a single week in March-April 2013. We compared across age groups laboratory adherence to pre-specified radiation-related best practices, radiation effective dose (ED; a whole-body measure reflecting the amount of radiation to each organ and its relative sensitivity to radiation's deleterious effects), and the proportion of patients with ED ≤ 9 mSv, a target level specified in guidelines. Among 7911 patients undergoing MPI in 308 laboratories in 65 countries, mean ED was 10.0 ± 4.5 mSv with slightly higher exposure among younger age groups (trend p value < 0.001). There was no difference in the proportion of patients with ED ≤ 9 mSv across age groups, or in adherence to best practices based on the median age of patients in a laboratory. In contemporary nuclear cardiology practice, the age of the patient appears not to impact protocol selection and radiation dose, contrary to professional society guidelines. Copyright © 2018. Published by Elsevier B.V.
Asih, Sali; Mayer, Tom G; Williams, Mark; Choi, Yun Hee; Gatchel, Robert J
2015-12-01
The objectives of this study: (1) to assess whether Multidimensional Pain Inventory (MPI) profiles predicted differential responses to a functional restoration program (FRP) in chronic disabling occupational musculoskeletal disorder (CDOMD) patients; (2) to examine whether coping style improves following FRP; and (3) to determine whether discharge MPI profiles predict discharge psychosocial and 1-year socioeconomic outcomes. Consecutive CDOMD patients (N=716) were classified into Adaptive Coper (AC, n=209), Interpersonally Distressed (ID, n=154), Dysfunctional (DYS, n=310), and Anomalous (n=43) using the MPI, and reclassified at discharge. Profiles were compared on psychosocial measures and 1-year socioeconomic outcomes. An intent-to-treat sample analyzed the effect of drop-outs on treatment responsiveness. The MPI classification significantly predicted program completion (P=0.001), although the intent-to-treat analyses found no significant effects of drop-out on treatment responsiveness. There was a significant increase in the number of patients who became AC or Anomalous at FRP discharge and a decrease in those who were ID or DYS. Patients who changed or remained as DYS at FRP discharge reported the highest levels of pain, disability, and depression. No significant interaction effect was found between MPI group and time for pain intensity or disability. All groups improved on psychosocial measures at discharge. DYS patients had decreased work retention and a greater health care utilization at 1 year. An FRP was clinically effective for CDOMD patients regardless of initial MPI profiles. The FRP modified profiles, with patients changing from negative to positive profiles. Discharge DYS were more likely to have poor 1-year outcomes. Those classified as Anomalous had a good prognosis for functional recovery similar to ACs.
The effect of patient anxiety and depression on motion during myocardial perfusion SPECT imaging.
Lyra, Vassiliki; Kallergi, Maria; Rizos, Emmanouil; Lamprakopoulos, Georgios; Chatziioannou, Sofia N
2016-08-22
Patient motion during myocardial perfusion SPECT imaging (MPI) may be triggered by a patient's physical and/or psychological discomfort. The aim of this study was to investigate the impact of state anxiety (patient's reaction to exam-related stress), trait anxiety (patient's personality characteristic) and depression on patient motion during MPI. All patients that underwent MPI in our department in a six-month period were prospectively enrolled. One hundred eighty-three patients (45 females; 138 males) filled in the State-Trait Anxiety Inventory (STAI) and the Beck Depression Inventory (BDI), along with a short questionnaire regarding their age, height and weight, level of education in years, occupation, and marital status. Cardiovascular and other co-morbidity factors were also evaluated. Through inspection of raw data on cinematic display, the presence or absence of patient motion was registered and classified into mild, moderate and severe, for both phases involved in image acquisition. The correlation of patient motion in the stress and delay phases of MPI and each of the other variables was investigated and the corresponding Pearson's coefficients of association were calculated. The anxiety-motion (r = 0.43, P < 0.0001) and depression-motion (r = 0.32, P < 0.0001) correlation results were moderately strong and statistically significant for the female but not the male patients. All the other variables did not demonstrate any association with motion in MPI, except a weak correlation between age and motion in females (r = 0.23, P < 0.001). The relationship between anxiety-motion and depression-motion identified in female patients represents the first supporting evidence of psychological discomfort as predisposing factor for patient motion during MPI.
NASA Astrophysics Data System (ADS)
Dhavalikar, Rohan; Rinaldi, Carlos
2016-12-01
Magnetic nanoparticles in alternating magnetic fields (AMFs) transfer some of the field's energy to their surroundings in the form of heat, a property that has attracted significant attention for use in cancer treatment through hyperthermia and in developing magnetic drug carriers that can be actuated to release their cargo externally using magnetic fields. To date, most work in this field has focused on the use of AMFs that actuate heat release by nanoparticles over large regions, without the ability to select specific nanoparticle-loaded regions for heating while leaving other nanoparticle-loaded regions unaffected. In parallel, magnetic particle imaging (MPI) has emerged as a promising approach to image the distribution of magnetic nanoparticle tracers in vivo, with sub-millimeter spatial resolution. The underlying principle in MPI is the application of a selection magnetic field gradient, which defines a small region of low bias field, superimposed with an AMF (of lower frequency and amplitude than those normally used to actuate heating by the nanoparticles) to obtain a signal which is proportional to the concentration of particles in the region of low bias field. Here we extend previous models for estimating the energy dissipation rates of magnetic nanoparticles in uniform AMFs to provide theoretical predictions of how the selection magnetic field gradient used in MPI can be used to selectively actuate heating by magnetic nanoparticles in the low bias field region of the selection magnetic field gradient. Theoretical predictions are given for the spatial decay in energy dissipation rate under magnetic field gradients representative of those that can be achieved with current MPI technology. These results underscore the potential of combining MPI and higher amplitude/frequency actuation AMFs to achieve selective magnetic fluid hyperthermia (MFH) guided by MPI.
Korosoglou, G; Hansen, A; Bekeredjian, R; Filusch, A; Hardt, S; Wolf, D; Schellberg, D; Katus, H A; Kuecherer, H
2006-03-01
To evaluate whether myocardial parametric imaging (MPI) is superior to visual assessment for the evaluation of myocardial viability. Myocardial contrast echocardiography (MCE) was assessed in 11 pigs before, during, and after left anterior descending coronary artery occlusion and in 32 patients with ischaemic heart disease by using intravenous SonoVue administration. In experimental studies perfusion defect area assessment by MPI was compared with visually guided perfusion defect planimetry. Histological assessment of necrotic tissue was the standard reference. In clinical studies viability was assessed on a segmental level by (1) visual analysis of myocardial opacification; (2) quantitative estimation of myocardial blood flow in regions of interest; and (3) MPI. Functional recovery between three and six months after revascularisation was the standard reference. In experimental studies, compared with visually guided perfusion defect planimetry, planimetric assessment of infarct size by MPI correlated more significantly with histology (r2 = 0.92 versus r2 = 0.56) and had a lower intraobserver variability (4% v 15%, p < 0.05). In clinical studies, MPI had higher specificity (66% v 43%, p < 0.05) than visual MCE and good accuracy (81%) for viability detection. It was less time consuming (3.4 (1.6) v 9.2 (2.4) minutes per image, p < 0.05) than quantitative blood flow estimation by regions of interest and increased the agreement between observers interpreting myocardial perfusion (kappa = 0.87 v kappa = 0.75, p < 0.05). MPI is useful for the evaluation of myocardial viability both in animals and in patients. It is less time consuming than quantification analysis by regions of interest and less observer dependent than visual analysis. Thus, strategies incorporating this technique may be valuable for the evaluation of myocardial viability in clinical routine.
Weinsaft, Jonathan W; Manoushagian, Shant J; Patel, Taral; Shakoor, Aqsa; Kim, Robert J; Mirchandani, Sunil; Lin, Fay; Wong, Franklin J; Szulc, Massimiliano; Okin, Peter M; Kligfield, Paul D; Min, James K
2009-01-01
To assess the utility of stress electrocardiography (ECG) for identifying the presence and severity of obstructive coronary artery disease (CAD) defined by coronary computed tomographic angiography (CCTA) among patients with normal nuclear myocardial perfusion imaging (MPI). The study population comprised 119 consecutive patients with normal MPI who also underwent CCTA (interval 3.5+/-3.8 months). Stress ECG was performed at the time of MPI. CCTA and MPI were interpreted using established scoring systems, and CCTA was used to define the presence and extent of CAD, which was quantified by a coronary artery jeopardy score. Within this population, 28 patients (24%) had obstructive CAD identified by CCTA. The most common CAD pattern was single-vessel CAD (61%), although proximal vessel involvement was present in 46% of patients. Patients with CAD were nearly three times more likely to have positive standard test responses (1 mm ST-segment deviation) than patients with patent coronary arteries (36 vs. 13%, P=0.007). In multivariate analysis, a positive ST-segment test response was an independent marker for CAD (odds ratio: 2.02, confidence interval: 1.09-3.78, P=0.03) even after adjustment for a composite of clinical cardiac risk factors (odds ratio: 1.85, confidence interval: 1.05-3.23, P=0.03). Despite uniformly normal MPI, mean coronary jeopardy score was three-fold higher among patients with positive compared to those with negative ST-segment response to exercise or dobutamine stress (1.9+/-2.7 vs. 0.5+/-1.4, P=0.03). Stress-induced ST-segment deviation is an independent marker for obstructive CAD among patients with normal MPI. A positive stress ECG identifies patients with a greater anatomic extent of CAD as quantified by coronary jeopardy score.
Kosonen, Jukka; Kulmala, Juha-Pekka; Müller, Erich; Avela, Janne
2017-03-21
Anti-pronation orthoses, like medially posted insoles (MPI), have traditionally been used to treat various of lower limb problems. Yet, we know surprisingly little about their effects on overall foot motion and lower limb mechanics across walking and running, which represent highly different loading conditions. To address this issue, multi-segment foot and lower limb mechanics was examined among 11 overpronating men with normal (NORM) and MPI insoles during walking (self-selected speed 1.70±0.19m/s vs 1.72±0.20m/s, respectively) and running (4.04±0.17m/s vs 4.10±0.13m/s, respectively). The kinematic results showed that MPI reduced the peak forefoot eversion movement in respect to both hindfoot and tibia across walking and running when compared to NORM (p<0.05-0.01). No differences were found in hindfoot eversion between conditions. The kinetic results showed no insole effects in walking, but during running MPI shifted center of pressure medially under the foot (p<0.01) leading to an increase in frontal plane moments at the hip (p<0.05) and knee (p<0.05) joints and a reduction at the ankle joint (p<0.05). These findings indicate that MPI primarily controlled the forefoot motion across walking and running. While kinetic response to MPI was more pronounced in running than walking, kinematic effects were essentially similar across both modes. This suggests that despite higher loads placed upon lower limb during running, there is no need to have a stiffer insoles to achieve similar reduction in the forefoot motion than in walking. Copyright © 2017 Elsevier Ltd. All rights reserved.
Development of Modeling and Simulation for Magnetic Particle Inspection Using Finite Elements
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lee, Jun-Youl
2003-01-01
Magnetic particle inspection (MPI) is a widely used nondestructive inspection method for aerospace applications essentially limited to experiment-based approaches. The analysis of MPI characteristics that affect sensitivity and reliability contributes not only reductions in inspection design cost and time but also improvement of analysis of experimental data. Magnetic particles are easily attracted toward a high magnetic field gradient. Selection of a magnetic field source, which produces a magnetic field gradient large enough to detect a defect in a test sample or component, is an important factor in magnetic particle inspection. In this work a finite element method (FEM) has beenmore » employed for numerical calculation of the MPI simulation technique. The FEM method is known to be suitable for complicated geometries such as defects in samples. This thesis describes the research that is aimed at providing a quantitative scientific basis for magnetic particle inspection. A new FEM solver for MPI simulation has been developed in this research for not only nonlinear reversible permeability materials but also irreversible hysteresis materials that are described by the Jiles-Atherton model. The material is assumed to have isotropic ferromagnetic properties in this research (i.e., the magnetic properties of the material are identical in all directions in a single crystal). In the research, with a direct current field mode, an MPI situation has been simulated to measure the estimated volume of magnetic particles around defect sites before and after removing any external current fields. Currently, this new MPI simulation package is limited to solving problems with the single current source from either a solenoid or an axial directional current rod.« less
Impact of physical permafrost processes on hydrological change
NASA Astrophysics Data System (ADS)
Hagemann, Stefan; Blome, Tanja; Beer, Christian; Ekici, Altug
2015-04-01
Permafrost or perennially frozen ground is an important part of the terrestrial cryosphere; roughly one quarter of Earth's land surface is underlain by permafrost. As it is a thermal phenomenon, its characteristics are highly dependent on climatic factors. The impact of the currently observed warming, which is projected to persist during the coming decades due to anthropogenic CO2 input, certainly has effects for the vast permafrost areas of the high northern latitudes. The quantification of these effects, however, is scientifically still an open question. This is partly due to the complexity of the system, where several feedbacks are interacting between land and atmosphere, sometimes counterbalancing each other. Moreover, until recently, many global circulation models (GCMs) and Earth system models (ESMs) lacked the sufficient representation of permafrost physics in their land surface schemes. Within the European Union FP7 project PAGE21, the land surface scheme JSBACH of the Max-Planck-Institute for Meteorology ESM (MPI-ESM) has been equipped with the representation of relevant physical processes for permafrost studies. These processes include the effects of freezing and thawing of soil water for both energy and water cycles, thermal properties depending on soil water and ice contents, and soil moisture movement being influenced by the presence of soil ice. In the present study, it will be analysed how these permafrost relevant processes impact projected hydrological changes over northern hemisphere high latitude land areas. For this analysis, the atmosphere-land part of MPI-ESM, ECHAM6-JSBACH, is driven by prescribed SST and sea ice in an AMIP2-type setup with and without the newly implemented permafrost processes. Observed SST and sea ice for 1979-1999 are used to consider induced changes in the simulated hydrological cycle. In addition, simulated SST and sea ice are taken from a MPI-ESM simulation conducted for CMIP5 following the RCP8.5 scenario. The corresponding simulations with ECHAM6-JSBACH are used to assess differences in projected hydrological changes induced by the permafrost relevant processes.
The Particle Accelerator Simulation Code PyORBIT
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gorlov, Timofey V; Holmes, Jeffrey A; Cousineau, Sarah M
2015-01-01
The particle accelerator simulation code PyORBIT is presented. The structure, implementation, history, parallel and simulation capabilities, and future development of the code are discussed. The PyORBIT code is a new implementation and extension of algorithms of the original ORBIT code that was developed for the Spallation Neutron Source accelerator at the Oak Ridge National Laboratory. The PyORBIT code has a two level structure. The upper level uses the Python programming language to control the flow of intensive calculations performed by the lower level code implemented in the C++ language. The parallel capabilities are based on MPI communications. The PyORBIT ismore » an open source code accessible to the public through the Google Open Source Projects Hosting service.« less
Optimization of the design of Gas Cherenkov Detectors for ICF diagnosis
NASA Astrophysics Data System (ADS)
Liu, Bin; Hu, Huasi; Han, Hetong; Lv, Huanwen; Li, Lan
2018-07-01
A design method, which combines a genetic algorithm (GA) with Monte-Carlo simulation, is established and applied to two different types of Cherenkov detectors, namely, Gas Cherenkov Detector (GCD) and Gamma Reaction History (GRH). For accelerating the optimization program, open Message Passing Interface (MPI) is used in the Geant4 simulation. Compared with the traditional optical ray-tracing method, the performances of these detectors have been improved with the optimization method. The efficiency for GCD system, with a threshold of 6.3 MeV, is enhanced by ∼20% and time response improved by ∼7.2%. For the GRH system, with threshold of 10 MeV, the efficiency is enhanced by ∼76% in comparison with previously published results.
NASA Astrophysics Data System (ADS)
Eckert, C. H. J.; Zenker, E.; Bussmann, M.; Albach, D.
2016-10-01
We present an adaptive Monte Carlo algorithm for computing the amplified spontaneous emission (ASE) flux in laser gain media pumped by pulsed lasers. With the design of high power lasers in mind, which require large size gain media, we have developed the open source code HASEonGPU that is capable of utilizing multiple graphic processing units (GPUs). With HASEonGPU, time to solution is reduced to minutes on a medium size GPU cluster of 64 NVIDIA Tesla K20m GPUs and excellent speedup is achieved when scaling to multiple GPUs. Comparison of simulation results to measurements of ASE in Y b 3 + : Y AG ceramics show perfect agreement.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Computational Research Division, Lawrence Berkeley National Laboratory; NERSC, Lawrence Berkeley National Laboratory; Computer Science Department, University of California, Berkeley
2009-05-04
We apply auto-tuning to a hybrid MPI-pthreads lattice Boltzmann computation running on the Cray XT4 at National Energy Research Scientific Computing Center (NERSC). Previous work showed that multicore-specific auto-tuning can improve the performance of lattice Boltzmann magnetohydrodynamics (LBMHD) by a factor of 4x when running on dual- and quad-core Opteron dual-socket SMPs. We extend these studies to the distributed memory arena via a hybrid MPI/pthreads implementation. In addition to conventional auto-tuning at the local SMP node, we tune at the message-passing level to determine the optimal aspect ratio as well as the correct balance between MPI tasks and threads permore » MPI task. Our study presents a detailed performance analysis when moving along an isocurve of constant hardware usage: fixed total memory, total cores, and total nodes. Overall, our work points to approaches for improving intra- and inter-node efficiency on large-scale multicore systems for demanding scientific applications.« less
Accelerating list management for MPI.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hemmert, K. Scott; Rodrigues, Arun F.; Underwood, Keith Douglas
2005-07-01
The latency and throughput of MPI messages are critically important to a range of parallel scientific applications. In many modern networks, both of these performance characteristics are largely driven by the performance of a processor on the network interface. Because of the semantics of MPI, this embedded processor is forced to traverse a linked list of posted receives each time a message is received. As this list grows long, the latency of message reception grows and the throughput of MPI messages decreases. This paper presents a novel hardware feature to handle list management functions on a network interface. By movingmore » functions such as list insertion, list traversal, and list deletion to the hardware unit, latencies are decreased by up to 20% in the zero length queue case with dramatic improvements in the presence of long queues. Similarly, the throughput is increased by up to 10% in the zero length queue case and by nearly 100% in the presence queues of 30 messages.« less
Porting the AVS/Express scientific visualization software to Cray XT4.
Leaver, George W; Turner, Martin J; Perrin, James S; Mummery, Paul M; Withers, Philip J
2011-08-28
Remote scientific visualization, where rendering services are provided by larger scale systems than are available on the desktop, is becoming increasingly important as dataset sizes increase beyond the capabilities of desktop workstations. Uptake of such services relies on access to suitable visualization applications and the ability to view the resulting visualization in a convenient form. We consider five rules from the e-Science community to meet these goals with the porting of a commercial visualization package to a large-scale system. The application uses message-passing interface (MPI) to distribute data among data processing and rendering processes. The use of MPI in such an interactive application is not compatible with restrictions imposed by the Cray system being considered. We present details, and performance analysis, of a new MPI proxy method that allows the application to run within the Cray environment yet still support MPI communication required by the application. Example use cases from materials science are considered.
Andreu, Yolanda; Galdon, Maria J; Durá, Estrella; Ferrando, Maite; Pascual, Juan; Turk, Dennis C; Jiménez, Yolanda; Poveda, Rafael
2006-01-01
Background This paper seeks to analyse the psychometric and structural properties of the Multidimensional Pain Inventory (MPI) in a sample of temporomandibular disorder patients. Methods The internal consistency of the scales was obtained. Confirmatory Factor Analysis was carried out to test the MPI structure section by section in a sample of 114 temporomandibular disorder patients. Results Nearly all scales obtained good reliability indexes. The original structure could not be totally confirmed. However, with a few adjustments we obtained a satisfactory structural model of the MPI which was slightly different from the original: certain items and the Self control scale were eliminated; in two cases, two original scales were grouped in one factor, Solicitous and Distracting responses on the one hand, and Social activities and Away from home activities, on the other. Conclusion The MPI has been demonstrated to be a reliable tool for the assessment of pain in temporomandibular disorder patients. Some divergences to be taken into account have been clarified. PMID:17169143
Projection x-space magnetic particle imaging.
Goodwill, Patrick W; Konkle, Justin J; Zheng, Bo; Saritas, Emine U; Conolly, Steven M
2012-05-01
Projection magnetic particle imaging (MPI) can improve imaging speed by over 100-fold over traditional 3-D MPI. In this work, we derive the 2-D x-space signal equation, 2-D image equation, and introduce the concept of signal fading and resolution loss for a projection MPI imager. We then describe the design and construction of an x-space projection MPI scanner with a field gradient of 2.35 T/m across a 10 cm magnet free bore. The system has an expected resolution of 3.5 × 8.0 mm using Resovist tracer, and an experimental resolution of 3.8 × 8.4 mm resolution. The system images 2.5 cm × 5.0 cm partial field-of views (FOVs) at 10 frames/s, and acquires a full field-of-view of 10 cm × 5.0 cm in 4 s. We conclude by imaging a resolution phantom, a complex "Cal" phantom, mice injected with Resovist tracer, and experimentally confirm the theoretically predicted x-space spatial resolution.
Dhavalikar, R; Hensley, D; Maldonado-Camargo, L; Croft, L R; Ceron, S; Goodwill, P W; Conolly, S M; Rinaldi, C
2016-08-03
Magnetic Particle Imaging (MPI) is an emerging tomographic imaging technology that detects magnetic nanoparticle tracers by exploiting their non-linear magnetization properties. In order to predict the behavior of nanoparticles in an imager, it is possible to use a non-imaging MPI relaxometer or spectrometer to characterize the behavior of nanoparticles in a controlled setting. In this paper we explore the use of ferrohydrodynamic magnetization equations for predicting the response of particles in an MPI relaxometer. These include a magnetization equation developed by Shliomis (Sh) which has a constant relaxation time and a magnetization equation which uses a field-dependent relaxation time developed by Martsenyuk, Raikher and Shliomis (MRSh). We compare the predictions from these models with measurements and with the predictions based on the Langevin function that assumes instantaneous magnetization response of the nanoparticles. The results show good qualitative and quantitative agreement between the ferrohydrodynamic models and the measurements without the use of fitting parameters and provide further evidence of the potential of ferrohydrodynamic modeling in MPI.
Gayed, Isis; Gohar, Salman; Liao, Zhongxing; McAleer, Mary; Bassett, Roland; Yusuf, Syed Wamique
2009-06-01
This study aims to identify the clinical implications of myocardial perfusion defects after chemoradiation therapy (CRT) in patients with esophageal and lung cancer. We retrospectively compared myocardial perfusion imaging (MPI) results before and after CRT in 16 patients with esophageal cancer and 24 patients with lung cancer. New MPI defects in the radiation therapy (RT) fields were considered related to RT. Follow-up to evaluate for cardiac complications and their relation with the results of MPI was performed. Statistical analysis identified predictors of cardiac morbidities. Eleven females and twenty nine males at a mean age of 66.7 years were included. Five patients (31%) with esophageal cancer and seven patients (29%) with lung cancer developed myocardial ischemia in the RT field at mean intervals of 7.0 and 8.4 months after RT. The patients were followed-up for mean intervals of 15 and 23 months in the esophageal and lung cancer groups, respectively. Seven patients in each of the esophageal (44%) and lung (29%) cancer patients (P = 0.5) developed cardiac complications of which one patient with esophageal cancer died of complete heart block. Six out of the fourteen patients (43%) with cardiac complication had new ischemia on MPI after CRT of which only one developed angina. The remaining eight patients with cardiac complications had normal MPI results. MPI result was not a statistically significant predictor of future cardiac complications after CRT. A history of congestive heart failure (CHF) (P = 0.003) or arrhythmia (P = 0.003) is a significant predictor of cardiac morbidity after CRT in univariate analysis but marginal predictors when multivariate analysis was performed (P = 0.06 and 0.06 for CHF and arrhythmia, respectively). Cardiac complications after CRT are more common in esophageal than lung cancer patients but the difference is not statistically significant. MPI abnormalities are frequently seen after CRT but are not predictive of future cardiac complications. A history of arrhythmia or CHF is significantly associated with cardiac complications after CRT.
Hardware Architectures for Data-Intensive Computing Problems: A Case Study for String Matching
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tumeo, Antonino; Villa, Oreste; Chavarría-Miranda, Daniel
DNA analysis is an emerging application of high performance bioinformatic. Modern sequencing machinery are able to provide, in few hours, large input streams of data, which needs to be matched against exponentially growing databases of known fragments. The ability to recognize these patterns effectively and fastly may allow extending the scale and the reach of the investigations performed by biology scientists. Aho-Corasick is an exact, multiple pattern matching algorithm often at the base of this application. High performance systems are a promising platform to accelerate this algorithm, which is computationally intensive but also inherently parallel. Nowadays, high performance systems alsomore » include heterogeneous processing elements, such as Graphic Processing Units (GPUs), to further accelerate parallel algorithms. Unfortunately, the Aho-Corasick algorithm exhibits large performance variability, depending on the size of the input streams, on the number of patterns to search and on the number of matches, and poses significant challenges on current high performance software and hardware implementations. An adequate mapping of the algorithm on the target architecture, coping with the limit of the underlining hardware, is required to reach the desired high throughputs. In this paper, we discuss the implementation of the Aho-Corasick algorithm for GPU-accelerated high performance systems. We present an optimized implementation of Aho-Corasick for GPUs and discuss its tradeoffs on the Tesla T10 and he new Tesla T20 (codename Fermi) GPUs. We then integrate the optimized GPU code, respectively, in a MPI-based and in a pthreads-based load balancer to enable execution of the algorithm on clusters and large sharedmemory multiprocessors (SMPs) accelerated with multiple GPUs.« less
Specialized Computer Systems for Environment Visualization
NASA Astrophysics Data System (ADS)
Al-Oraiqat, Anas M.; Bashkov, Evgeniy A.; Zori, Sergii A.
2018-06-01
The need for real time image generation of landscapes arises in various fields as part of tasks solved by virtual and augmented reality systems, as well as geographic information systems. Such systems provide opportunities for collecting, storing, analyzing and graphically visualizing geographic data. Algorithmic and hardware software tools for increasing the realism and efficiency of the environment visualization in 3D visualization systems are proposed. This paper discusses a modified path tracing algorithm with a two-level hierarchy of bounding volumes and finding intersections with Axis-Aligned Bounding Box. The proposed algorithm eliminates the branching and hence makes the algorithm more suitable to be implemented on the multi-threaded CPU and GPU. A modified ROAM algorithm is used to solve the qualitative visualization of reliefs' problems and landscapes. The algorithm is implemented on parallel systems—cluster and Compute Unified Device Architecture-networks. Results show that the implementation on MPI clusters is more efficient than Graphics Processing Unit/Graphics Processing Clusters and allows real-time synthesis. The organization and algorithms of the parallel GPU system for the 3D pseudo stereo image/video synthesis are proposed. With realizing possibility analysis on a parallel GPU-architecture of each stage, 3D pseudo stereo synthesis is performed. An experimental prototype of a specialized hardware-software system 3D pseudo stereo imaging and video was developed on the CPU/GPU. The experimental results show that the proposed adaptation of 3D pseudo stereo imaging to the architecture of GPU-systems is efficient. Also it accelerates the computational procedures of 3D pseudo-stereo synthesis for the anaglyph and anamorphic formats of the 3D stereo frame without performing optimization procedures. The acceleration is on average 11 and 54 times for test GPUs.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Golovanov, Georgy
The thesis is devoted to the study of processes with multiple parton interactions (MPI) in a ppbar collision collected by D0 detector at the Fermilab Tevatron collider at sqrt(s) = 1.96 TeV. The study includes measurements of MPI event fraction and effective cross section, a process-independent parameter related to the effective interaction region inside the nucleon. The measurements are done using events with a photon and three hadronic jets in the final state. The measured effective cross section is used to estimate background from MPI for WH production at the Tevatron energy
A Generator-Produced Gallium-68 Radiopharmaceutical for PET Imaging of Myocardial Perfusion
Sharma, Vijay; Sivapackiam, Jothilingam; Harpstrite, Scott E.; Prior, Julie L.; Gu, Hannah; Rath, Nigam P.; Piwnica-Worms, David
2014-01-01
Lipophilic cationic technetium-99m-complexes are widely used for myocardial perfusion imaging (MPI). However, inherent uncertainties in the supply chain of molybdenum-99, the parent isotope required for manufacturing 99Mo/99mTc generators, intensifies the need for discovery of novel MPI agents incorporating alternative radionuclides. Recently, germanium/gallium (Ge/Ga) generators capable of producing high quality 68Ga, an isotope with excellent emission characteristics for clinical PET imaging, have emerged. Herein, we report a novel 68Ga-complex identified through mechanism-based cell screening that holds promise as a generator-produced radiopharmaceutical for PET MPI. PMID:25353349
Hybrid cloud and cluster computing paradigms for life science applications
2010-01-01
Background Clouds and MapReduce have shown themselves to be a broadly useful approach to scientific computing especially for parallel data intensive applications. However they have limited applicability to some areas such as data mining because MapReduce has poor performance on problems with an iterative structure present in the linear algebra that underlies much data analysis. Such problems can be run efficiently on clusters using MPI leading to a hybrid cloud and cluster environment. This motivates the design and implementation of an open source Iterative MapReduce system Twister. Results Comparisons of Amazon, Azure, and traditional Linux and Windows environments on common applications have shown encouraging performance and usability comparisons in several important non iterative cases. These are linked to MPI applications for final stages of the data analysis. Further we have released the open source Twister Iterative MapReduce and benchmarked it against basic MapReduce (Hadoop) and MPI in information retrieval and life sciences applications. Conclusions The hybrid cloud (MapReduce) and cluster (MPI) approach offers an attractive production environment while Twister promises a uniform programming environment for many Life Sciences applications. Methods We used commercial clouds Amazon and Azure and the NSF resource FutureGrid to perform detailed comparisons and evaluations of different approaches to data intensive computing. Several applications were developed in MPI, MapReduce and Twister in these different environments. PMID:21210982
Hybrid cloud and cluster computing paradigms for life science applications.
Qiu, Judy; Ekanayake, Jaliya; Gunarathne, Thilina; Choi, Jong Youl; Bae, Seung-Hee; Li, Hui; Zhang, Bingjing; Wu, Tak-Lon; Ruan, Yang; Ekanayake, Saliya; Hughes, Adam; Fox, Geoffrey
2010-12-21
Clouds and MapReduce have shown themselves to be a broadly useful approach to scientific computing especially for parallel data intensive applications. However they have limited applicability to some areas such as data mining because MapReduce has poor performance on problems with an iterative structure present in the linear algebra that underlies much data analysis. Such problems can be run efficiently on clusters using MPI leading to a hybrid cloud and cluster environment. This motivates the design and implementation of an open source Iterative MapReduce system Twister. Comparisons of Amazon, Azure, and traditional Linux and Windows environments on common applications have shown encouraging performance and usability comparisons in several important non iterative cases. These are linked to MPI applications for final stages of the data analysis. Further we have released the open source Twister Iterative MapReduce and benchmarked it against basic MapReduce (Hadoop) and MPI in information retrieval and life sciences applications. The hybrid cloud (MapReduce) and cluster (MPI) approach offers an attractive production environment while Twister promises a uniform programming environment for many Life Sciences applications. We used commercial clouds Amazon and Azure and the NSF resource FutureGrid to perform detailed comparisons and evaluations of different approaches to data intensive computing. Several applications were developed in MPI, MapReduce and Twister in these different environments.
Advances in PET myocardial perfusion imaging: F-18 labeled tracers.
Rischpler, Christoph; Park, Min-Jae; Fung, George S K; Javadi, Mehrbod; Tsui, Benjamin M W; Higuchi, Takahiro
2012-01-01
Coronary artery disease and its related cardiac disorders represent the most common cause of death in the USA and Western world. Despite advancements in treatment and accompanying improvements in outcome with current diagnostic and therapeutic modalities, it is the correct assignment of these diagnostic techniques and treatment options which are crucial. From a diagnostic standpoint, SPECT myocardial perfusion imaging (MPI) using traditional radiotracers like thallium-201 chloride, Tc-99m sestamibi or Tc-99m tetrofosmin is the most utilized imaging technique. However, PET MPI using N-13 ammonia, rubidium-82 chloride or O-15 water is increasing in availability and usage as a result of the growing number of medical centers with new-generation PET/CT systems taking advantage of the superior imaging properties of PET over SPECT. The routine clinical use of PET MPI is still limited, in part because of the short half-life of conventional PET MPI tracers. The disadvantages of these conventional PET tracers include expensive onsite production and inconvenient on-scanner tracer administration making them unsuitable for physical exercise stress imaging. Recently, two F-18 labeled radiotracers with longer radioactive half-lives than conventional PET imaging agents have been introduced. These are flurpiridaz F 18 (formerly known as F-18 BMS747158-02) and F-18 fluorobenzyltriphenylphosphonium. These longer half-life F-18 labeled perfusion tracers can overcome the production and protocol limitations of currently used radiotracers for PET MPI.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gowda, Giri; Sagurthi, Someswar Rao; Savithri, H. S.
2008-02-01
The cloning, expression, purification, crystallization and preliminary X-ray crystallographic studies of mannose 6-phosphate isomerase from S. typhimurium are reported. Mannose 6-phosphate isomerase (MPI; EC 5.3.1.8) catalyzes the reversible isomerization of d-mannose 6-phosphate (M6P) and d-fructose 6-phosphate (F6P). In the eukaryotes and prokaryotes investigated to date, the enzyme has been reported to play a crucial role in d-mannose metabolism and supply of the activated mannose donor guanosine diphosphate d-mannose (GDP-d-mannose). In the present study, MPI was cloned from Salmonella typhimurium, overexpressed in Escherichia coli and purified using Ni–NTA affinity column chromatography. Purified MPI crystallized in space group P2{sub 1}2{sub 1}2{sub 1},more » with unit-cell parameters a = 36.03, b = 92.2, c = 111.01 Å. A data set extending to 1.66 Å resolution was collected with 98.8% completeness using an image-plate detector system mounted on a rotating-anode X-ray generator. The asymmetric unit of the crystal cell was compatible with the presence of a monomer of MPI. A preliminary structure solution of the enzyme has been obtained by molecular replacement using Candida albicans MPI as the phasing model and the program Phaser. Further refinement and model building are in progress.« less
Nakatani, Akiho; Li, Xuan; Miyamoto, Junki; Igarashi, Miki; Watanabe, Hitoshi; Sutou, Asuka; Watanabe, Keita; Motoyama, Takayasu; Tachibana, Nobuhiko; Kohno, Mitsutaka; Inoue, Hiroshi; Kimura, Ikuo
2018-07-02
The 8-globulin-rich mung bean protein (MPI) suppresses hepatic lipogenesis in rodent models and reduces fasting plasma glucose and insulin levels in obese adults. However, its effects on mitigating high fat diet (HFD)-induced obesity and the mechanism underlying these effects remain to be elucidated. Herein, we examined the metabolic phenotype, intestinal bile acid (BA) pool, and gut microbiota of conventionally raised (CONV-R) male C57BL/6 mice and germ-free (GF) mice that were randomized to receive either regular HFD or HFD containing mung bean protein isolate (MPI) instead of the dairy protein present in regular HFD. MPI intake significantly reduced HFD-induced weight gain and adipose tissue accumulation, and attenuated hepatic steatosis. Enhancement in the secretion of intestinal glucagon-like peptide-1 (GLP-1) and an enlarged cecal and fecal BA pool of dramatically elevated secondary/primary BA ratio were observed in mice that had consumed MPI. These effects were abolished in GF mice, indicating that the effects were dependent upon the presence of the microbiota. As revealed by 16S rRNA gene sequence analysis, MPI intake also elicited dramatic changes in the gut microbiome, such as an expansion of taxa belonging to the phylum Bacteroidetes along with a reduced abundance of the Firmicutes. Copyright © 2018 Elsevier Inc. All rights reserved.
Ortiz, Javier U; Torres, Ximena; Eixarch, Elisenda; Bennasar, Mar; Cruz-Lemini, Monica; Gómez, Olga; Lobmaier, Silvia M; Martínez, Josep M; Gratacós, Eduard; Crispi, Fatima
2018-01-19
To evaluate left myocardial performance index (MPI) and time intervals in fetuses with twin-to-twin transfusion syndrome (TTTS) before and after laser surgery. Fifty-one fetal pairs with TTTS and 47 uncomplicated monochorionic twin pairs were included. Left ventricular isovolumetric contraction time (ICT), ejection time (ET), and isovolumetric relaxation time (IRT) were measured using conventional Doppler. Recipients showed prolonged ICT (46 ± 12 vs. 31 ± 8 vs. 30 ± 5 ms; p < 0.001) and IRT (51 ± 9 vs. 43 ± 8 vs. 43 ± 5 ms; p < 0.001) and higher MPI (0.57 ± 0.12 vs. 0.47 ± 0.09 vs. 0.44 ± 0.05; p < 0.001) than donors and controls. Donors showed shorter ET than recipients and controls (157 ± 12 vs. 169 ± 10 vs. 168 ± 10 ms; p < 0.001) and higher MPI than controls (0.47 ± 0.09 vs. 0.44 ± 0.05; p = 0.006). Preoperative MPI changes were observed in all TTTS stages. Time intervals partially improved after surgery. Donor and recipient twins had higher MPI due to different changes in the time intervals, possibly reflecting the state of hypovolemia in the donor and hypervolemia and pressure overload in the recipient. © 2018 S. Karger AG, Basel.
NASA Astrophysics Data System (ADS)
Brovkin, Victor; Lorenz, Stephan; Raddatz, Thomas
2017-04-01
Plants influence climate through changes in the land surface biophysics (albedo, transpiration) and concentrations of the atmospheric greenhouse gases. One of the interesting periods to investigate a climatic role of terrestrial biosphere is the Holocene, when, despite of the relatively steady global climate, the atmospheric CO2 grew by about 20 ppm from 7 kyr BP to pre-industrial. We use a new setup of the Max Planck Institute Earth System Model MPI-ESM1 consisting of the latest version of the atmospheric model ECHAM6, including the land surface model JSBACH3 with carbon cycle and vegetation dynamics, coupled to the ocean circulation model MPI-OM, which includes the HAMOCC model of ocean biogeochemistry. The model has been run for several simulations over the Holocene period of the last 8000 years under the forcing data sets of orbital insolation, atmospheric greenhouse gases, volcanic aerosols, solar irradiance and stratospheric ozone, as well as land-use changes. In response to this forcing, the land carbon storage increased by about 60 PgC between 8 and 4 kyr BP, stayed relatively constant until 2 kyr BP, and decreased by about 90 PgC by 1850 AD due to land use changes. Vegetation and soil carbon changes significantly affected atmospheric CO2 during the periods of strong volcanic eruptions. In response to the eruption-caused cooling, the land initially stores more carbon as respiration decreases, but then it releases even more carbon due to productivity decrease. This decadal- scale variability helps to quantify the vegetation and land carbon feedbacks during the past periods when the temporal resolution of the ice-core CO2 record is not sufficient to capture fast CO2 variations. From a set of Holocene simulations with prescribed or interactive atmospheric CO2, we get estimates of climate-carbon feedback useful for future climate studies. Members of the Hamburg Holocene Team: Jürgen Bader1, Sebastian Bathiany2, Victor Brovkin1, Martin Claussen1,3, Traute Crüger1, Roberta D'agostino1, Anne Dallmeyer1, Sabine Egerer1, Vivienne Groner1, Matthias Heinze1, Tatiana Ilyina1, Johann Jungclaus1, Thomas Kleinen1, Alexander Lemburg1, Stephan Lorenz1, Thomas Raddatz1, Hauke Schmidt1, Gerhard Schmiedl3, Bjorn Stevens1, Claudia Timmreck1, Matthew Toohey4 1Max-Planck-Institut für Meteorologie, D 2Wageningen University, NL 3CEN, Universität Hamburg, D 4GEOMAR Helmholtz Zentrum für Ozeanforschung Kiel, D
Climate Projections over Mediterranean Basin under RCP8.5 and RCP4.5 emission scenarios
NASA Astrophysics Data System (ADS)
Ilhan, Asli; Ünal, Yurdanur S.
2017-04-01
Climate Projections over Mediterranean Basin under RCP8.5 and RCP4.5 emission scenarios A. ILHAN ve Y. S. UNAL Istanbul Technical University, Department of Meteorology In the study, 50 km resolution downscaled results of two different Earth System Models (ESM) HadGEM2-ES and MPI-ESM with regional climate model of RegCM are used to estimate present and future climate conditions over Mediterranean Basin. The purpose of this study is to compare the projections of two ESMs under Representative Concentration Pathways 4.5 (RCP4.5) and 8.5 (RCP8.5) over the region of interest seasonally and annually with 50 km resolution. Temperature and precipitation parameters for reference period (1971-2000) and future (2015-2100) are analyzed. The average temperature and total precipitation distributions of each downscaled ESM simulations were compared with observation data (Climate Research Unit-CRU data) to explore the capability of each model for the representation of the current climate. According to reference period values of CRU, HadGEM2-ES and MPI-ESM, it is seen that both models are warmer and wetter than observations and have positive temperature biases only around Caspian Sea and positive precipitation biases over Eastern and Central Europe. The future projections (from 2015 to 2100) of HadGEM2-ES and MPI-ESM-MR simulations under RCP4.5 and RCP8.5 emission scenarios are compared with reference period (from 1971 to 2000) and analyzed for temperature and precipitation parameters. The downscaled HadGEM2-ES forced by RCP8.5 scenario produces higher temperatures than the MPI-ESM-MR. The reasons of this warming can be sensitivity of HadGEM2-ES to greenhouse gases and high radiative forcing (+8.5 W/m2). On the other hand, MPI-ESM produce more precipitation than HadGEM2-ES. In order to analyze regional responses of the climate model chains, five main regions are selected which are Turkey, Central Europe, Western Europe, Eastern Europe and North Africa. The average biases of the HadGEM2-ES+RegCM and MPI-ESM-MR+RegCM model chains are also calculated for temperature and precipitation variables, and future expectations in each region are discussed under RCP4.5 and RCP8.5 scenarios. According to the regional analysis, North Africa is the warmest region for HadGEM2-ES and MPI-ESM-MR, and Central Europe warms up similar to North Africa in MPI-ESM-MR coupled simulations under both RCPs. In addition, Eastern Europe is expected to be the wettest region in both models and in both emission scenarios. On the other hand, the driest conditions are expected over Western Europe for MPI-ESM-MR and over Turkey for HadGEM2-ES under RCPs.
Sachse, Carola
2011-01-01
The Max Planck Institute for Molecular Genetics (MPIMG) in Berlin-Dahlem dates its establishment to 1964. Its homepage makes no mention of its predecessor institutes, the Kaiser Wilhelm Institute for Anthropology, Human Genetics and Eugenics (KWIA) and the subsequent MPI for Comparative Genetics and Hereditary Pathology (MPIVEE). This article traces the two critical phases of transition regarding the constellations of academic staff, institutional and epistemic ruptures and continuities specific to the era. Only one of the five department heads from the final war years, Hans Nachtsheim, remained a researcher within the Max Planck Society (MPG); he nevertheless continued to advocate the pre-war and wartime eugenic agenda in the life sciences and social policy. The generational change of 1959/60 became a massive struggle within the institute, in which microbial genetics (with Fritz Kaudewitz) was pitted against human genetics (with Friedrich Vogel) and managed to establish itself after a fresh change in personnel in 1964/65. For the Dahlem institute, this involved a far-reaching reorientation of its research, but for the genetically oriented life sciences in the Max Planck Society as a whole it only meant that molecular biology, which was already being pursued in the West German institutes, gained an additional facility. With this realignment of research traditions, the Society was able to draw a line under the Nazi past without having to address it head-on.
78 FR 9793 - Airworthiness Directives; Bell Helicopter Textron Helicopters
Federal Register 2010, 2011, 2012, 2013, 2014
2013-02-12
...-numbered main rotor hub inboard strap fittings (fittings). This AD requires magnetic particle inspecting..., data, or views. We also invite comments relating to the economic, environmental, energy, or federalism..., perform a magnetic particle inspection (MPI) of each fitting for a crack. If an MPI was already performed...
Students' Performance on a Symmetry Task
ERIC Educational Resources Information Center
Ho, Siew Yin; Logan, Tracy
2013-01-01
This paper describes Singapore and Australian Grade 6 students' (n=1,187) performance on a symmetry task in a recently developed Mathematics Processing Instrument (MPI). The MPI comprised tasks sourced from Australia and Singapore's national assessments, NAPLAN and PSLE. Only half of the cohort solved the item successfully. It is possible that…
WinHPC System Programming | High-Performance Computing | NREL
Programming WinHPC System Programming Learn how to build and run an MPI (message passing interface (mpi.h) and library (msmpi.lib) are. To build from the command line, run... Start > Intel Software Development Tools > Intel C++ Compiler Professional... > C++ Build Environment for applications running
Federal Register 2010, 2011, 2012, 2013, 2014
2012-07-16
... Leased Workers From Echelon Service Company, Sun Associated Industries, INC., MPI Consultants LLC... International, including on-site leased workers from Echelon Service Company, Sun Associated Industries, Inc... Company, Sun Associated Industries, Inc., MPI Consultants LLC, Alliance Engineering, Inc., Washington...
MiKlip-PRODEF: Probabilistic Decadal Forecast for Central and Western Europe
NASA Astrophysics Data System (ADS)
Reyers, Mark; Haas, Rabea; Ludwig, Patrick; Pinto, Joaquim
2013-04-01
The demand for skilful climate predictions on time-scales of several years to decades has increased in recent years, in particular for economic, societal and political terms. Within the BMBF MiKlip consortium, a decadal prediction system on the global to local scale is currently being developed. The subproject PRODEF is part of the MiKlip-Module C, which aims at the regionalisation of decadal predictability for Central and Western Europe. In PRODEF, a combined statistical-dynamical downscaling (SDD) and a probabilistic forecast tool are developed and applied to the new Earth system model of the Max-Planck Institute Hamburg (MPI-ESM), which is part of the CMIP5 experiment. Focus is given on the decadal predictability of windstorms, related wind gusts as well as wind energy potentials. SDD combines the benefits of both high resolution dynamical downscaling and purely statistical downscaling of GCM output. Hence, the SDD approach is used to obtain a very large ensemble of highly resolved decadal forecasts. With respect to the focal points of PRODEF, a clustering of temporal evolving atmospheric fields, a circulation weather type (CWT) analysis, and a storm damage indices analysis is applied to the full ensemble of the decadal hindcast experiments of the MPI-ESM in its lower resolution (MPI-ESM-LR). The ensemble consists of up to ten realisations per yearly initialised decadal hindcast experiments for the period 1960-2010 (altogether 287 realisations). Representatives of CWTs / clusters and single storm episodes are dynamical downscaled with the regional climate model COSMO-CLM with a horizontal resolution of 0.22°. For each model grid point, the distributions of the local climate parameters (e.g. surface wind gusts) are determined for different periods (e.g. each decades) by recombining dynamical downscaled episodes weighted with the respective weather type frequencies. The applicability of the SDD approach is illustrated with examples of decadal forecasts of the MPI-ESM. We are able to perform a bias correction of the frequencies of large scale weather types and to quantify the uncertainties of decadal predictability on large and local scale arising from different initial conditions. Further, probability density functions of local parameters like e.g. wind gusts for different periods and decades derived from the SDD approach is compared to observations and reanalysis data. Skill scores are used to quantify the decadal predictability for different leading time periods and to analyse whether the SDD approach shows systematic errors for some regions.
Kjaergaard, Thomas; Baudin, Pablo; Bykov, Dmytro; ...
2016-11-16
Here, we present a scalable cross-platform hybrid MPI/OpenMP/OpenACC implementation of the Divide–Expand–Consolidate (DEC) formalism with portable performance on heterogeneous HPC architectures. The Divide–Expand–Consolidate formalism is designed to reduce the steep computational scaling of conventional many-body methods employed in electronic structure theory to linear scaling, while providing a simple mechanism for controlling the error introduced by this approximation. Our massively parallel implementation of this general scheme has three levels of parallelism, being a hybrid of the loosely coupled task-based parallelization approach and the conventional MPI +X programming model, where X is either OpenMP or OpenACC. We demonstrate strong and weak scalabilitymore » of this implementation on heterogeneous HPC systems, namely on the GPU-based Cray XK7 Titan supercomputer at the Oak Ridge National Laboratory. Using the “resolution of the identity second-order Moller–Plesset perturbation theory” (RI-MP2) as the physical model for simulating correlated electron motion, the linear-scaling DEC implementation is applied to 1-aza-adamantane-trione (AAT) supramolecular wires containing up to 40 monomers (2440 atoms, 6800 correlated electrons, 24 440 basis functions and 91 280 auxiliary functions). This represents the largest molecular system treated at the MP2 level of theory, demonstrating an efficient removal of the scaling wall pertinent to conventional quantum many-body methods.« less
García-Borreguero, Diego; Allen, Richard P; Kohnen, Ralf; Högl, Birgit; Trenkwalder, Claudia; Oertel, Wolfgang; Hening, Wayne A; Paulus, Walter; Rye, David; Walters, Arthur; Winkelmann, Juliane; Earley, Christopher J
2007-08-01
Augmentation of symptom severity is the main complication of dopaminergic treatment of restless legs syndrome (RLS). The current article reports on the considerations of augmentation that were made during a European Restless Legs Syndrome Study Group (EURLSSG)-sponsored Consensus Conference in April 2006 at the Max Planck Institute (MPI) in Munich, Germany, the conclusions of which were endorsed by the International RLS Study Group (IRLSSG) and the World Association of Sleep Medicine (WASM). The Consensus Conference sought to develop a better understanding of augmentation and generate a better operational definition for its clinical identification. Current concepts of the pathophysiology, clinical features, and therapy of RLS augmentation were evaluated by subgroups who presented a summary of their findings for general consideration and discussion. Recent data indicating sensitivity and specificity of augmentation features for identification of augmentation were also evaluated. The diagnostic criteria of augmentation developed at the National Institutes of Health (NIH) conference in 2002 were reviewed in light of current data and theoretical understanding of augmentation. The diagnostic value and criteria for each of the accepted features of augmentation were considered by the group. A consensus was then developed for a revised statement of the diagnostic criteria for augmentation. Five major diagnostic features of augmentation were identified: usual time of RLS symptom onset each day, number of body parts with RLS symptoms, latency to symptoms at rest, severity of the symptoms when they occur, and effects of dopaminergic medication on symptoms. The quantitative data available relating the time of RLS onset and the presence of other features indicated optimal augmentation criteria of either a 4-h advance in usual starting time for RLS symptoms or a combination of the occurrence of other features. A paradoxical response to changes in medication dose also indicates augmentation. Clinical significance of augmentation is defined. The Consensus Conference agreed upon new operational criteria for the clinical diagnosis of RLS augmentation: the MPI diagnostic criteria for augmentation. Areas needing further consideration for validating these criteria and for understanding the underlying biology of RLS augmentation are indicated.
Overby, Casey Lynnette; Pathak, Jyotishman; Gottesman, Omri; Haerian, Krystl; Perotte, Adler; Murphy, Sean; Bruce, Kevin; Johnson, Stephanie; Talwalkar, Jayant; Shen, Yufeng; Ellis, Steve; Kullo, Iftikhar; Chute, Christopher; Friedman, Carol; Bottinger, Erwin; Hripcsak, George; Weng, Chunhua
2013-01-01
Objective To describe a collaborative approach for developing an electronic health record (EHR) phenotyping algorithm for drug-induced liver injury (DILI). Methods We analyzed types and causes of differences in DILI case definitions provided by two institutions—Columbia University and Mayo Clinic; harmonized two EHR phenotyping algorithms; and assessed the performance, measured by sensitivity, specificity, positive predictive value, and negative predictive value, of the resulting algorithm at three institutions except that sensitivity was measured only at Columbia University. Results Although these sites had the same case definition, their phenotyping methods differed by selection of liver injury diagnoses, inclusion of drugs cited in DILI cases, laboratory tests assessed, laboratory thresholds for liver injury, exclusion criteria, and approaches to validating phenotypes. We reached consensus on a DILI phenotyping algorithm and implemented it at three institutions. The algorithm was adapted locally to account for differences in populations and data access. Implementations collectively yielded 117 algorithm-selected cases and 23 confirmed true positive cases. Discussion Phenotyping for rare conditions benefits significantly from pooling data across institutions. Despite the heterogeneity of EHRs and varied algorithm implementations, we demonstrated the portability of this algorithm across three institutions. The performance of this algorithm for identifying DILI was comparable with other computerized approaches to identify adverse drug events. Conclusions Phenotyping algorithms developed for rare and complex conditions are likely to require adaptive implementation at multiple institutions. Better approaches are also needed to share algorithms. Early agreement on goals, data sources, and validation methods may improve the portability of the algorithms. PMID:23837993
Performance Evaluation of Remote Memory Access (RMA) Programming on Shared Memory Parallel Computers
NASA Technical Reports Server (NTRS)
Jin, Hao-Qiang; Jost, Gabriele; Biegel, Bryan A. (Technical Monitor)
2002-01-01
The purpose of this study is to evaluate the feasibility of remote memory access (RMA) programming on shared memory parallel computers. We discuss different RMA based implementations of selected CFD application benchmark kernels and compare them to corresponding message passing based codes. For the message-passing implementation we use MPI point-to-point and global communication routines. For the RMA based approach we consider two different libraries supporting this programming model. One is a shared memory parallelization library (SMPlib) developed at NASA Ames, the other is the MPI-2 extensions to the MPI Standard. We give timing comparisons for the different implementation strategies and discuss the performance.
Scalable NIC-based reduction on large-scale clusters
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moody, A.; Fernández, J. C.; Petrini, F.
2003-01-01
Many parallel algorithms require effiaent support for reduction mllectives. Over the years, researchers have developed optimal reduction algonduns by taking inm account system size, dam size, and complexities of reduction operations. However, all of these algorithm have assumed the faa that the reduction precessing takes place on the host CPU. Modem Network Interface Cards (NICs) sport programmable processors with substantial memory and thus introduce a fresh variable into the equation This raises the following intersting challenge: Can we take advantage of modern NICs to implementJost redudion operations? In this paper, we take on this challenge in the context of large-scalemore » clusters. Through experiments on the 960-node, 1920-processor or ASCI Linux Cluster (ALC) located at the Lawrence Livermore National Laboratory, we show that NIC-based reductions indeed perform with reduced latency and immed consistency over host-based aleorithms for the wmmon case and that these benefits scale as the system grows. In the largest configuration tested--1812 processors-- our NIC-based algorithm can sum a single element vector in 73 ps with 32-bi integers and in 118 with Mbit floating-point numnbers. These results represent an improvement, respeaively, of 121% and 39% with resvect w the {approx}roductionle vel MPI library« less
Azad, Ariful; Ouzounis, Christos A; Kyrpides, Nikos C; Buluç, Aydin
2018-01-01
Abstract Biological networks capture structural or functional properties of relevant entities such as molecules, proteins or genes. Characteristic examples are gene expression networks or protein–protein interaction networks, which hold information about functional affinities or structural similarities. Such networks have been expanding in size due to increasing scale and abundance of biological data. While various clustering algorithms have been proposed to find highly connected regions, Markov Clustering (MCL) has been one of the most successful approaches to cluster sequence similarity or expression networks. Despite its popularity, MCL’s scalability to cluster large datasets still remains a bottleneck due to high running times and memory demands. Here, we present High-performance MCL (HipMCL), a parallel implementation of the original MCL algorithm that can run on distributed-memory computers. We show that HipMCL can efficiently utilize 2000 compute nodes and cluster a network of ∼70 million nodes with ∼68 billion edges in ∼2.4 h. By exploiting distributed-memory environments, HipMCL clusters large-scale networks several orders of magnitude faster than MCL and enables clustering of even bigger networks. HipMCL is based on MPI and OpenMP and is freely available under a modified BSD license. PMID:29315405
BiCluE - Exact and heuristic algorithms for weighted bi-cluster editing of biomedical data
2013-01-01
Background The explosion of biological data has dramatically reformed today's biology research. The biggest challenge to biologists and bioinformaticians is the integration and analysis of large quantity of data to provide meaningful insights. One major problem is the combined analysis of data from different types. Bi-cluster editing, as a special case of clustering, which partitions two different types of data simultaneously, might be used for several biomedical scenarios. However, the underlying algorithmic problem is NP-hard. Results Here we contribute with BiCluE, a software package designed to solve the weighted bi-cluster editing problem. It implements (1) an exact algorithm based on fixed-parameter tractability and (2) a polynomial-time greedy heuristics based on solving the hardest part, edge deletions, first. We evaluated its performance on artificial graphs. Afterwards we exemplarily applied our implementation on real world biomedical data, GWAS data in this case. BiCluE generally works on any kind of data types that can be modeled as (weighted or unweighted) bipartite graphs. Conclusions To our knowledge, this is the first software package solving the weighted bi-cluster editing problem. BiCluE as well as the supplementary results are available online at http://biclue.mpi-inf.mpg.de. PMID:24565035
Azad, Ariful; Pavlopoulos, Georgios A.; Ouzounis, Christos A.; ...
2018-01-05
Biological networks capture structural or functional properties of relevant entities such as molecules, proteins or genes. Characteristic examples are gene expression networks or protein–protein interaction networks, which hold information about functional affinities or structural similarities. Such networks have been expanding in size due to increasing scale and abundance of biological data. While various clustering algorithms have been proposed to find highly connected regions, Markov Clustering (MCL) has been one of the most successful approaches to cluster sequence similarity or expression networks. Despite its popularity, MCL’s scalability to cluster large datasets still remains a bottleneck due to high running times andmore » memory demands. In this paper, we present High-performance MCL (HipMCL), a parallel implementation of the original MCL algorithm that can run on distributed-memory computers. We show that HipMCL can efficiently utilize 2000 compute nodes and cluster a network of ~70 million nodes with ~68 billion edges in ~2.4 h. By exploiting distributed-memory environments, HipMCL clusters large-scale networks several orders of magnitude faster than MCL and enables clustering of even bigger networks. Finally, HipMCL is based on MPI and OpenMP and is freely available under a modified BSD license.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Azad, Ariful; Pavlopoulos, Georgios A.; Ouzounis, Christos A.
Biological networks capture structural or functional properties of relevant entities such as molecules, proteins or genes. Characteristic examples are gene expression networks or protein–protein interaction networks, which hold information about functional affinities or structural similarities. Such networks have been expanding in size due to increasing scale and abundance of biological data. While various clustering algorithms have been proposed to find highly connected regions, Markov Clustering (MCL) has been one of the most successful approaches to cluster sequence similarity or expression networks. Despite its popularity, MCL’s scalability to cluster large datasets still remains a bottleneck due to high running times andmore » memory demands. In this paper, we present High-performance MCL (HipMCL), a parallel implementation of the original MCL algorithm that can run on distributed-memory computers. We show that HipMCL can efficiently utilize 2000 compute nodes and cluster a network of ~70 million nodes with ~68 billion edges in ~2.4 h. By exploiting distributed-memory environments, HipMCL clusters large-scale networks several orders of magnitude faster than MCL and enables clustering of even bigger networks. Finally, HipMCL is based on MPI and OpenMP and is freely available under a modified BSD license.« less
Chapter 13. Exploring Use of the Reserved Core
DOE Office of Scientific and Technical Information (OSTI.GOV)
Holmen, John; Humphrey, Alan; Berzins, Martin
2015-07-29
In this chapter, we illustrate benefits of thinking in terms of thread management techniques when using a centralized scheduler model along with interoperability of MPI and PThread. This is facilitated through an exploration of thread placement strategies for an algorithm modeling radiative heat transfer with special attention to the 61st core. This algorithm plays a key role within the Uintah Computational Framework (UCF) and current efforts taking place at the University of Utah to model next-generation, large-scale clean coal boilers. In such simulations, this algorithm models the dominant form of heat transfer and consumes a large portion of compute time.more » Exemplified by a real-world example, this chapter presents our early efforts in porting a key portion of a scalability-centric codebase to the Intel Xeon Phi coprocessor. Specifically, this chapter presents results from our experiments profiling the native execution of a reverse Monte-Carlo ray tracing-based radiation model on a single coprocessor. These results demonstrate that our fastest run configurations utilized the 61st core and that performance was not profoundly impacted when explicitly oversubscribing the coprocessor operating system thread. Additionally, this chapter presents a portion of radiation model source code, a MIC-centric UCF cross-compilation example, and less conventional thread management technique for developers utilizing the PThreads threading model.« less
Applying Bayesian Item Selection Approaches to Adaptive Tests Using Polytomous Items
ERIC Educational Resources Information Center
Penfield, Randall D.
2006-01-01
This study applied the maximum expected information (MEI) and the maximum posterior-weighted information (MPI) approaches of computer adaptive testing item selection to the case of a test using polytomous items following the partial credit model. The MEI and MPI approaches are described. A simulation study compared the efficiency of ability…
Optimising the Parallelisation of OpenFOAM Simulations
2014-06-01
UNCLASSIFIED UNCLASSIFIED Optimising the Parallelisation of OpenFOAM Simulations Shannon Keough Maritime Division Defence...Science and Technology Organisation DSTO-TR-2987 ABSTRACT The OpenFOAM computational fluid dynamics toolbox allows parallel computation of...performance of a given high performance computing cluster with several OpenFOAM cases, running using a combination of MPI libraries and corresponding MPI
32 CFR 637.2 - Use of MPI and DAC Detectives/Investigators.
Code of Federal Regulations, 2011 CFR
2011-07-01
... employed in the following investigations: (a) Offenses for which the maximum punishment listed in the Table of Maximum Punishment, Manual for Courts-Martial, United States, 2002 is confinement for 1 year or... MPI. The same punishment criteria apply. (b) Property-related offenses when the value is less than $1...
32 CFR 637.2 - Use of MPI and DAC Detectives/Investigators.
Code of Federal Regulations, 2014 CFR
2014-07-01
... employed in the following investigations: (a) Offenses for which the maximum punishment listed in the Table of Maximum Punishment, Manual for Courts-Martial, United States, 2002 is confinement for 1 year or... MPI. The same punishment criteria apply. (b) Property-related offenses when the value is less than $1...
32 CFR 637.2 - Use of MPI and DAC Detectives/Investigators.
Code of Federal Regulations, 2013 CFR
2013-07-01
... employed in the following investigations: (a) Offenses for which the maximum punishment listed in the Table of Maximum Punishment, Manual for Courts-Martial, United States, 2002 is confinement for 1 year or... MPI. The same punishment criteria apply. (b) Property-related offenses when the value is less than $1...
32 CFR 637.2 - Use of MPI and DAC Detectives/Investigators.
Code of Federal Regulations, 2012 CFR
2012-07-01
... employed in the following investigations: (a) Offenses for which the maximum punishment listed in the Table of Maximum Punishment, Manual for Courts-Martial, United States, 2002 is confinement for 1 year or... MPI. The same punishment criteria apply. (b) Property-related offenses when the value is less than $1...
Tardif, Keith D; Rogers, Aaron; Cassiano, Jared; Roth, Bruce L; Cimbora, Daniel M; McKinnon, Rena; Peterson, Ashley; Douce, Thomas B; Robinson, Rosann; Dorweiler, Irene; Davis, Thaylon; Hess, Mark A; Ostanin, Kirill; Papac, Damon I; Baichwal, Vijay; McAlexander, Ian; Willardsen, J Adam; Saunders, Michael; Christophe, Hoarau; Kumar, D Vijay; Wettstein, Daniel A; Carlson, Robert O; Williams, Brandi L
2011-12-01
Mps1 is a dual specificity protein kinase that is essential for the bipolar attachment of chromosomes to the mitotic spindle and for maintaining the spindle assembly checkpoint until all chromosomes are properly attached. Mps1 is expressed at high levels during mitosis and is abundantly expressed in cancer cells. Disruption of Mps1 function induces aneuploidy and cell death. We report the identification of MPI-0479605, a potent and selective ATP competitive inhibitor of Mps1. Cells treated with MPI-0479605 undergo aberrant mitosis, resulting in aneuploidy and formation of micronuclei. In cells with wild-type p53, this promotes the induction of a postmitotic checkpoint characterized by the ATM- and RAD3-related-dependent activation of the p53-p21 pathway. In both wild-type and p53 mutant cells lines, there is a growth arrest and inhibition of DNA synthesis. Subsequently, cells undergo mitotic catastrophe and/or an apoptotic response. In xenograft models, MPI-0479605 inhibits tumor growth, suggesting that drugs targeting Mps1 may have utility as novel cancer therapeutics.
Dhavalikar, R; Hensley, D; Maldonado-Camargo, L; Croft, L R; Ceron, S; Goodwill, P W; Conolly, S M
2016-01-01
Magnetic Particle Imaging (MPI) is an emerging tomographic imaging technology that detects magnetic nanoparticle tracers by exploiting their non-linear magnetization properties. In order to predict the behavior of nanoparticles in an imager, it is possible to use a non-imaging MPI relaxometer or spectrometer to characterize the behavior of nanoparticles in a controlled setting. In this paper we explore the use of ferrohydrodynamic magnetization equations for predicting the response of particles in an MPI relaxometer. These include a magnetization equation developed by Shliomis (Sh) which has a constant relaxation time and a magnetization equation which uses a field-dependent relaxation time developed by Martsenyuk, Raikher and Shliomis (MRSh). We compare the predictions from these models with measurements and with the predictions based on the Langevin function that assumes instantaneous magnetization response of the nanoparticles. The results show good qualitative and quantitative agreement between the ferrohydrodynamic models and the measurements without the use of fitting parameters and provide further evidence of the potential of ferrohydrodynamic modeling in MPI. PMID:27867219
Shehzad, Danish; Bozkuş, Zeki
2016-01-01
Increase in complexity of neuronal network models escalated the efforts to make NEURON simulation environment efficient. The computational neuroscientists divided the equations into subnets amongst multiple processors for achieving better hardware performance. On parallel machines for neuronal networks, interprocessor spikes exchange consumes large section of overall simulation time. In NEURON for communication between processors Message Passing Interface (MPI) is used. MPI_Allgather collective is exercised for spikes exchange after each interval across distributed memory systems. The increase in number of processors though results in achieving concurrency and better performance but it inversely affects MPI_Allgather which increases communication time between processors. This necessitates improving communication methodology to decrease the spikes exchange time over distributed memory systems. This work has improved MPI_Allgather method using Remote Memory Access (RMA) by moving two-sided communication to one-sided communication, and use of recursive doubling mechanism facilitates achieving efficient communication between the processors in precise steps. This approach enhanced communication concurrency and has improved overall runtime making NEURON more efficient for simulation of large neuronal network models.
Bozkuş, Zeki
2016-01-01
Increase in complexity of neuronal network models escalated the efforts to make NEURON simulation environment efficient. The computational neuroscientists divided the equations into subnets amongst multiple processors for achieving better hardware performance. On parallel machines for neuronal networks, interprocessor spikes exchange consumes large section of overall simulation time. In NEURON for communication between processors Message Passing Interface (MPI) is used. MPI_Allgather collective is exercised for spikes exchange after each interval across distributed memory systems. The increase in number of processors though results in achieving concurrency and better performance but it inversely affects MPI_Allgather which increases communication time between processors. This necessitates improving communication methodology to decrease the spikes exchange time over distributed memory systems. This work has improved MPI_Allgather method using Remote Memory Access (RMA) by moving two-sided communication to one-sided communication, and use of recursive doubling mechanism facilitates achieving efficient communication between the processors in precise steps. This approach enhanced communication concurrency and has improved overall runtime making NEURON more efficient for simulation of large neuronal network models. PMID:27413363
Multipath interference test method for distributed amplifiers
NASA Astrophysics Data System (ADS)
Okada, Takahiro; Aida, Kazuo
2005-12-01
A method for testing distributed amplifiers is presented; the multipath interference (MPI) is detected as a beat spectrum between the multipath signal and the direct signal using a binary frequency shifted keying (FSK) test signal. The lightwave source is composed of a DFB-LD that is directly modulated by a pulse stream passing through an equalizer, and emits the FSK signal of the frequency deviation of about 430MHz at repetition rate of 80-100 kHz. The receiver consists of a photo-diode and an electrical spectrum analyzer (ESA). The base-band power spectrum peak appeared at the frequency of the FSK frequency deviation can be converted to amount of MPI using a calibration chart. The test method has improved the minimum detectable MPI as low as -70 dB, compared to that of -50 dB of the conventional test method. The detailed design and performance of the proposed method are discussed, including the MPI simulator for calibration procedure, computer simulations for evaluating the error caused by the FSK repetition rate and the fiber length under test and experiments on singlemode fibers and distributed Raman amplifier.
Glauber gluons and multiple parton interactions
NASA Astrophysics Data System (ADS)
Gaunt, Jonathan R.
2014-07-01
We show that for hadronic transverse energy E T in hadron-hadron collisions, the classic Collins-Soper-Sterman (CSS) argument for the cancellation of Glauber gluons breaks down at the level of two Glauber gluons exchanged between the spectators. Through an argument that relates the diagrams with these Glauber gluons to events containing additional soft scatterings, we suggest that this failure of the CSS cancellation actually corresponds to a failure of the `standard' factorisation formula with hard, soft and collinear functions to describe E T at leading power. This is because the observable receives a leading power contribution from multiple parton interaction (or spectator-spectator Glauber) processes. We also suggest that the same argument can be used to show that a whole class of observables, which we refer to as MPI sensitive observables, do not obey the standard factorisation at leading power. MPI sensitive observables are observables whose distributions in hadron-hadron collisions are disrupted strongly by the presence of multiple parton interactions (MPI) in the event. Examples of further MPI sensitive observables include the beam thrust B {/a, b +} and transverse thrust.
Sánchez-Ayala, Alfonso; Vilanova, Larissa Soares Reis; Costa, Marina Abrantes; Farias-Neto, Arcelino
2014-01-01
The aim of this study was to evaluate the reproducibility of the condensation silicone Optosil Comfort® as an artificial test food for masticatory performance evaluation. Twenty dentate subjects with mean age of 23.3±0.7 years were selected. Masticatory performance was evaluated using the simple (MPI), the double (IME) and the multiple sieve methods. Trials were carried out five times by three examiners: three times by the first, and once by the second and third examiners. Friedman's test was used to find the differences among time trials. Reproducibility was determined by the intra-class correlation (ICC) test (α=0.05). No differences among time trials were found, except for MPI-4 mm (p=0.022) from the first examiner results. The intra-examiner reproducibility (ICC) of almost all data was high (ICC≥0.92, p<0.001), being moderate only for MPI-0.50 mm (ICC=0.89, p<0.001). The inter-examiner reproducibility was high (ICC>0.93, p<0.001) for all results. For the multiple sieve method, the average mean of absolute difference from repeated measurements were lower than 1 mm. This trend was observed only from MPI-0.50 to MPI-1.4 for the single sieve method, and from IME-0.71/0.50 to IME-1.40/1.00 for the double sieve method. The results suggest that regardless of the method used, the reproducibility of Optosil Comfort® is high.
Chen, Chun; Li, Dianfu; Miao, Changqing; Feng, Jianlin; Zhou, Yanli; Cao, Kejiang; Lloyd, Michael S; Chen, Ji
2012-07-01
The purpose of this study was to evaluate left ventricular (LV) mechanical dyssynchrony in patients with Wolff-Parkinson-White (WPW) syndrome pre- and post-radiofrequency catheter ablation (RFA) using phase analysis of gated single photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI). Forty-five WPW patients were enrolled and had gated SPECT MPI pre- and 2-3 days post-RFA. Electrophysiological study (EPS) was used to locate accessory pathways (APs) and categorize the patients according to the AP locations (septal, left and right free wall). Electrocardiography (ECG) was performed pre- and post-RFA to confirm successful elimination of the APs. Phase analysis of gated SPECT MPI was used to assess LV dyssynchrony pre- and post-RFA. Among the 45 patients, 3 had gating errors, and thus 42 had SPECT phase analysis. Twenty-two patients (52.4%) had baseline LV dyssynchrony. Baseline LV dyssynchrony was more prominent in the patients with septal APs than in the patients with left or right APs (p < 0.05). RFA improved LV synchrony in the entire cohort and in the patients with septal APs (p < 0.01). Phase analysis of gated SPECT MPI demonstrated that LV mechanical dyssynchrony can be present in patients with WPW syndrome. Septal APs result in the greatest degree of LV mechanical dyssynchrony and afford the most benefit after RFA. This study supports further investigation in the relationship between electrical and mechanical activation using EPS and phase analysis of gated SPECT MPI.
A satellite simulator for TRMM PR applied to climate model simulations
NASA Astrophysics Data System (ADS)
Spangehl, T.; Schroeder, M.; Bodas-Salcedo, A.; Hollmann, R.; Riley Dellaripa, E. M.; Schumacher, C.
2017-12-01
Climate model simulations have to be compared against observation based datasets in order to assess their skill in representing precipitation characteristics. Here we use a satellite simulator for TRMM PR in order to evaluate simulations performed with MPI-ESM (Earth system model of the Max Planck Institute for Meteorology in Hamburg, Germany) performed within the MiKlip project (https://www.fona-miklip.de/, funded by Federal Ministry of Education and Research in Germany). While classical evaluation methods focus on geophysical parameters such as precipitation amounts, the application of the satellite simulator enables an evaluation in the instrument's parameter space thereby reducing uncertainties on the reference side. The CFMIP Observation Simulator Package (COSP) provides a framework for the application of satellite simulators to climate model simulations. The approach requires the introduction of sub-grid cloud and precipitation variability. Radar reflectivities are obtained by applying Mie theory, with the microphysical assumptions being chosen to match the atmosphere component of MPI-ESM (ECHAM6). The results are found to be sensitive to the methods used to distribute the convective precipitation over the sub-grid boxes. Simple parameterization methods are used to introduce sub-grid variability of convective clouds and precipitation. In order to constrain uncertainties a comprehensive comparison with sub-grid scale convective precipitation variability which is deduced from TRMM PR observations is carried out.
Korosoglou, G; Hansen, A; Bekeredjian, R; Filusch, A; Hardt, S; Wolf, D; Schellberg, D; Katus, H A; Kuecherer, H
2006-01-01
Objective To evaluate whether myocardial parametric imaging (MPI) is superior to visual assessment for the evaluation of myocardial viability. Methods and results Myocardial contrast echocardiography (MCE) was assessed in 11 pigs before, during, and after left anterior descending coronary artery occlusion and in 32 patients with ischaemic heart disease by using intravenous SonoVue administration. In experimental studies perfusion defect area assessment by MPI was compared with visually guided perfusion defect planimetry. Histological assessment of necrotic tissue was the standard reference. In clinical studies viability was assessed on a segmental level by (1) visual analysis of myocardial opacification; (2) quantitative estimation of myocardial blood flow in regions of interest; and (3) MPI. Functional recovery between three and six months after revascularisation was the standard reference. In experimental studies, compared with visually guided perfusion defect planimetry, planimetric assessment of infarct size by MPI correlated more significantly with histology (r2 = 0.92 versus r2 = 0.56) and had a lower intraobserver variability (4% v 15%, p < 0.05). In clinical studies, MPI had higher specificity (66% v 43%, p < 0.05) than visual MCE and good accuracy (81%) for viability detection. It was less time consuming (3.4 (1.6) v 9.2 (2.4) minutes per image, p < 0.05) than quantitative blood flow estimation by regions of interest and increased the agreement between observers interpreting myocardial perfusion (κ = 0.87 v κ = 0.75, p < 0.05). Conclusion MPI is useful for the evaluation of myocardial viability both in animals and in patients. It is less time consuming than quantification analysis by regions of interest and less observer dependent than visual analysis. Thus, strategies incorporating this technique may be valuable for the evaluation of myocardial viability in clinical routine. PMID:15939722
Adaptive Neuron Apoptosis for Accelerating Deep Learning on Large Scale Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Siegel, Charles M.; Daily, Jeffrey A.; Vishnu, Abhinav
Machine Learning and Data Mining (MLDM) algorithms are becoming ubiquitous in {\\em model learning} from the large volume of data generated using simulations, experiments and handheld devices. Deep Learning algorithms -- a class of MLDM algorithms -- are applied for automatic feature extraction, and learning non-linear models for unsupervised and supervised algorithms. Naturally, several libraries which support large scale Deep Learning -- such as TensorFlow and Caffe -- have become popular. In this paper, we present novel techniques to accelerate the convergence of Deep Learning algorithms by conducting low overhead removal of redundant neurons -- {\\em apoptosis} of neurons --more » which do not contribute to model learning, during the training phase itself. We provide in-depth theoretical underpinnings of our heuristics (bounding accuracy loss and handling apoptosis of several neuron types), and present the methods to conduct adaptive neuron apoptosis. We implement our proposed heuristics with the recently introduced TensorFlow and using its recently proposed extension with MPI. Our performance evaluation on two difference clusters -- one connected with Intel Haswell multi-core systems, and other with nVIDIA GPUs -- using InfiniBand, indicates the efficacy of the proposed heuristics and implementations. Specifically, we are able to improve the training time for several datasets by 2-3x, while reducing the number of parameters by 30x (4-5x on average) on datasets such as ImageNet classification. For the Higgs Boson dataset, our implementation improves the accuracy (measured by Area Under Curve (AUC)) for classification from 0.88/1 to 0.94/1, while reducing the number of parameters by 3x in comparison to existing literature, while achieving a 2.44x speedup in comparison to the default (no apoptosis) algorithm.« less
Jung, Jaewoon; Mori, Takaharu; Kobayashi, Chigusa; Matsunaga, Yasuhiro; Yoda, Takao; Feig, Michael; Sugita, Yuji
2015-07-01
GENESIS (Generalized-Ensemble Simulation System) is a new software package for molecular dynamics (MD) simulations of macromolecules. It has two MD simulators, called ATDYN and SPDYN. ATDYN is parallelized based on an atomic decomposition algorithm for the simulations of all-atom force-field models as well as coarse-grained Go-like models. SPDYN is highly parallelized based on a domain decomposition scheme, allowing large-scale MD simulations on supercomputers. Hybrid schemes combining OpenMP and MPI are used in both simulators to target modern multicore computer architectures. Key advantages of GENESIS are (1) the highly parallel performance of SPDYN for very large biological systems consisting of more than one million atoms and (2) the availability of various REMD algorithms (T-REMD, REUS, multi-dimensional REMD for both all-atom and Go-like models under the NVT, NPT, NPAT, and NPγT ensembles). The former is achieved by a combination of the midpoint cell method and the efficient three-dimensional Fast Fourier Transform algorithm, where the domain decomposition space is shared in real-space and reciprocal-space calculations. Other features in SPDYN, such as avoiding concurrent memory access, reducing communication times, and usage of parallel input/output files, also contribute to the performance. We show the REMD simulation results of a mixed (POPC/DMPC) lipid bilayer as a real application using GENESIS. GENESIS is released as free software under the GPLv2 licence and can be easily modified for the development of new algorithms and molecular models. WIREs Comput Mol Sci 2015, 5:310-323. doi: 10.1002/wcms.1220.
Ben-Haim, Simona; Kacperski, Krzysztof; Hain, Sharon; Van Gramberg, Dean; Hutton, Brian F; Erlandsson, Kjell; Sharir, Tali; Roth, Nathaniel; Waddington, Wendy A; Berman, Daniel S; Ell, Peter J
2010-08-01
We compared simultaneous dual-radionuclide (DR) stress and rest myocardial perfusion imaging (MPI) with a novel solid-state cardiac camera and a conventional SPECT camera with separate stress and rest acquisitions. Of 27 consecutive patients recruited, 24 (64.5+/-11.8 years of age, 16 men) were injected with 74 MBq of (201)Tl (rest) and 250 MBq (99m)Tc-MIBI (stress). Conventional MPI acquisition times for stress and rest are 21 min and 16 min, respectively. Rest (201)Tl for 6 min and simultaneous DR 15-min list mode gated scans were performed on a D-SPECT cardiac scanner. In 11 patients DR D-SPECT was performed first and in 13 patients conventional stress (99m)Tc-MIBI SPECT imaging was performed followed by DR D-SPECT. The DR D-SPECT data were processed using a spill-over and scatter correction method. DR D-SPECT images were compared with rest (201)Tl D-SPECT and with conventional SPECT images by visual analysis employing the 17-segment model and a five-point scale (0 normal, 4 absent) to calculate the summed stress and rest scores. Image quality was assessed on a four-point scale (1 poor, 4 very good) and gut activity was assessed on a four-point scale (0 none, 3 high). Conventional MPI studies were abnormal at stress in 17 patients and at rest in 9 patients. In the 17 abnormal stress studies DR D-SPECT MPI showed 113 abnormal segments and conventional MPI showed 93 abnormal segments. In the nine abnormal rest studies DR D-SPECT showed 45 abnormal segments and conventional MPI showed 48 abnormal segments. The summed stress and rest scores on conventional SPECT and DR D-SPECT were highly correlated (r=0.9790 and 0.9694, respectively). The summed scores of rest (201)Tl D-SPECT and DR-DSPECT were also highly correlated (r=0.9968, p<0.0001 for all). In six patients stress perfusion defects were significantly larger on stress DR D-SPECT images, and five of these patients were imaged earlier by D-SPECT than by conventional SPECT. Fast and high-quality simultaneous DR MPI is feasible with D-SPECT in a single imaging session with comparable diagnostic performance and image quality to conventional SPECT and to a separate rest (201)Tl D-SPECT acquisition.
Lyngholm, Ann Marie; Pedersen, Begitte H; Petersen, Lars J
2008-09-01
Intestinal activity at the inferior myocardial wall represents an issue for assessment of myocardial perfusion imaging (MPI) with 99mTc-labelled tracers. The aim of this study was to investigate the effect of time and food on upper abdominal activity in 99mTc-tetrofosmin MPI. The study population consisted of 152 consecutive patients referred for routine MPI. All patients underwent 2-day stress-rest 99mTc-tetrofosmin single-photon emission computed tomography MPI. Before stress testing, patients were randomized in a factorial design to four different regimens. Group A: early scan (image acquisition initiated within 15 min after injection of the tracer) and no food; group B: early scan and food (two pieces of white bread with butter and a minimum of 450 ml of water); group C: late scan (image acquisition 30-60 min after injection of the tracer) and no food; and group D: late and scan with food. Patients underwent standard bicycle exercise or pharmacological stress test. The degree of upper abdominal activity was evaluated by trained observers blinded to the randomization code. The primary endpoint was the proportion of accepted scans in the intention-to-treat population in stress MPI. The results showed statistical significant impact on both time and food on upper abdominal activity. The primary endpoint showed that the acceptance rate improved from 55% in group A to 100% success rate in group D. An early scan reduced the acceptance rate by 30% versus a late scan [hazard ratio 0.70, 95% confidence interval 0.58-0.84; P<0.0001], whereas the addition of food improved the success rate versus no food by 27% (hazard ratio 1.27, 95% confidence interval 1.07-1.51; P=0.006). No significant interaction between food and time was observed. An analysis of accepted scans according to the actual scan time and food consumption confirmed the findings of the intention-to-treat analysis. In addition, similar findings were seen in 116 of 152 patients with a rest MPI (success rate of 53% in group A vs. 96% in group D). A combination of solid food and water administered after injection of the tracer and delayed image acquisition led to significant and clinically relevant decrease of interfering upper abdominal activity in 99mTc-tetrofosmin MPI.
Ramses-GPU: Second order MUSCL-Handcock finite volume fluid solver
NASA Astrophysics Data System (ADS)
Kestener, Pierre
2017-10-01
RamsesGPU is a reimplementation of RAMSES (ascl:1011.007) which drops the adaptive mesh refinement (AMR) features to optimize 3D uniform grid algorithms for modern graphics processor units (GPU) to provide an efficient software package for astrophysics applications that do not need AMR features but do require a very large number of integration time steps. RamsesGPU provides an very efficient C++/CUDA/MPI software implementation of a second order MUSCL-Handcock finite volume fluid solver for compressible hydrodynamics as a magnetohydrodynamics solver based on the constraint transport technique. Other useful modules includes static gravity, dissipative terms (viscosity, resistivity), and forcing source term for turbulence studies, and special care was taken to enhance parallel input/output performance by using state-of-the-art libraries such as HDF5 and parallel-netcdf.
CFD Analysis and Design Optimization Using Parallel Computers
NASA Technical Reports Server (NTRS)
Martinelli, Luigi; Alonso, Juan Jose; Jameson, Antony; Reuther, James
1997-01-01
A versatile and efficient multi-block method is presented for the simulation of both steady and unsteady flow, as well as aerodynamic design optimization of complete aircraft configurations. The compressible Euler and Reynolds Averaged Navier-Stokes (RANS) equations are discretized using a high resolution scheme on body-fitted structured meshes. An efficient multigrid implicit scheme is implemented for time-accurate flow calculations. Optimum aerodynamic shape design is achieved at very low cost using an adjoint formulation. The method is implemented on parallel computing systems using the MPI message passing interface standard to ensure portability. The results demonstrate that, by combining highly efficient algorithms with parallel computing, it is possible to perform detailed steady and unsteady analysis as well as automatic design for complex configurations using the present generation of parallel computers.
A practical guide to replica-exchange Wang—Landau simulations
NASA Astrophysics Data System (ADS)
Vogel, Thomas; Li, Ying Wai; Landau, David P.
2018-04-01
This paper is based on a series of tutorial lectures about the replica-exchange Wang-Landau (REWL) method given at the IX Brazilian Meeting on Simulational Physics (BMSP 2017). It provides a practical guide for the implementation of the method. A complete example code for a model system is available online. In this paper, we discuss the main parallel features of this code after a brief introduction to the REWL algorithm. The tutorial section is mainly directed at users who have written a single-walker Wang–Landau program already but might have just taken their first steps in parallel programming using the Message Passing Interface (MPI). In the last section, we answer “frequently asked questions” from users about the implementation of REWL for different scientific problems.
The influence of hunger on meal to pellet intervals in barred owls
Duke, G.E.; Fuller, M.R.; Huberty, B.J.
1980-01-01
1. Barred owls fed at a sub-maintenance (SM) level had significantly (P < 0.01) longer meal to pellet intervals (MPI)/g eaten/kg body weight (BW) than those fed at an above maintenance (AM) level; MPI/g per kg for owls fed at a maintenance (M) level was intermediate but significantly (P < 0.01) different from both SM and AM.2. During SM feeding, MPI/g per kg gradually increased.3. The proportion of a meal occurring in a pellet was less in “hungry” owls whether losing weight (SM) or gaining (AM) as compared to owls maintaining their normal body weight (M).4. SM fed owls appear to be able to increase digestion time as well as thoroughness of digestion.
Contaminant studies in the Sierra Nevadas
Sparling, Don; Fellers, Gary M.
2002-01-01
1. 1. Barred owls fed at a sub-maintenance (SM) level had significantly (P < 0.01) longer meal to pellet intervals (MPI)/g eaten/kg body weight (BW) than those fed at an above maintenance (AM) level; MPI/g per kg for owls fed at a maintenance (M) level was intermediate but significantly (P < 0.01) different from both SM and AM. 2. 2. During SM feeding, MPI/g per kg gradually increased. 3. 3. The proportion of a meal occurring in a pellet was less in ?hungry? owls whether losing weight (SM) or gaining (AM) as compared to owls maintaining their normal body weight (M). 4. 4. SM fed owls appear to be able to increase digestion time as well as thoroughness of digestion.
Parallel PAB3D: Experiences with a Prototype in MPI
NASA Technical Reports Server (NTRS)
Guerinoni, Fabio; Abdol-Hamid, Khaled S.; Pao, S. Paul
1998-01-01
PAB3D is a three-dimensional Navier Stokes solver that has gained acceptance in the research and industrial communities. It takes as computational domain, a set disjoint blocks covering the physical domain. This is the first report on the implementation of PAB3D using the Message Passing Interface (MPI), a standard for parallel processing. We discuss briefly the characteristics of tile code and define a prototype for testing. The principal data structure used for communication is derived from preprocessing "patching". We describe a simple interface (COMMSYS) for MPI communication, and some general techniques likely to be encountered when working on problems of this nature. Last, we identify levels of improvement from the current version and outline future work.
A Posteriori Error Bounds for the Empirical Interpolation Method
2010-03-18
paramètres (x̄1, x̄2) ≡ µ ∈ DII ≡ [0.4, 0.6]2 et α = 0.1 fixé, les résultats sont similaires au cas d’un seul paramètre (Fig. 2). 1. Introduction...and denote the set of all distinct multi-indices β of dimension P of length I by MPI . The cardinality of MPI is given by card (MPI ) = ( P+I−1 I...operations, and we compute the interpolation errors ‖F (β)(·; τ) − F (β)M (·; τ)‖L∞(Ω), 0 < |β| < p − 1, for all τ ∈ Φ, in O(nΦMN ) ∑p−1 j=0 card (MPj
Murase, Kenya; Konishi, Takashi; Takeuchi, Yuki; Takata, Hiroshige; Saito, Shigeyoshi
2013-07-01
Our purpose in this study was to investigate the behavior of signal harmonics in magnetic particle imaging (MPI) by experimental and simulation studies. In the experimental studies, we made an apparatus for MPI in which both a drive magnetic field (DMF) and a selection magnetic field (SMF) were generated with a Maxwell coil pair. The MPI signals from magnetic nanoparticles (MNPs) were detected with a solenoid coil. The odd- and even-numbered harmonics were calculated by Fourier transformation with or without background subtraction. The particle size of the MNPs was measured by transmission electron microscopy (TEM), dynamic light-scattering, and X-ray diffraction methods. In the simulation studies, the magnetization and particle size distribution of MNPs were assumed to obey the Langevin theory of paramagnetism and a log-normal distribution, respectively. The odd- and even-numbered harmonics were calculated by Fourier transformation under various conditions of DMF and SMF and for three different particle sizes. The behavior of the harmonics largely depended on the size of the MNPs. When we used the particle size obtained from the TEM image, the simulation results were most similar to the experimental results. The similarity between the experimental and simulation results for the even-numbered harmonics was better than that for the odd-numbered harmonics. This was considered to be due to the fact that the odd-numbered harmonics were more sensitive to background subtraction than were the even-numbered harmonics. This study will be useful for a better understanding, optimization, and development of MPI and for designing MNPs appropriate for MPI.
Fiechter, Michael; Ghadri, Jelena R; Wolfrum, Mathias; Kuest, Silke M; Pazhenkottil, Aju P; Nkoulou, Rene N; Herzog, Bernhard A; Gebhard, Cathérine; Fuchs, Tobias A; Gaemperli, Oliver; Kaufmann, Philipp A
2012-03-01
Low yield of invasive coronary angiography and unnecessary coronary interventions have been identified as key cost drivers in cardiology for evaluation of coronary artery disease (CAD). This has fuelled the search for noninvasive techniques providing comprehensive functional and anatomical information on coronary lesions. We have evaluated the impact of implementation of a novel hybrid cadmium-zinc-telluride (CZT)/64-slice CT camera into the daily clinical routine on downstream resource utilization. Sixty-two patients with known or suspected CAD were referred for same-day single-session hybrid evaluation with CZT myocardial perfusion imaging (MPI) and coronary CT angiography (CCTA). Hybrid MPI/CCTA images from the integrated CZT/CT camera served for decision-making towards conservative versus invasive management. Based on the hybrid images patients were classified into those with and those without matched findings. Matched findings were defined as the combination of MPI defect with a stenosis by CCTA in the coronary artery subtending the respective territory. All patients with normal MPI and CCTA as well as those with isolated MPI or CCTA finding or combined but unmatched findings were categorized as "no match". All 23 patients with a matched finding underwent invasive coronary angiography and 21 (91%) were revascularized. Of the 39 patients with no match, 5 (13%, p < 0.001 vs matched) underwent catheterization and 3 (8%, p < 0.001 vs matched) were revascularized. Cardiac hybrid imaging in CAD evaluation has a profound impact on patient management and may contribute to optimal downstream resource utilization.
NASA Astrophysics Data System (ADS)
Çatıkkaş, Berna; Aktan, Ebru; Yalçın, Ergin
2016-08-01
This work deals with the optimized molecular structure, vibrational spectra, nonlinear optic (NLO) and frontier molecule orbital (FMO) properties of 1-Methyl-2-phenyl-3-(1,3,4-thiadiazol-2-yldiazenyl)-1H-indole (MPI) by quantum chemical calculations. The Fourier transform infrared (FT-MIR and FT-FIR) and Raman spectra of 1-Methyl-2-phenyl-3-(1,3,4-thiadiazol-2-yldiazenyl)-1H-indole (MPI) were recorded in the region (4000-400 cm-1 and 400-30 cm-1) and (3200-92 cm-1), respectively. The analysis and complete vibrational assignments of the fundamental modes of the MPI molecule were carried out by using the observed FT-IR and FT-Raman data and calculated Total Energy Distribution (TED) according to Scaled Quantum Mechanics procedure. The calculated geometrical parameters of the MPI molecule are in agreement with the obtained values from XRD studies. On the other hand, the difference between the scaled and observed wavenumber values of the most of the fundamentals are very small. 1H NMR and 13C NMR chemical shift values, and energy gap between LUMO-HOMO and molecular electrostatic potential (MEP) were investigated by using density functional theory (B3LYP) methods. UV/Visible spectra and λ maximum absorption values, the oscillator strengths in the chloroform, methanol and DMSO solvation in combination with different basis sets were calculated by using the time-dependent density functional theory (TD-DFT). Additionally, the predicted nonlinear optical (NLO) properties of the MPI are quite greater than that of urea at the B3LYP/6-31++G(d,p) level.
Zucoloto, Miriane Lucindo; Maroco, João; Duarte Bonini Campos, Juliana Alvares
2015-01-01
To evaluate the psychometric properties of the Multidimensional Pain Inventory (MPI) in a Brazilian sample of patients with orofacial pain. A total of 1,925 adult patients, who sought dental care in the School of Dentistry of São Paulo State University's Araraquara campus, were invited to participate; 62.5% (n=1,203) agreed to participate. Of these, 436 presented with orofacial pain and were included. The mean age was 39.9 (SD=13.6) years and 74.5% were female. Confirmatory factor analysis was conducted using χ²/df, comparative fit index, goodness of fit index, and root mean square error of approximation as indices of goodness of fit. Convergent validity was estimated by the average variance extracted and composite reliability, and internal consistency by Cronbach's alpha standardized coefficient (α). The stability of the models was tested in independent samples (test and validation; dental pain and orofacial pain). The factorial invariance was estimated by multigroup analysis (Δχ²). Factorial, convergent validity, and internal consistency were adequate in all three parts of the MPI. To achieve this adequate fit for Part 1, item 15 needed to be deleted (λ=0.13). Discriminant validity was compromised between the factors "activities outside the home" and "social activities" of Part 3 of the MPI in the total sample, validation sample, and in patients with dental pain and with orofacial pain. A strong invariance between different subsamples from the three parts of the MPI was detected. The MPI produced valid, reliable, and stable data for pain assessment among Brazilian patients with orofacial pain.
Zhou, Yanli; Faber, Tracy L.; Patel, Zenic; Folks, Russell D.; Cheung, Alice A.; Garcia, Ernest V.; Soman, Prem; Li, Dianfu; Cao, Kejiang; Chen, Ji
2013-01-01
Objective Left ventricular (LV) function and dyssynchrony parameters measured from serial gated single-photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI) using blinded processing had a poorer repeatability than when manual side-by-side processing was used. The objective of this study was to validate whether an automatic alignment tool can reduce the variability of LV function and dyssynchrony parameters in serial gated SPECT MPI. Methods Thirty patients who had undergone serial gated SPECT MPI were prospectively enrolled in this study. Thirty minutes after the first acquisition, each patient was repositioned and a gated SPECT MPI image was reacquired. The two data sets were first processed blinded from each other by the same technologist in different weeks. These processed data were then realigned by the automatic tool, and manual side-by-side processing was carried out. All processing methods used standard iterative reconstruction and Butterworth filtering. The Emory Cardiac Toolbox was used to measure the LV function and dyssynchrony parameters. Results The automatic tool failed in one patient, who had a large, severe scar in the inferobasal wall. In the remaining 29 patients, the repeatability of the LV function and dyssynchrony parameters after automatic alignment was significantly improved from blinded processing and was comparable to manual side-by-side processing. Conclusion The automatic alignment tool can be an alternative method to manual side-by-side processing to improve the repeatability of LV function and dyssynchrony measurements by serial gated SPECT MPI. PMID:23211996
Furuhashi, Tatsuhiko; Moroi, Masao; Joki, Nobuhiko; Hase, Hiroki; Masai, Hirofumi; Kunimasa, Taeko; Fukuda, Hiroshi; Sugi, Kaoru
2013-02-01
Pretest probability of coronary artery disease (CAD) facilitates diagnosis and risk stratification of CAD. Stress myocardial perfusion imaging (MPI) and chronic kidney disease (CKD) are established major predictors of cardiovascular events. However, the role of CKD to assess pretest probability of CAD has been unclear. This study evaluates the role of CKD to assess the predictive value of cardiovascular events under consideration of pretest probability in patients who underwent stress MPI. Patients with no history of CAD underwent stress MPI (n = 310; male = 166; age = 70; CKD = 111; low/intermediate/high pretest probability = 17/194/99) and were followed for 24 months. Cardiovascular events included cardiac death and nonfatal acute coronary syndrome. Cardiovascular events occurred in 15 of the 310 patients (4.8 %), but not in those with low pretest probability which included 2 CKD patients. In patients with intermediate to high pretest probability (n = 293), multivariate Cox regression analysis identified only CKD [hazard ratio (HR) = 4.88; P = 0.022) and summed stress score of stress MPI (HR = 1.50; P < 0.001) as independent and significant predictors of cardiovascular events. Cardiovascular events were not observed in patients with low pretest probability. In patients with intermediate to high pretest probability, CKD and stress MPI are independent predictors of cardiovascular events considering the pretest probability of CAD in patients with no history of CAD. In assessing pretest probability of CAD, CKD might be an important factor for assessing future cardiovascular prognosis.
Pascual, Thomas N B; Mercuri, Mathew; El-Haj, Noura; Bom, Henry Hee-Sung; Lele, Vikram; Al-Mallah, Mouaz H; Luxenburg, Osnat; Karthikeyan, Ganesan; Vitola, Joao; Mahmarian, John J; Better, Nathan; Shaw, Leslee J; Rehani, Madan M; Kashyap, Ravi; Paez, Diana; Dondi, Maurizio; Einstein, Andrew J
2017-03-24
This paper examines the current status of radiation exposure to patients in myocardial perfusion imaging (MPI) in Asia.Methods and Results:Laboratories voluntarily provided information on MPI performed over a 1-week period. Eight best practice criteria regarding MPI were predefined by an expert panel. Implementation of ≥6 best practices (quality index [QI] ≥6) was pre-specified as a desirable goal for keeping radiation exposure at a low level. Radiation effective dose (ED) in 1,469 patients and QI of 69 laboratories in Asia were compared against data from 239 laboratories in the rest of the world (RoW). Mean ED was significantly higher in Asia (11.4 vs. 9.6 mSv; P<0.0001), with significantly lower doses in South-East vs. East Asia (9.7 vs. 12.7 mSv; P<0.0001). QI in Asia was lower than in RoW. In comparison with RoW, Asian laboratories used thallium more frequently, used weight-based technetium dosing less frequently, and trended towards a lower rate of stress-only imaging. MPI radiation dose in Asia is higher than that in the RoW and linked to less consistent use of laboratory best practices such as avoidance of thallium, weight-based dosing, and use of stress-only imaging. Given that MPI is performed in Asia within a diverse array of medical contexts, laboratory-specific adoption of best practices offers numerous opportunities to improve quality of care.
SurF: an innovative framework in biosecurity and animal health surveillance evaluation.
Muellner, Petra; Watts, Jonathan; Bingham, Paul; Bullians, Mark; Gould, Brendan; Pande, Anjali; Riding, Tim; Stevens, Paul; Vink, Daan; Stärk, Katharina Dc
2018-05-16
Surveillance for biosecurity hazards is being conducted by the New Zealand Competent Authority, the Ministry for Primary Industries (MPI) to support New Zealand's biosecurity system. Surveillance evaluation should be an integral part of the surveillance life cycle, as it provides a means to identify and correct problems and to sustain and enhance the existing strengths of a surveillance system. The surveillance evaluation Framework (SurF) presented here was developed to provide a generic framework within which the MPI biosecurity surveillance portfolio, and all of its components, can be consistently assessed. SurF is an innovative, cross-sectoral effort that aims to provide a common umbrella for surveillance evaluation in the animal, plant, environment and aquatic sectors. It supports the conduct of the following four distinct components of an evaluation project: (i) motivation for the evaluation, (ii) scope of the evaluation, (iii) evaluation design and implementation and (iv) reporting and communication of evaluation outputs. Case studies, prepared by MPI subject matter experts, are included in the framework to guide users in their assessment. Three case studies were used in the development of SurF in order to assure practical utility and to confirm usability of SurF across all included sectors. It is anticipated that the structured approach and information provided by SurF will not only be of benefit to MPI but also to other New Zealand stakeholders. Although SurF was developed for internal use by MPI, it could be applied to any surveillance system in New Zealand or elsewhere. © 2018 2018 The Authors. Transboundary and Emerging Diseases Published by Blackwell Verlag GmbH.
Chen, Chun; Miao, Changqing; Feng, Jianlin; Zhou, Yanli; Cao, Kejiang; Lloyd, Michael S.; Chen, Ji
2013-01-01
Purpose The purpose of this study was to evaluate left ventricular (LV) mechanical dyssynchrony in patients with Wolff-Parkinson-White (WPW) syndrome pre- and post-radiofrequency catheter ablation (RFA) using phase analysis of gated single photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI). Methods Forty-five WPW patients were enrolled and had gated SPECT MPI pre- and 2–3 days post-RFA. Electrophysiological study (EPS) was used to locate accessory pathways (APs) and categorize the patients according to the AP locations (septal, left and right free wall). Electrocardiography (ECG) was performed pre- and post-RFA to confirm successful elimination of the APs. Phase analysis of gated SPECT MPI was used to assess LV dyssynchrony pre- and post-RFA. Results Among the 45 patients, 3 had gating errors, and thus 42 had SPECT phase analysis. Twenty-two patients (52.4 %) had baseline LV dyssynchrony. Baseline LV dyssynchrony was more prominent in the patients with septal APs than in the patients with left or right APs (p<0.05). RFA improved LV synchrony in the entire cohort and in the patients with septal APs (p<0.01). Conclusion Phase analysis of gated SPECT MPI demonstrated that LV mechanical dyssynchrony can be present in patients with WPW syndrome. Septal APs result in the greatest degree of LV mechanical dyssynchrony and afford the most benefit after RFA. This study supports further investigation in the relationship between electrical and mechanical activation using EPS and phase analysis of gated SPECT MPI. PMID:22532253
2012-03-01
on the standard Navy Handgun Qualification Course. Results partially supported the hypotheses. The simulation group showed greater improvement in MPI...standard Navy Handgun Qualification Course. Results partially supported the hypotheses. The simulation group showed greater improvement in MPI than the...14 3. Navy Handgun Qualification Course Firing Sequence ..................15 F. PROCEDURES
Performance Comparison of HPF and MPI Based NAS Parallel Benchmarks
NASA Technical Reports Server (NTRS)
Saini, Subhash
1997-01-01
Compilers supporting High Performance Form (HPF) features first appeared in late 1994 and early 1995 from Applied Parallel Research (APR), Digital Equipment Corporation, and The Portland Group (PGI). IBM introduced an HPF compiler for the IBM RS/6000 SP2 in April of 1996. Over the past two years, these implementations have shown steady improvement in terms of both features and performance. The performance of various hardware/ programming model (HPF and MPI) combinations will be compared, based on latest NAS Parallel Benchmark results, thus providing a cross-machine and cross-model comparison. Specifically, HPF based NPB results will be compared with MPI based NPB results to provide perspective on performance currently obtainable using HPF versus MPI or versus hand-tuned implementations such as those supplied by the hardware vendors. In addition, we would also present NPB, (Version 1.0) performance results for the following systems: DEC Alpha Server 8400 5/440, Fujitsu CAPP Series (VX, VPP300, and VPP700), HP/Convex Exemplar SPP2000, IBM RS/6000 SP P2SC node (120 MHz), NEC SX-4/32, SGI/CRAY T3E, and SGI Origin2000. We would also present sustained performance per dollar for Class B LU, SP and BT benchmarks.
Pope, Bernard J; Fitch, Blake G; Pitman, Michael C; Rice, John J; Reumann, Matthias
2011-10-01
Future multiscale and multiphysics models that support research into human disease, translational medical science, and treatment can utilize the power of high-performance computing (HPC) systems. We anticipate that computationally efficient multiscale models will require the use of sophisticated hybrid programming models, mixing distributed message-passing processes [e.g., the message-passing interface (MPI)] with multithreading (e.g., OpenMP, Pthreads). The objective of this study is to compare the performance of such hybrid programming models when applied to the simulation of a realistic physiological multiscale model of the heart. Our results show that the hybrid models perform favorably when compared to an implementation using only the MPI and, furthermore, that OpenMP in combination with the MPI provides a satisfactory compromise between performance and code complexity. Having the ability to use threads within MPI processes enables the sophisticated use of all processor cores for both computation and communication phases. Considering that HPC systems in 2012 will have two orders of magnitude more cores than what was used in this study, we believe that faster than real-time multiscale cardiac simulations can be achieved on these systems.
NASA Technical Reports Server (NTRS)
Hockney, George; Lee, Seungwon
2008-01-01
A computer program known as PyPele, originally written as a Pythonlanguage extension module of a C++ language program, has been rewritten in pure Python language. The original version of PyPele dispatches and coordinates parallel-processing tasks on cluster computers and provides a conceptual framework for spacecraft-mission- design and -analysis software tools to run in an embarrassingly parallel mode. The original version of PyPele uses SSH (Secure Shell a set of standards and an associated network protocol for establishing a secure channel between a local and a remote computer) to coordinate parallel processing. Instead of SSH, the present Python version of PyPele uses Message Passing Interface (MPI) [an unofficial de-facto standard language-independent application programming interface for message- passing on a parallel computer] while keeping the same user interface. The use of MPI instead of SSH and the preservation of the original PyPele user interface make it possible for parallel application programs written previously for the original version of PyPele to run on MPI-based cluster computers. As a result, engineers using the previously written application programs can take advantage of embarrassing parallelism without need to rewrite those programs.
Meinel, Felix G; Schoepf, U Joseph; Townsend, Jacob C; Flowers, Brian A; Geyer, Lucas L; Ebersberger, Ullrich; Krazinski, Aleksander W; Kunz, Wolfgang G; Thierfelder, Kolja M; Baker, Deborah W; Khan, Ashan M; Fernandes, Valerian L; O'Brien, Terrence X
2018-06-15
We aimed to determine the diagnostic yield and accuracy of coronary CT angiography (CCTA) in patients referred for invasive coronary angiography (ICA) based on clinical concern for coronary artery disease (CAD) and an abnormal nuclear stress myocardial perfusion imaging (MPI) study. We enrolled 100 patients (84 male, mean age 59.6 ± 8.9 years) with an abnormal MPI study and subsequent referral for ICA. Each patient underwent CCTA prior to ICA. We analyzed the prevalence of potentially obstructive CAD (≥50% stenosis) on CCTA and calculated the diagnostic accuracy of ≥50% stenosis on CCTA for the detection of clinically significant CAD on ICA (defined as any ≥70% stenosis or ≥50% left main stenosis). On CCTA, 54 patients had at least one ≥50% stenosis. With ICA, 45 patients demonstrated clinically significant CAD. A positive CCTA had 100% sensitivity and 84% specificity with a 100% negative predictive value and 83% positive predictive value for clinically significant CAD on a per patient basis in MPI positive symptomatic patients. In conclusion, almost half (48%) of patients with suspected CAD and an abnormal MPI study demonstrate no obstructive CAD on CCTA.
Taupitz, Matthias; Ariza de Schellenberger, Angela; Kosch, Olaf; Eberbeck, Dietmar; Wagner, Susanne; Trahms, Lutz; Hamm, Bernd; Schnorr, Jörg
2018-01-01
Synthesis of novel magnetic multicore particles (MCP) in the nano range, involves alkaline precipitation of iron(II) chloride in the presence of atmospheric oxygen. This step yields green rust, which is oxidized to obtain magnetic nanoparticles, which probably consist of a magnetite/maghemite mixed-phase. Final growth and annealing at 90°C in the presence of a large excess of carboxymethyl dextran gives MCP very promising magnetic properties for magnetic particle imaging (MPI), an emerging medical imaging modality, and magnetic resonance imaging (MRI). The magnetic nanoparticles are biocompatible and thus potential candidates for future biomedical applications such as cardiovascular imaging, sentinel lymph node mapping in cancer patients, and stem cell tracking. The new MCP that we introduce here have three times higher magnetic particle spectroscopy performance at lower and middle harmonics and five times higher MPS signal strength at higher harmonics compared with Resovist®. In addition, the new MCP have also an improved in vivo MPI performance compared to Resovist®, and we here report the first in vivo MPI investigation of this new generation of magnetic nanoparticles. PMID:29300729
Rugged optical mirrors for the operation of Fourier-Transform Spectrometers in rough environments
NASA Astrophysics Data System (ADS)
Feist, Dietrich G.
2014-05-01
The Total Carbon Column Observing Network (TCCON) and the Network for the Detection of Atmospheric Composition Change (NDACC) operate a growing number of Fourier-Transform Spectrometers (FTS) that measure the total column of several atmospheric trace gases. For these measurements, the sun is used as a light source. This is typically achieved by a solar tracker that uses a pair of optical mirrors to guide the sunlight into the instrument. There is a growing demand to operate these instruments in remote locations that fill the gaps in the global observation network. Besides the logistical challenges of running a remote site, the environment at these locations can be very harsh compared to the sheltered environment of the instruments' home institutions. While the FTS itself is usually well protected inside a building or container, the solar tracker and especially its mirrors are exposed to the environment. There they may suffer from - temperature fluctuations - high humidity - sea salt corrosion at coastal sites - dirt and dust - air pollution from anthropogenic sources - deposition from plants or animals The Max Planck Institute for Biogeochemistry (MPI-BGC) operates a TCCON station on Ascension Island, about 200 m from the sea. Under the rough conditions at this site, typical optical mirrors that are made for laboratory conditions are destroyed by sea salt spray within a few weeks. Besides, typical gold-coated mirrors cannot be cleaned as their soft surface is easily scratched or damaged. To overcome these problems, the MPI-BGC has developed optical mirrors that - offer good reflectivity in the near and mid infrared - are highly resistant to salt and chlorine - have a hard surface so that they can be cleaned often and easily - are not affected by organic solvents - last for months in very harsh environments - can be reused after polishing These mirrors could be applied to most TCCON and NDACC sites. This way, the network could be expanded to regions where operation would have been too challenging so far.
Manipulating the motion of large molecules: Information from the molecular frame
NASA Astrophysics Data System (ADS)
Küpper, Jochen
2011-05-01
Large molecules have complex potential-energy surfaces with many local minima. They exhibit multiple stereoisomers, even at the low temperatures (~1 K) in a molecular beam, with rich intra- and intermolecular dynamics. Over the last years, we have developed methods to manipulate the motion of large, complex molecules and to select their quantum states. We have exploited this state-selectivity, for example, to spatially separate individual structural isomers of complex molecules and to demonstrate unprecedented degrees of laser alignment and mixed-field orientation of these molecules. Such clean, well-defined samples strongly benefit, or simply allow, novel experiments on the dynamics of complex molecules, for instance, femtosecond pump-probe measurements, X-ray or electron diffraction of molecular ensembles (including diffraction-from-within experiments), or tomographic reconstructions of molecular orbitals. These samples could also be very advantageous for metrology applications, such as, for example, matter-wave interferometry or the search for electroweak interactions in chiral molecules. Moreover, they provide an extreme level of control for stereo-dynamically controlled reaction dynamics. We have recently exploited these state-selected and oriented samples to measure photoelectron angular distributions in the molecular frame (MFPADs) from non-resonant femtosecond-laser photoionization and using the X-ray Free-Electron-Laser LCLS. We have also investigated X-ray diffraction imaging and, using ion momentum imaging, the induced radiation damage of these samples using the LCLS. This work was carried out within a collaboration for which J. Küpper, H. Chapman, and D. Rolles are spokespersons. The collaboration consists of CFEL (DESY, MPG, University Hamburg), Fritz-Haber-Institute Berlin, MPI Nuclear Physics Heidelberg, MPG Semi-conductor Lab, Aarhus University, FOM AMOLF Amsterdam, Lund University, MPI Medical Research Heidelberg, TU Berlin, Max Born Institute Berlin, and SLAC Menlo Park, CA, USA. The experiments were carried out using CAMP (designed and built by the MPG-ASG at CFEL) at the LCLS (operated by Stanford University on behalf of the US DOE).
NASA Astrophysics Data System (ADS)
Petibon, Yoann; Guehl, Nicolas J.; Reese, Timothy G.; Ebrahimi, Behzad; Normandin, Marc D.; Shoup, Timothy M.; Alpert, Nathaniel M.; El Fakhri, Georges; Ouyang, Jinsong
2017-01-01
PET is an established modality for myocardial perfusion imaging (MPI) which enables quantification of absolute myocardial blood flow (MBF) using dynamic imaging and kinetic modeling. However, heart motion and partial volume effects (PVE) significantly limit the spatial resolution and quantitative accuracy of PET MPI. Simultaneous PET-MR offers a solution to the motion problem in PET by enabling MR-based motion correction of PET data. The aim of this study was to develop a motion and PVE correction methodology for PET MPI using simultaneous PET-MR, and to assess its impact on both static and dynamic PET MPI using 18F-Flurpiridaz, a novel 18F-labeled perfusion tracer. Two dynamic 18F-Flurpiridaz MPI scans were performed on healthy pigs using a PET-MR scanner. Cardiac motion was tracked using a dedicated tagged-MRI (tMR) sequence. Motion fields were estimated using non-rigid registration of tMR images and used to calculate motion-dependent attenuation maps. Motion correction of PET data was achieved by incorporating tMR-based motion fields and motion-dependent attenuation coefficients into image reconstruction. Dynamic and static PET datasets were created for each scan. Each dataset was reconstructed as (i) Ungated, (ii) Gated (end-diastolic phase), and (iii) Motion-Corrected (MoCo), each without and with point spread function (PSF) modeling for PVE correction. Myocardium-to-blood concentration ratios (MBR) and apparent wall thickness were calculated to assess image quality for static MPI. For dynamic MPI, segment- and voxel-wise MBF values were estimated by non-linear fitting of a 2-tissue compartment model to tissue time-activity-curves. MoCo and Gating respectively decreased mean apparent wall thickness by 15.1% and 14.4% and increased MBR by 20.3% and 13.6% compared to Ungated images (P < 0.01). Combined motion and PSF correction (MoCo-PSF) yielded 30.9% (15.7%) lower wall thickness and 82.2% (20.5%) higher MBR compared to Ungated data reconstructed without (with) PSF modeling (P < 0.01). For dynamic PET, mean MBF across all segments were comparable for MoCo (0.72 ± 0.21 ml/min/ml) and Gating (0.69 ± 0.18 ml/min/ml). Ungated data yielded significantly lower mean MBF (0.59 ± 0.16 ml/min/ml). Mean MBF for MoCo-PSF was 0.80 ± 0.22 ml/min/ml, which was 37.9% (25.0%) higher than that obtained from Ungated data without (with) PSF correction (P < 0.01). The developed methodology holds promise to improve the image quality and sensitivity of PET MPI studies performed using PET-MR.
Mechanic: The MPI/HDF code framework for dynamical astronomy
NASA Astrophysics Data System (ADS)
Słonina, Mariusz; Goździewski, Krzysztof; Migaszewski, Cezary
2015-01-01
We introduce the Mechanic, a new open-source code framework. It is designed to reduce the development effort of scientific applications by providing unified API (Application Programming Interface) for configuration, data storage and task management. The communication layer is based on the well-established Message Passing Interface (MPI) standard, which is widely used on variety of parallel computers and CPU-clusters. The data storage is performed within the Hierarchical Data Format (HDF5). The design of the code follows core-module approach which allows to reduce the user’s codebase and makes it portable for single- and multi-CPU environments. The framework may be used in a local user’s environment, without administrative access to the cluster, under the PBS or Slurm job schedulers. It may become a helper tool for a wide range of astronomical applications, particularly focused on processing large data sets, such as dynamical studies of long-term orbital evolution of planetary systems with Monte Carlo methods, dynamical maps or evolutionary algorithms. It has been already applied in numerical experiments conducted for Kepler-11 (Migaszewski et al., 2012) and νOctantis planetary systems (Goździewski et al., 2013). In this paper we describe the basics of the framework, including code listings for the implementation of a sample user’s module. The code is illustrated on a model Hamiltonian introduced by (Froeschlé et al., 2000) presenting the Arnold diffusion. The Arnold web is shown with the help of the MEGNO (Mean Exponential Growth of Nearby Orbits) fast indicator (Goździewski et al., 2008a) applied onto symplectic SABAn integrators family (Laskar and Robutel, 2001).
DISPATCH: a numerical simulation framework for the exa-scale era - I. Fundamentals
NASA Astrophysics Data System (ADS)
Nordlund, Åke; Ramsey, Jon P.; Popovas, Andrius; Küffmeier, Michael
2018-06-01
We introduce a high-performance simulation framework that permits the semi-independent, task-based solution of sets of partial differential equations, typically manifesting as updates to a collection of `patches' in space-time. A hybrid MPI/OpenMP execution model is adopted, where work tasks are controlled by a rank-local `dispatcher' which selects, from a set of tasks generally much larger than the number of physical cores (or hardware threads), tasks that are ready for updating. The definition of a task can vary, for example, with some solving the equations of ideal magnetohydrodynamics (MHD), others non-ideal MHD, radiative transfer, or particle motion, and yet others applying particle-in-cell (PIC) methods. Tasks do not have to be grid based, while tasks that are, may use either Cartesian or orthogonal curvilinear meshes. Patches may be stationary or moving. Mesh refinement can be static or dynamic. A feature of decisive importance for the overall performance of the framework is that time-steps are determined and applied locally; this allows potentially large reductions in the total number of updates required in cases when the signal speed varies greatly across the computational domain, and therefore a corresponding reduction in computing time. Another feature is a load balancing algorithm that operates `locally' and aims to simultaneously minimize load and communication imbalance. The framework generally relies on already existing solvers, whose performance is augmented when run under the framework, due to more efficient cache usage, vectorization, local time-stepping, plus near-linear and, in principle, unlimited OpenMP and MPI scaling.
2013-01-01
Background Molecular imaging using magnetic nanoparticles (MNPs)—magnetic particle imaging (MPI)—has attracted interest for the early diagnosis of cancer and cardiovascular disease. However, because a steep local magnetic field distribution is required to obtain a defined image, sophisticated hardware is required. Therefore, it is desirable to realize excellent image quality even with low-performance hardware. In this study, the spatial resolution of MPI was evaluated using an image reconstruction method based on the correlation information of the magnetization signal in a time domain and by applying MNP samples made from biocompatible ferucarbotran that have adjusted particle diameters. Methods The magnetization characteristics and particle diameters of four types of MNP samples made from ferucarbotran were evaluated. A numerical analysis based on our proposed method that calculates the image intensity from correlation information between the magnetization signal generated from MNPs and the system function was attempted, and the obtained image quality was compared with that using the prototype in terms of image resolution and image artifacts. Results MNP samples obtained by adjusting ferucarbotran showed superior properties to conventional ferucarbotran samples, and numerical analysis showed that the same image quality could be obtained using a gradient magnetic field generator with 0.6 times the performance. However, because image blurring was included theoretically by the proposed method, an algorithm will be required to improve performance. Conclusions MNP samples obtained by adjusting ferucarbotran showed magnetizing properties superior to conventional ferucarbotran samples, and by using such samples, comparable image quality (spatial resolution) could be obtained with a lower gradient magnetic field intensity. PMID:23734917
Myocardial perfusion imaging in patients with a recent, normal exercise test.
Bovin, Ann; Klausen, Ib C; Petersen, Lars J
2013-03-26
To investigate the added value of myocardial perfusion scintigraphy imaging (MPI) in consecutive patients with suspected coronary artery disease (CAD) and a recent, normal exercise electrocardiography (ECG). This study was a retrospective analysis of consecutive patients referred for MPI during a 2-year period from 2006-2007 at one clinic. All eligible patients were suspected of suffering from CAD, and had performed a satisfactory bicycle exercise test (i.e., peak heart rate > 85% of the expected, age-predicted maximum) within 6 mo of referral, their exercise ECG was had no signs of ischemia, there was no exercise-limiting angina, and no cardiac events occurred between the exercise test and referral. The patients subsequently underwent a standard 2-d, stress-rest exercise MPI. Ischemia was defined based on visual scoring supported by quantitative segmental analysis (i.e., sum of stress score > 3). The results of cardiac catheterization were analyzed, and clinical follow up was performed by review of electronic medical files. A total of 56 patients fulfilled the eligibility criteria. Most patients had a low or intermediate ATPIII pre-test risk of CAD (6 patients had a high pre-test risk). The referral exercise test showed a mean Duke score of 5 (range: 2 to 11), which translated to a low post-exercise risk in 66% and intermediate risk in 34%. A total of seven patients were reported with ischemia by MPI. Three of these patients had high ATPIII pre-test risk scores. Six of these seven patients underwent cardiac catheterization, which showed significant stenosis in one patient with a high pre-test risk of CAD, and indeterminate lesions in three patients (two of whom had high pre-test risk scores). With MPI as a gate keeper for catheterization, no significant, epicardial stenosis was observed in any of the 50 patients (0%, 95% confidence interval 0.0 to 7.1) with low to intermediate pre-test risk of CAD and a negative exercise test. No cardiac events occurred in any patients within a median follow up period of > 1200 d. The added diagnostic value of MPI in patients with low or intermediate risk of CAD and a recent, normal exercise test is marginal.
Jung, Jaewoon; Mori, Takaharu; Kobayashi, Chigusa; Matsunaga, Yasuhiro; Yoda, Takao; Feig, Michael; Sugita, Yuji
2015-01-01
GENESIS (Generalized-Ensemble Simulation System) is a new software package for molecular dynamics (MD) simulations of macromolecules. It has two MD simulators, called ATDYN and SPDYN. ATDYN is parallelized based on an atomic decomposition algorithm for the simulations of all-atom force-field models as well as coarse-grained Go-like models. SPDYN is highly parallelized based on a domain decomposition scheme, allowing large-scale MD simulations on supercomputers. Hybrid schemes combining OpenMP and MPI are used in both simulators to target modern multicore computer architectures. Key advantages of GENESIS are (1) the highly parallel performance of SPDYN for very large biological systems consisting of more than one million atoms and (2) the availability of various REMD algorithms (T-REMD, REUS, multi-dimensional REMD for both all-atom and Go-like models under the NVT, NPT, NPAT, and NPγT ensembles). The former is achieved by a combination of the midpoint cell method and the efficient three-dimensional Fast Fourier Transform algorithm, where the domain decomposition space is shared in real-space and reciprocal-space calculations. Other features in SPDYN, such as avoiding concurrent memory access, reducing communication times, and usage of parallel input/output files, also contribute to the performance. We show the REMD simulation results of a mixed (POPC/DMPC) lipid bilayer as a real application using GENESIS. GENESIS is released as free software under the GPLv2 licence and can be easily modified for the development of new algorithms and molecular models. WIREs Comput Mol Sci 2015, 5:310–323. doi: 10.1002/wcms.1220 PMID:26753008
Comparative Implementation of High Performance Computing for Power System Dynamic Simulations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jin, Shuangshuang; Huang, Zhenyu; Diao, Ruisheng
Dynamic simulation for transient stability assessment is one of the most important, but intensive, computations for power system planning and operation. Present commercial software is mainly designed for sequential computation to run a single simulation, which is very time consuming with a single processer. The application of High Performance Computing (HPC) to dynamic simulations is very promising in accelerating the computing process by parallelizing its kernel algorithms while maintaining the same level of computation accuracy. This paper describes the comparative implementation of four parallel dynamic simulation schemes in two state-of-the-art HPC environments: Message Passing Interface (MPI) and Open Multi-Processing (OpenMP).more » These implementations serve to match the application with dedicated multi-processor computing hardware and maximize the utilization and benefits of HPC during the development process.« less
NASA Astrophysics Data System (ADS)
Hobson, Michael; Graff, Philip; Feroz, Farhan; Lasenby, Anthony
2014-05-01
Machine-learning methods may be used to perform many tasks required in the analysis of astronomical data, including: data description and interpretation, pattern recognition, prediction, classification, compression, inference and many more. An intuitive and well-established approach to machine learning is the use of artificial neural networks (NNs), which consist of a group of interconnected nodes, each of which processes information that it receives and then passes this product on to other nodes via weighted connections. In particular, I discuss the first public release of the generic neural network training algorithm, called SkyNet, and demonstrate its application to astronomical problems focusing on its use in the BAMBI package for accelerated Bayesian inference in cosmology, and the identification of gamma-ray bursters. The SkyNet and BAMBI packages, which are fully parallelised using MPI, are available at http://www.mrao.cam.ac.uk/software/.
NASA Astrophysics Data System (ADS)
Bolis, A.; Cantwell, C. D.; Moxey, D.; Serson, D.; Sherwin, S. J.
2016-09-01
A hybrid parallelisation technique for distributed memory systems is investigated for a coupled Fourier-spectral/hp element discretisation of domains characterised by geometric homogeneity in one or more directions. The performance of the approach is mathematically modelled in terms of operation count and communication costs for identifying the most efficient parameter choices. The model is calibrated to target a specific hardware platform after which it is shown to accurately predict the performance in the hybrid regime. The method is applied to modelling turbulent flow using the incompressible Navier-Stokes equations in an axisymmetric pipe and square channel. The hybrid method extends the practical limitations of the discretisation, allowing greater parallelism and reduced wall times. Performance is shown to continue to scale when both parallelisation strategies are used.
Response properties of the refractory auditory nerve fiber.
Miller, C A; Abbas, P J; Robinson, B K
2001-09-01
The refractory characteristics of auditory nerve fibers limit their ability to accurately encode temporal information. Therefore, they are relevant to the design of cochlear prostheses. It is also possible that the refractory property could be exploited by prosthetic devices to improve information transfer, as refractoriness may enhance the nerve's stochastic properties. Furthermore, refractory data are needed for the development of accurate computational models of auditory nerve fibers. We applied a two-pulse forward-masking paradigm to a feline model of the human auditory nerve to assess refractory properties of single fibers. Each fiber was driven to refractoriness by a single (masker) current pulse delivered intracochlearly. Properties of firing efficiency, latency, jitter, spike amplitude, and relative spread (a measure of dynamic range and stochasticity) were examined by exciting fibers with a second (probe) pulse and systematically varying the masker-probe interval (MPI). Responses to monophasic cathodic current pulses were analyzed. We estimated the mean absolute refractory period to be about 330 micros and the mean recovery time constant to be about 410 micros. A significant proportion of fibers (13 of 34) responded to the probe pulse with MPIs as short as 500 micros. Spike amplitude decreased with decreasing MPI, a finding relevant to the development of computational nerve-fiber models, interpretation of gross evoked potentials, and models of more central neural processing. A small mean decrement in spike jitter was noted at small MPI values. Some trends (such as spike latency-vs-MPI) varied across fibers, suggesting that sites of excitation varied across fibers. Relative spread was found to increase with decreasing MPI values, providing direct evidence that stochastic properties of fibers are altered under conditions of refractoriness.
Effect of balloon mitral valvotomy on left ventricular function in rheumatic mitral stenosis.
Rajesh, Gopalan Nair; Sreekumar, Pradeep; Haridasan, Vellani; Sajeev, C G; Bastian, Cicy; Vinayakumar, D; Kadermuneer, P; Mathew, Dolly; George, Biju; Krishnan, M N
Mitral stenosis (MS) is found to produce left ventricular (LV) dysfunction in some studies. We sought to study the left ventricular function in patients with rheumatic MS undergoing balloon mitral valvotomy (BMV). Ours is the first study to analyze effect of BMV on mitral annular plane systolic excursion (MAPSE), and to quantify prevalence of longitudinal left ventricular dysfunction in rheumatic MS. In this prospective cohort study, we included 43 patients with severe rheumatic mitral stenosis undergoing BMV. They were compared to twenty controls whose distribution of age and gender were similar to that of patients. The parameters compared were LV ejection fraction (EF) by modified Simpson's method, mitral annular systolic velocity (MASV), MAPSE, mitral annular early diastolic velocity (E'), and myocardial performance index (MPI). These parameters were reassessed immediately following BMV and after 3 months of procedure. MASV, MAPSE, E', and EF were significantly lower and MPI was higher in mitral stenosis group compared to controls. Impaired longitudinal LV function was present in 77% of study group. MAPSE and EF did not show significant change after BMV while MPI, MASV, and E' improved significantly. MASV and E' showed improvement immediately after BMV, while MPI decreased only at 3 months follow-up. There were significantly lower mitral annular motion parameters including MAPSE in patients with rheumatic mitral stenosis. Those with atrial fibrillation had higher MPI. Immediately after BMV, there was improvement in LV long axis function with a gradual improvement in global LV function. There was no significant change of MAPSE after BMV. Copyright © 2015 Cardiological Society of India. Published by Elsevier B.V. All rights reserved.
Besli, Feyzullah; Basar, Cengiz; Kecebas, Mesut; Turker, Yasin
2015-03-01
This study evaluated the response to electrical cardioversion (EC) and the effect on the myocardial performance index (MPI) in patients with persistent and long-standing persistent atrial fibrillation (AF). We enrolled 103 patients (mean age 69.6 ± 8.9 years, 40.7% males) with a diagnosis of persistent and long-standing persistent AF. EC was applied to all patients after one g of amiodarone administration. Echocardiographic findings before EC were compared in patients with successful versus unsuccessful cardioversions and in patients with maintained sinus rhythm (SR) versus those with AF recurrence at the end of the first month. We also compared echocardiographic data before EC versus at the end of the first month in the same patients with maintained SR. SR was achieved in 72.8% of patients and was continued at the end of the first month in 69.3% of the patients. The MPI value of all patients was found to be 0.73 ± 0.21. The size of the left atrium was determined to be an independent predictor of the maintenance of SR at 1 month. In subgroup analyses, when we compared echocardiographic findings before EC and at the end of the first month in patients with maintained SR, the MPI (0.66 ± 0.14 vs 0.56 ± 0.09, p < 0.001) values were significantly decreased. Our study is the first to show impairment of the MPI, which is an indicator of systolic and diastolic function, in patients with persistent and long-standing persistent AF and improvement of the MPI after successful EC.
Bouyoucef, Salah E; Mercuri, Mathew; Pascual, Thomas N; Allam, Adel H; Vangu, Mboyo; Vitola, João V; Better, Nathan; Karthikeyan, Ganesan; Mahmarian, John J; Rehani, Madan M; Kashyap, Ravi; Dondi, Maurizio; Paez, Diana; Einstein, Andrew J
While nuclear myocardial perfusion imaging (MPI) offers many benefits to patients with known or suspected cardiovascular disease, concerns exist regarding radiation-associated health effects. Little is known regarding MPI practice in Africa. We sought to characterise radiation doses and the use of MPI best practices that could minimise radiation in African nuclear cardiology laboratories, and compare these to practice worldwide. Demographics and clinical characteristics were collected for a consecutive sample of 348 patients from 12 laboratories in six African countries over a one-week period from March to April 2013. Radiation effective dose (ED) was estimated for each patient. A quality index (QI) enumerating adherence to eight best practices, identified a priori by an IAEA expert panel, was calculated for each laboratory. We compared these metrics with those from 7 563 patients from 296 laboratories outside Africa. Median (interquartile range) patient ED in Africa was similar to that of the rest of the world [9.1 (5.1-15.6) vs 10.3 mSv (6.8-12.6), p = 0.14], although a larger proportion of African patients received a low ED, ≤ 9 mSv targeted in societal recommendations (49.7 vs 38.2%, p < 0.001). Bestpractice adherence was higher among African laboratories (QI score: 6.3 ± 1.2 vs 5.4 ± 1.3, p = 0.013). However, median ED varied significantly among African laboratories (range: 2.0-16.3 mSv; p < 0.0001) and QI range was 4-8. Patient radiation dose from MPI in Africa was similar to that in the rest of the world, and adherence to best practices was relatively high in African laboratories. Nevertheless there remain opportunities to further reduce radiation exposure to African patients from MPI.
Einstein, Andrew J; Pascual, Thomas N B; Mercuri, Mathew; Karthikeyan, Ganesan; Vitola, João V; Mahmarian, John J; Better, Nathan; Bouyoucef, Salah E; Hee-Seung Bom, Henry; Lele, Vikram; Magboo, V Peter C; Alexánderson, Erick; Allam, Adel H; Al-Mallah, Mouaz H; Flotats, Albert; Jerome, Scott; Kaufmann, Philipp A; Luxenburg, Osnat; Shaw, Leslee J; Underwood, S Richard; Rehani, Madan M; Kashyap, Ravi; Paez, Diana; Dondi, Maurizio
2015-07-07
To characterize patient radiation doses from nuclear myocardial perfusion imaging (MPI) and the use of radiation-optimizing 'best practices' worldwide, and to evaluate the relationship between laboratory use of best practices and patient radiation dose. We conducted an observational cross-sectional study of protocols used for all 7911 MPI studies performed in 308 nuclear cardiology laboratories in 65 countries for a single week in March-April 2013. Eight 'best practices' relating to radiation exposure were identified a priori by an expert committee, and a radiation-related quality index (QI) devised indicating the number of best practices used by a laboratory. Patient radiation effective dose (ED) ranged between 0.8 and 35.6 mSv (median 10.0 mSv). Average laboratory ED ranged from 2.2 to 24.4 mSv (median 10.4 mSv); only 91 (30%) laboratories achieved the median ED ≤ 9 mSv recommended by guidelines. Laboratory QIs ranged from 2 to 8 (median 5). Both ED and QI differed significantly between laboratories, countries, and world regions. The lowest median ED (8.0 mSv), in Europe, coincided with high best-practice adherence (mean laboratory QI 6.2). The highest doses (median 12.1 mSv) and low QI (4.9) occurred in Latin America. In hierarchical regression modelling, patients undergoing MPI at laboratories following more 'best practices' had lower EDs. Marked worldwide variation exists in radiation safety practices pertaining to MPI, with targeted EDs currently achieved in a minority of laboratories. The significant relationship between best-practice implementation and lower doses indicates numerous opportunities to reduce radiation exposure from MPI globally. © The Author 2015. Published by Oxford University Press on behalf of the European Society of Cardiology.
Einstein, Andrew J.; Pascual, Thomas N. B.; Mercuri, Mathew; Karthikeyan, Ganesan; Vitola, João V.; Mahmarian, John J.; Better, Nathan; Bouyoucef, Salah E.; Hee-Seung Bom, Henry; Lele, Vikram; Magboo, V. Peter C.; Alexánderson, Erick; Allam, Adel H.; Al-Mallah, Mouaz H.; Flotats, Albert; Jerome, Scott; Kaufmann, Philipp A.; Luxenburg, Osnat; Shaw, Leslee J.; Underwood, S. Richard; Rehani, Madan M.; Kashyap, Ravi; Paez, Diana; Dondi, Maurizio
2015-01-01
Aims To characterize patient radiation doses from nuclear myocardial perfusion imaging (MPI) and the use of radiation-optimizing ‘best practices’ worldwide, and to evaluate the relationship between laboratory use of best practices and patient radiation dose. Methods and results We conducted an observational cross-sectional study of protocols used for all 7911 MPI studies performed in 308 nuclear cardiology laboratories in 65 countries for a single week in March–April 2013. Eight ‘best practices’ relating to radiation exposure were identified a priori by an expert committee, and a radiation-related quality index (QI) devised indicating the number of best practices used by a laboratory. Patient radiation effective dose (ED) ranged between 0.8 and 35.6 mSv (median 10.0 mSv). Average laboratory ED ranged from 2.2 to 24.4 mSv (median 10.4 mSv); only 91 (30%) laboratories achieved the median ED ≤ 9 mSv recommended by guidelines. Laboratory QIs ranged from 2 to 8 (median 5). Both ED and QI differed significantly between laboratories, countries, and world regions. The lowest median ED (8.0 mSv), in Europe, coincided with high best-practice adherence (mean laboratory QI 6.2). The highest doses (median 12.1 mSv) and low QI (4.9) occurred in Latin America. In hierarchical regression modelling, patients undergoing MPI at laboratories following more ‘best practices’ had lower EDs. Conclusion Marked worldwide variation exists in radiation safety practices pertaining to MPI, with targeted EDs currently achieved in a minority of laboratories. The significant relationship between best-practice implementation and lower doses indicates numerous opportunities to reduce radiation exposure from MPI globally. PMID:25898845
Doukky, Rami; Hayes, Kathleen; Frogge, Nathan; Nazir, Noreen T; Collado, Fareed M; Williams, Kim A
2015-05-01
The impact of health insurance carrier and socioeconomic status (SES) on the adherence to appropriate use criteria (AUC) for radionuclide myocardial perfusion imaging (MPI) is unknown. Health insurance carrier's prior authorization and patient's SES impact adherence to AUC for MPI in a fee-for-service setting. We conducted a prospective cohort study of 1511 consecutive patients who underwent outpatient MPI in a multi-site, office-based, fee-for-service setting. The patients were stratified according to the 2009 AUC into appropriate/uncertain appropriateness and inappropriate use groups. Insurance status was categorized as Medicare (does not require prior authorization) vs commercial (requires prior authorization). Socioeconomic status was determined by the median household income in the ZIP code of residence. The proportion of patients with Medicare was 33% vs 67% with commercial insurance. The rate of inappropriate use was higher among patients with commercial insurance vs Medicare (55% vs 24%; P < 0.001); this difference was not significant after adjusting for confounders known to impact AUC determination (odds ratio: 1.06, 95% confidence interval: 0.62-1.82, P = 0.82). The mean annual household income in the residential areas of patients with inappropriate use as compared to those with appropriate/uncertain use was $72 000 ± 21 000 vs $68 000 ± 20 000, respectively (P < 0.001). After adjusting for covariates known to impact AUC determination, SES (top vs bottom quartile income area) was not independently predictive of inappropriate MPI use (odds ratio: 0.9, 95% confidence interval: 0.53-1.52, P = 0.69). Insurance carriers prior authorization and SES do not seem to play a significant role in determining physicians adherence to AUC for MPI. © 2015 Wiley Periodicals, Inc.
Zhang, Li; Liu, Zhe; Hu, Ke-You; Tian, Qing-Bao; Wei, Ling-Ge; Zhao, Zhe; Shen, Hong-Rui; Hu, Jing
2015-01-01
Early detection of muscular dystrophy (MD)-associated cardiomyopathy is important because early medical treatment may slow cardiac remodeling and attenuate symptoms of cardiac dysfunction; however, no sensitive and standard diagnostic method for MD at an earlier stage has been well-recognized. Thus, the aim of this study was to test the early diagnostic value of technetium 99m-methoxyisobutylisonitrile ((99)Tc(m)-MIBI) gated myocardial perfusion imaging (G-MPI) for MD. Ninety-one patients underwent (99)Tc(m)-MIBI G-MPI examinations when they were diagnosed with Duchenne muscular dystrophy (DMD) (n=77) or Becker muscular dystrophy (BMD; n=14). (99)Tc(m)-MIBI G-MPI examinations were repeated in 43 DMD patients who received steroid treatments for 2 years as a follow-up examination. Myocardial defects were observed in nearly every segment of the left ventricular wall in both DMD and BMD patients compared with controls, especially in the inferior walls and the apices by using (99)Tc(m)-MIBI G-MPI. Cardiac wall movement impairment significantly correlated with age in the DMD and BMD groups (r s=0.534 [P<0.05] and r s=0.784 [P<0.05], respectively). Intermittent intravenous doses of glucocorticoids and continuation with oral steroid treatments significantly improved myocardial function in DMD patients (P<0.05), but not in BMD patients. (99)Tc(m)-MIBI G-MPI is a sensitive and safe approach for early evaluation of cardiomyopathy in patients with DMD or BMD, and can serve as a candidate method for the evaluation of progression, prognosis, and assessment of the effect of glucocorticoid treatment in these patients.
Bouyoucef, Salah E; Mercuri, Mathew; Einstein, Andrew J; Pascual, Thomas NB; Kashyap, Ravi; Dondi, Maurizio; Paez, Diana; Allam, Adel H; Vangu, Mboyo; Vitola, João V; Better, Nathan; Karthikeyan, Ganesan; Mahmarian, John J; Rehani, Madan M; Einstein, Andrew J
2017-01-01
Summary Objective: While nuclear myocardial perfusion imaging (MPI) offers many benefits to patients with known or suspected cardiovascular disease, concerns exist regarding radiationassociated health effects. Little is known regarding MPI practice in Africa. We sought to characterise radiation doses and the use of MPI best practices that could minimise radiation in African nuclear cardiology laboratories, and compare these to practice worldwide. Methods: Demographics and clinical characteristics were collected for a consecutive sample of 348 patients from 12 laboratories in six African countries over a one-week period from March to April 2013. Radiation effective dose (ED) was estimated for each patient. A quality index (QI) enumerating adherence to eight best practices, identified a priori by an IAEA expert panel, was calculated for each laboratory. We compared these metrics with those from 7 563 patients from 296 laboratories outside Africa. Results: to that of the rest of the world [9.1 (5.1–15.6) vs 10.3 mSv (6.8–12.6), p = 0.14], although a larger proportion of African patients received a low ED, ≤ 9 mSv targeted in societal recommendations (49.7 vs 38.2%, p < 0.001). Bestpractice adherence was higher among African laboratories (QI score: 6.3 ± 1.2 vs 5.4 ± 1.3, p = 0.013). However, median ED varied significantly among African laboratories (range: 2.0–16.3 mSv; p < 0.0001) and QI range was 4–8. Conclusion: Patient radiation dose from MPI in Africa was similar to that in the rest of the world, and adherence to best practices was relatively high in African laboratories. Nevertheless there remain opportunities to further reduce radiation exposure to African patients from MPI. PMID:28906538
Rastgou, Fereydoon; Shojaeifard, Maryam; Amin, Ahmad; Ghaedian, Tahereh; Firoozabadi, Hasan; Malek, Hadi; Yaghoobi, Nahid; Bitarafan-Rajabi, Ahmad; Haghjoo, Majid; Amouzadeh, Hedieh; Barati, Hossein
2014-12-01
Recently, the phase analysis of gated single-photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI) has become feasible via several software packages for the evaluation of left ventricular mechanical dyssynchrony. We compared two quantitative software packages, quantitative gated SPECT (QGS) and Emory cardiac toolbox (ECTb), with tissue Doppler imaging (TDI) as the conventional method for the evaluation of left ventricular mechanical dyssynchrony. Thirty-one patients with severe heart failure (ejection fraction ≤35%) and regular heart rhythm, who referred for gated-SPECT MPI, were enrolled. TDI was performed within 3 days after MPI. Dyssynchrony parameters derived from gated-SPECT MPI were analyzed by QGS and ECTb and were compared with the Yu index and septal-lateral wall delay measured by TDI. QGS and ECTb showed a good correlation for assessment of phase histogram bandwidth (PHB) and phase standard deviation (PSD) (r = 0.664 and r = 0.731, P < .001, respectively). However, the mean value of PHB and PSD by ECTb was significantly higher than that of QGS. No significant correlation was found between ECTb and QGS and the Yu index. Nevertheless, PHB, PSD, and entropy derived from QGS revealed a significant (r = 0.424, r = 0.478, r = 0.543, respectively; P < .02) correlation with septal-lateral wall delay. Despite a good correlation between QGS and ECTb software packages, different normal cut-off values of PSD and PHB should be defined for each software package. There was only a modest correlation between phase analysis of gated-SPECT MPI and TDI data, especially in the population of heart failure patients with both narrow and wide QRS complex.
Verra, Martin L; Angst, Felix; Brioschi, Roberto; Lehmann, Susanne; Keefe, Francis J; Staal, J Bart; de Bie, Rob A; Aeschlimann, André
2009-01-01
INTRODUCTION: The present study aimed to replicate and validate the empirically derived subgroup classification based on the Multidimensional Pain Inventory (MPI) in a sample of highly disabled fibromyalgia (FM) patients. Second, it examined how the identified subgroups differed in their response to an intensive, interdisciplinary inpatient pain management program. METHODS: Participants were 118 persons with FM who experienced persistent pain and were disabled. Subgroup classification was conducted by cluster analysis using MPI subscale scores at entry to the program. At program entry and discharge, participants completed the MPI, Medical Outcomes Study Short Form-36, Hospital Anxiety and Depression Scale and Coping Strategies Questionnaire. RESULTS: Cluster analysis identified three subgroups in the highly disabled sample that were similar to those described by other studies using less disabled samples of FM. The dysfunctional subgroup (DYS; 36% of the sample) showed the highest level of depression, the interpersonally distressed subgroup (ID; 24%) showed a modest level of depression and the adaptive copers subgroup (AC; 38%) showed the lowest depression scores in the MPI (negative mood), Medical Outcomes Study Short Form-36 (mental health), Hospital Anxiety and Depression Scale (depression) and Coping Strategies Questionnaire (catastrophizing). Significant differences in treatment outcome were observed among the three subgroups in terms of reduction of pain severity (as assessed using the MPI). The effect sizes were 1.42 for DYS, 1.32 for AC and 0.62 for ID (P=0.004 for pairwise comparison of ID-AC and P=0.018 for ID-DYS). DISCUSSION: These findings underscore the importance of assessing individuals’ differences in how they adjust to FM. PMID:20011715
Amer, Hamid; Niaz, Khalid; Hatazawa, Jun; Gasmelseed, Ahmed; Samiri, Hussain Al; Al Othman, Maram; Hammad, Mai Al
2017-11-01
We sought to determine the prognostic importance of adenosine-induced ischemic ECG changes in patients with normal single-photon emission computed tomography myocardial perfusion images (MPI). We carried out a retrospective analysis of 765 patients undergoing adenosine MPI between January 2013 and January 2015. Patients with baseline ECG abnormalities and/or abnormal scan were excluded. Overall, 67 (8.7%) patients had ischemic ECG changes during adenosine infusion in the form of ST depression of 1 mm or more. Of these, 29 [43% (3.8% of all patients)] had normal MPI (positive ECG group). An age-matched and sex-matched group of 108 patients with normal MPI without ECG changes served as control participants (negative ECG group). During a mean follow-up duration of 33.3±6.1 months, patients in the positive ECG group did not have significantly more adverse cardiac events than those in the negative ECG group. One (0.9%) patient in the negative ECG group had a nonfatal myocardial infarction (0.7% annual event rate after a negative MPI). Also in this group, two (1.8%) patients admitted with a diagnosis of CAD where they have been ruled out by angiography. A fourth case in this, in the negative ECG group, was admitted because of heart failure that proved to be secondary to a pulmonary cause and not CAD. A case only in the positive ECG group was admitted as a CAD that was ruled out by coronary angiography. Patients with normal myocardial perfusion scintigraphy in whom ST-segment depression develops during adenosine stress test appear to have no increased risk for future cardiac events compared with similar patients without ECG evidence of ischemia.
Influence of North Atlantic modes on European climate extremes
NASA Astrophysics Data System (ADS)
Proemmel, K.; Cubasch, U.
2017-12-01
It is well known that the North Atlantic strongly influences European climate. Only few studies exist that focus on its impact on climate extremes. We are interested in these extremes and the processes and mechanisms behind it. For the analysis of the North Atlantic Oscillation (NAO) we use simulations performed with the Max Planck Institute for Meteorology Earth System Model (MPI-ESM). The NAO has a strong impact especially on European winter and the changes in minimum temperature are even larger than in maximum temperature. The impact of the Atlantic Multi-decadal Variability (AMV) on climate extremes is analyzed in ECHAM6 simulations forced with AMV warm and AMV cold sea surface temperature patterns. We analyze different extreme indices and try to understand the processes.
GSD-1G and MPI-DING Reference Glasses for In Situ and Bulk Isotopic Determination
Jochum, K.P.; Wilson, S.A.; Abouchami, W.; Amini, M.; Chmeleff, J.; Eisenhauer, A.; Hegner, E.; Iaccheri, L.M.; Kieffer, B.; Krause, J.; McDonough, W.F.; Mertz-Kraus, R.; Raczek, I.; Rudnick, R.L.; Scholz, Donna K.; Steinhoefel, G.; Stoll, B.; Stracke, A.; Tonarini, S.; Weis, D.; Weis, U.; Woodhead, J.D.
2011-01-01
This paper contains the results of an extensive isotopic study of United States Geological Survey GSD-1G and MPI-DING reference glasses. Thirteen different laboratories were involved using high-precision bulk (TIMS, MC-ICP-MS) and microanalytical (LA-MC-ICP-MS, LA-ICP-MS) techniques. Detailed studies were performed to demonstrate the large-scale and small-scale homogeneity of the reference glasses. Together with previously published isotopic data from ten other laboratories, preliminary reference and information values as well as their uncertainties at the 95% confidence level were determined for H, O, Li, B, Si, Ca, Sr, Nd, Hf, Pb, Th and U isotopes using the recommendations of the International Association of Geoanalysts for certification of reference materials. Our results indicate that GSD-1G and the MPI-DING glasses are suitable reference materials for microanalytical and bulk analytical purposes. Ce document contient les r??sultats d'une importante ??tude isotopique des verres de r??f??rence USGS GSD-1G et MPI-DING. Treize laboratoires diff??rents ont particip?? au travers de techniques analytiques de haute pr??cision travaillant soit sur ??chantillon total (TIMS, MC-ICP-MS) soit par microanalyse ??in situ?? (LA-MC-ICP-MS, LA-ICP-MS). ?? 2010 The Authors. Geostandards and Geoanalytical Research ?? 2010 International Association of Geoanalysts.
NASA Astrophysics Data System (ADS)
Tomitaka, Asahi; Arami, Hamed; Gandhi, Sonu; Krishnan, Kannan M.
2015-10-01
Magnetic Particle Imaging (MPI) is a new real-time imaging modality, which promises high tracer mass sensitivity and spatial resolution directly generated from iron oxide nanoparticles. In this study, monodisperse iron oxide nanoparticles with median core diameters ranging from 14 to 26 nm were synthesized and their surface was conjugated with lactoferrin to convert them into brain glioma targeting agents. The conjugation was confirmed with the increase of the hydrodynamic diameters, change of zeta potential, and Bradford assay. Magnetic particle spectrometry (MPS), performed to evaluate the MPI performance of these nanoparticles, showed no change in signal after lactoferrin conjugation to nanoparticles for all core diameters, suggesting that the MPI signal is dominated by Néel relaxation and thus independent of hydrodynamic size difference or presence of coating molecules before and after conjugations. For this range of core sizes (14-26 nm), both MPS signal intensity and spatial resolution improved with increasing core diameter of nanoparticles. The lactoferrin conjugated iron oxide nanoparticles (Lf-IONPs) showed specific cellular internalization into C6 cells with a 5-fold increase in MPS signal compared to IONPs without lactoferrin, both after 24 h incubation. These results suggest that Lf-IONPs can be used as tracers for targeted brain glioma imaging using MPI.
NASA Astrophysics Data System (ADS)
Rahmer, J.; Antonelli, A.; Sfara, C.; Tiemann, B.; Gleich, B.; Magnani, M.; Weizenecker, J.; Borgert, J.
2013-06-01
Magnetic particle imaging (MPI) is a new medical imaging approach that is based on the nonlinear magnetization response of super-paramagnetic iron oxide nanoparticles (SPIOs) injected into the blood stream. To date, real-time MPI of the bolus passage of an approved MRI SPIO contrast agent injected into the tail vein of living mice has been demonstrated. However, nanoparticles are rapidly removed from the blood stream by the mononuclear phagocyte system. Therefore, imaging applications for long-term monitoring require the repeated administration of bolus injections, which complicates quantitative comparisons due to the temporal variations in concentration. Encapsulation of SPIOs into red blood cells (RBCs) has been suggested to increase the blood circulation time of nanoparticles. This work presents first evidence that SPIO-loaded RBCs can be imaged in the blood pool of mice several hours after injection using MPI. This finding is supported by magnetic particle spectroscopy performed to quantify the iron concentration in blood samples extracted from the mice 3 and 24 h after injection of SPIO-loaded RBCs. Based on these results, new MPI applications can be envisioned, such as permanent 3D real-time visualization of the vessel tree during interventional procedures, bleeding monitoring after stroke, or long-term monitoring and treatment control of cardiovascular diseases.
Implementing Multidisciplinary and Multi-Zonal Applications Using MPI
NASA Technical Reports Server (NTRS)
Fineberg, Samuel A.
1995-01-01
Multidisciplinary and multi-zonal applications are an important class of applications in the area of Computational Aerosciences. In these codes, two or more distinct parallel programs or copies of a single program are utilized to model a single problem. To support such applications, it is common to use a programming model where a program is divided into several single program multiple data stream (SPMD) applications, each of which solves the equations for a single physical discipline or grid zone. These SPMD applications are then bound together to form a single multidisciplinary or multi-zonal program in which the constituent parts communicate via point-to-point message passing routines. Unfortunately, simple message passing models, like Intel's NX library, only allow point-to-point and global communication within a single system-defined partition. This makes implementation of these applications quite difficult, if not impossible. In this report it is shown that the new Message Passing Interface (MPI) standard is a viable portable library for implementing the message passing portion of multidisciplinary applications. Further, with the extension of a portable loader, fully portable multidisciplinary application programs can be developed. Finally, the performance of MPI is compared to that of some native message passing libraries. This comparison shows that MPI can be implemented to deliver performance commensurate with native message libraries.
Levy, Andrew E; Shah, Nishant R; Matheny, Michael E; Reeves, Ruth M; Gobbel, Glenn T; Bradley, Steven M
2018-04-25
Reporting standards promote clarity and consistency of stress myocardial perfusion imaging (MPI) reports, but do not require an assessment of post-test risk. Natural Language Processing (NLP) tools could potentially help estimate this risk, yet it is unknown whether reports contain adequate descriptive data to use NLP. Among VA patients who underwent stress MPI and coronary angiography between January 1, 2009 and December 31, 2011, 99 stress test reports were randomly selected for analysis. Two reviewers independently categorized each report for the presence of critical data elements essential to describing post-test ischemic risk. Few stress MPI reports provided a formal assessment of post-test risk within the impression section (3%) or the entire document (4%). In most cases, risk was determinable by combining critical data elements (74% impression, 98% whole). If ischemic risk was not determinable (25% impression, 2% whole), inadequate description of systolic function (9% impression, 1% whole) and inadequate description of ischemia (5% impression, 1% whole) were most commonly implicated. Post-test ischemic risk was determinable but rarely reported in this sample of stress MPI reports. This supports the potential use of NLP to help clarify risk. Further study of NLP in this context is needed.
Recent developments in bend-insensitive and ultra-bend-insensitive fibers
NASA Astrophysics Data System (ADS)
Boivin, David; de Montmorillon, Louis-Anne; Provost, Lionel; Montaigne, Nelly; Gooijer, Frans; Aldea, Eugen; Jensma, Jaap; Sillard, Pierre
2010-02-01
Designed to overcome the limitations in case of extreme bending conditions, Bend- and Ultra-Bend-Insensitive Fibers (BIFs and UBIFs) appear as ideal solutions for use in FTTH networks and in components, pigtails or patch-cords for ever demanding applications such as military or sensing. Recently, however, questions have been raised concerning the Multi-Path-Interference (MPI) levels in these fibers. Indeed, they are potentially subject to interferences between the fundamental mode and the higher-order mode that is also bend resistant. This MPI is generated because of discrete discontinuities such as staples, bends and splices/connections that occur on distance scales that become comparable to the laser coherent length. In this paper, we will demonstrate the high MPI tolerance of all-solid single-trench-assisted BIFs and UBIFs. We will present the first comprehensive study combining theoretical and experimental points of view to quantify the impact of fusion splices on coherent MPI. To be complete, results for mechanical splices will also be reported. Finally, we will show how the single-trench- assisted concept combined with the versatile PCVD process allows to tightly control the distributions of fibers characteristics. Such controls are needed to massively produce BIFs and to meet the more stringent specifications of the UBIFs.
Toward Abstracting the Communication Intent in Applications to Improve Portability and Productivity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mintz, Tiffany M; Hernandez, Oscar R; Kartsaklis, Christos
Programming with communication libraries such as the Message Passing Interface (MPI) obscures the high-level intent of the communication in an application and makes static communication analysis difficult to do. Compilers are unaware of communication libraries specifics, leading to the exclusion of communication patterns from any automated analysis and optimizations. To overcome this, communication patterns can be expressed at higher-levels of abstraction and incrementally added to existing MPI applications. In this paper, we propose the use of directives to clearly express the communication intent of an application in a way that is not specific to a given communication library. Our communicationmore » directives allow programmers to express communication among processes in a portable way, giving hints to the compiler on regions of computations that can be overlapped with communication and relaxing communication constraints on the ordering, completion and synchronization of the communication imposed by specific libraries such as MPI. The directives can then be translated by the compiler into message passing calls that efficiently implement the intended pattern and be targeted to multiple communication libraries. Thus far, we have used the directives to express point-to-point communication patterns in C, C++ and Fortran applications, and have translated them to MPI and SHMEM.« less
NASA Astrophysics Data System (ADS)
Nuraini, Lutviasari; Prifiharni, Siska; Priyotomo, Gadang; Sundjono, Gunawan, Hadi; Purawiardi, Ibrahim
2018-05-01
The performance of carbon steel, galvanized steel and aluminium after one month exposed in the atmospheric coastal area, which is in Limbangan and Karangsong Beach, West Java, Indonesia was evaluated. The corrosion rate was determined by weight loss method and the morphology of the steel after exposed was observed by Scanning Electron Microscopy(SEM)/Energy Dispersive X-Ray Analysis(EDX). The site was monitored to determine the chloride content in the marine atmosphere. Then, the corrosion products formed at carbon steel were characterized by X-Ray diffraction (XRD). The result showed the aggressively corrosion in Karangsong beach, indicated from the corrosion rate of carbon steel, galvanized steel and aluminium were 38.514 mpy; 4.7860 mpy and 0.5181 mpy, respectively. While in Limbangan Beach the corrosion rate of specimen carbon steel, galvanized steel and aluminium were 3.339; 0.219 and 0.166 mpy, respectively. The chloride content was found to be the main factor that influences in the atmospheric corrosion process in this area. Chloride content accumulated in Karangsong and Limbangan was 497 mg/m2.day and 117 mg/m2.day, respectively. The XRD Analysis on each carbon steel led to the characterization of a complex mixture of iron oxides phases.
High frequency QRS ECG predicts ischemic defects during myocardial perfusion imaging
NASA Technical Reports Server (NTRS)
2004-01-01
Changes in high frequency QRS components of the electrocardiogram (HF QRS ECG) (150-250 Hz) are more sensitive than changes in conventional ST segments for detecting myocardial ischemia. We investigated the accuracy of 12-lead HF QRS ECG in detecting ischemia during adenosine tetrofosmin myocardial perfusion imaging (MPI). 12-lead HF QRS ECG recordings were obtained from 45 patients before and during adenosine technetium-99 tetrofosmin MPI tests. Before the adenosine infusions, recordings of HF QRS were analyzed according to a morphological score that incorporated the number, type and location of reduced amplitude zones (RAZs) present in the 12 leads. During the adenosine infusions, recordings of HF QRS were analyzed according to the maximum percentage changes (in both the positive and negative directions) that occurred in root mean square (RMS) voltage amplitudes within the 12 leads. The best set of prospective HF QRS criteria had a sensitivity of 94% and a specificity of 83% for correctly identifying the MPI result. The sensitivity of simultaneous ST segment changes (18%) was significantly lower than that of any individual HF QRS criterion (P less than 0.00l). Analysis of 12-lead HF QRS ECG is highly sensitive and specific for detecting ischemic perfusion defects during adenosine MPI stress tests and significantly more sensitive than analysis of conventional ST segments.
High frequency QRS ECG predicts ischemic defects during myocardial perfusion imaging
NASA Technical Reports Server (NTRS)
Rahman, Atiar
2006-01-01
Background: Changes in high frequency QRS components of the electrocardiogram (HF QRS ECG) (150-250 Hz) are more sensitive than changes in conventional ST segments for detecting myocardial ischemia. We investigated the accuracy of 12-lead HF QRS ECG in detecting ischemia during adenosine tetrofosmin myocardial perfusion imaging (MPI). Methods and Results: 12-lead HF QRS ECG recordings were obtained from 45 patients before and during adenosine technetium-99 tetrofosmin MPI tests. Before the adenosine infusions, recordings of HF QRS were analyzed according to a morphological score that incorporated the number, type and location of reduced amplitude zones (RAZs) present in the 12 leads. During the adenosine infusions, recordings of HF QRS were analyzed according to the maximum percentage changes (in both the positive and negative directions) that occurred in root mean square (RMS) voltage amplitudes within the 12 leads. The best set of prospective HF QRS criteria had a sensitivity of 94% and a specificity of 83% for correctly identifying the MPI result. The sensitivity of simultaneous ST segment changes (18%) was significantly lower than that of any individual HF QRS criterion (P<0.001). Conclusions: Analysis of 12-lead HF QRS ECG is highly sensitive and specific for detecting ischemic perfusion defects during adenosine MPI stress tests and significantly more sensitive than analysis of conventional ST segments.
SPOTting Model Parameters Using a Ready-Made Python Package
NASA Astrophysics Data System (ADS)
Houska, Tobias; Kraft, Philipp; Chamorro-Chavez, Alejandro; Breuer, Lutz
2017-04-01
The choice for specific parameter estimation methods is often more dependent on its availability than its performance. We developed SPOTPY (Statistical Parameter Optimization Tool), an open source python package containing a comprehensive set of methods typically used to calibrate, analyze and optimize parameters for a wide range of ecological models. SPOTPY currently contains eight widely used algorithms, 11 objective functions, and can sample from eight parameter distributions. SPOTPY has a model-independent structure and can be run in parallel from the workstation to large computation clusters using the Message Passing Interface (MPI). We tested SPOTPY in five different case studies to parameterize the Rosenbrock, Griewank and Ackley functions, a one-dimensional physically based soil moisture routine, where we searched for parameters of the van Genuchten-Mualem function and a calibration of a biogeochemistry model with different objective functions. The case studies reveal that the implemented SPOTPY methods can be used for any model with just a minimal amount of code for maximal power of parameter optimization. They further show the benefit of having one package at hand that includes number of well performing parameter search methods, since not every case study can be solved sufficiently with every algorithm or every objective function.
Using Formal Grammars to Predict I/O Behaviors in HPC: The Omnisc'IO Approach
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dorier, Matthieu; Ibrahim, Shadi; Antoniu, Gabriel
2016-08-01
The increasing gap between the computation performance of post-petascale machines and the performance of their I/O subsystem has motivated many I/O optimizations including prefetching, caching, and scheduling. In order to further improve these techniques, modeling and predicting spatial and temporal I/O patterns of HPC applications as they run has become crucial. In this paper we present Omnisc'IO, an approach that builds a grammar-based model of the I/O behavior of HPC applications and uses it to predict when future I/O operations will occur, and where and how much data will be accessed. To infer grammars, Omnisc'IO is based on StarSequitur, amore » novel algorithm extending Nevill-Manning's Sequitur algorithm. Omnisc'IO is transparently integrated into the POSIX and MPI I/O stacks and does not require any modification in applications or higher-level I/O libraries. It works without any prior knowledge of the application and converges to accurate predictions of any N future I/O operations within a couple of iterations. Its implementation is efficient in both computation time and memory footprint.« less
SPOTting Model Parameters Using a Ready-Made Python Package.
Houska, Tobias; Kraft, Philipp; Chamorro-Chavez, Alejandro; Breuer, Lutz
2015-01-01
The choice for specific parameter estimation methods is often more dependent on its availability than its performance. We developed SPOTPY (Statistical Parameter Optimization Tool), an open source python package containing a comprehensive set of methods typically used to calibrate, analyze and optimize parameters for a wide range of ecological models. SPOTPY currently contains eight widely used algorithms, 11 objective functions, and can sample from eight parameter distributions. SPOTPY has a model-independent structure and can be run in parallel from the workstation to large computation clusters using the Message Passing Interface (MPI). We tested SPOTPY in five different case studies to parameterize the Rosenbrock, Griewank and Ackley functions, a one-dimensional physically based soil moisture routine, where we searched for parameters of the van Genuchten-Mualem function and a calibration of a biogeochemistry model with different objective functions. The case studies reveal that the implemented SPOTPY methods can be used for any model with just a minimal amount of code for maximal power of parameter optimization. They further show the benefit of having one package at hand that includes number of well performing parameter search methods, since not every case study can be solved sufficiently with every algorithm or every objective function.
SPOTting Model Parameters Using a Ready-Made Python Package
Houska, Tobias; Kraft, Philipp; Chamorro-Chavez, Alejandro; Breuer, Lutz
2015-01-01
The choice for specific parameter estimation methods is often more dependent on its availability than its performance. We developed SPOTPY (Statistical Parameter Optimization Tool), an open source python package containing a comprehensive set of methods typically used to calibrate, analyze and optimize parameters for a wide range of ecological models. SPOTPY currently contains eight widely used algorithms, 11 objective functions, and can sample from eight parameter distributions. SPOTPY has a model-independent structure and can be run in parallel from the workstation to large computation clusters using the Message Passing Interface (MPI). We tested SPOTPY in five different case studies to parameterize the Rosenbrock, Griewank and Ackley functions, a one-dimensional physically based soil moisture routine, where we searched for parameters of the van Genuchten-Mualem function and a calibration of a biogeochemistry model with different objective functions. The case studies reveal that the implemented SPOTPY methods can be used for any model with just a minimal amount of code for maximal power of parameter optimization. They further show the benefit of having one package at hand that includes number of well performing parameter search methods, since not every case study can be solved sufficiently with every algorithm or every objective function. PMID:26680783
NASA Astrophysics Data System (ADS)
Lee, Y. C.; Thompson, H. M.; Gaskell, P. H.
2009-12-01
FILMPAR is a highly efficient and portable parallel multigrid algorithm for solving a discretised form of the lubrication approximation to three-dimensional, gravity-driven, continuous thin film free-surface flow over substrates containing micro-scale topography. While generally applicable to problems involving heterogeneous and distributed features, for illustrative purposes the algorithm is benchmarked on a distributed memory IBM BlueGene/P computing platform for the case of flow over a single trench topography, enabling direct comparison with complementary experimental data and existing serial multigrid solutions. Parallel performance is assessed as a function of the number of processors employed and shown to lead to super-linear behaviour for the production of mesh-independent solutions. In addition, the approach is used to solve for the case of flow over a complex inter-connected topographical feature and a description provided of how FILMPAR could be adapted relatively simply to solve for a wider class of related thin film flow problems. Program summaryProgram title: FILMPAR Catalogue identifier: AEEL_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEEL_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 530 421 No. of bytes in distributed program, including test data, etc.: 1 960 313 Distribution format: tar.gz Programming language: C++ and MPI Computer: Desktop, server Operating system: Unix/Linux Mac OS X Has the code been vectorised or parallelised?: Yes. Tested with up to 128 processors RAM: 512 MBytes Classification: 12 External routines: GNU C/C++, MPI Nature of problem: Thin film flows over functional substrates containing well-defined single and complex topographical features are of enormous significance, having a wide variety of engineering, industrial and physical applications. However, despite recent modelling advances, the accurate numerical solution of the equations governing such problems is still at a relatively early stage. Indeed, recent studies employing a simplifying long-wave approximation have shown that highly efficient numerical methods are necessary to solve the resulting lubrication equations in order to achieve the level of grid resolution required to accurately capture the effects of micro- and nano-scale topographical features. Solution method: A portable parallel multigrid algorithm has been developed for the above purpose, for the particular case of flow over submerged topographical features. Within the multigrid framework adopted, a W-cycle is used to accelerate convergence in respect of the time dependent nature of the problem, with relaxation sweeps performed using a fixed number of pre- and post-Red-Black Gauss-Seidel Newton iterations. In addition, the algorithm incorporates automatic adaptive time-stepping to avoid the computational expense associated with repeated time-step failure. Running time: 1.31 minutes using 128 processors on BlueGene/P with a problem size of over 16.7 million mesh points.
A portable approach for PIC on emerging architectures
NASA Astrophysics Data System (ADS)
Decyk, Viktor
2016-03-01
A portable approach for designing Particle-in-Cell (PIC) algorithms on emerging exascale computers, is based on the recognition that 3 distinct programming paradigms are needed. They are: low level vector (SIMD) processing, middle level shared memory parallel programing, and high level distributed memory programming. In addition, there is a memory hierarchy associated with each level. Such algorithms can be initially developed using vectorizing compilers, OpenMP, and MPI. This is the approach recommended by Intel for the Phi processor. These algorithms can then be translated and possibly specialized to other programming models and languages, as needed. For example, the vector processing and shared memory programming might be done with CUDA instead of vectorizing compilers and OpenMP, but generally the algorithm itself is not greatly changed. The UCLA PICKSC web site at http://www.idre.ucla.edu/ contains example open source skeleton codes (mini-apps) illustrating each of these three programming models, individually and in combination. Fortran2003 now supports abstract data types, and design patterns can be used to support a variety of implementations within the same code base. Fortran2003 also supports interoperability with C so that implementations in C languages are also easy to use. Finally, main codes can be translated into dynamic environments such as Python, while still taking advantage of high performing compiled languages. Parallel languages are still evolving with interesting developments in co-Array Fortran, UPC, and OpenACC, among others, and these can also be supported within the same software architecture. Work supported by NSF and DOE Grants.
Tree-based solvers for adaptive mesh refinement code FLASH - I: gravity and optical depths
NASA Astrophysics Data System (ADS)
Wünsch, R.; Walch, S.; Dinnbier, F.; Whitworth, A.
2018-04-01
We describe an OctTree algorithm for the MPI parallel, adaptive mesh refinement code FLASH, which can be used to calculate the gas self-gravity, and also the angle-averaged local optical depth, for treating ambient diffuse radiation. The algorithm communicates to the different processors only those parts of the tree that are needed to perform the tree-walk locally. The advantage of this approach is a relatively low memory requirement, important in particular for the optical depth calculation, which needs to process information from many different directions. This feature also enables a general tree-based radiation transport algorithm that will be described in a subsequent paper, and delivers excellent scaling up to at least 1500 cores. Boundary conditions for gravity can be either isolated or periodic, and they can be specified in each direction independently, using a newly developed generalization of the Ewald method. The gravity calculation can be accelerated with the adaptive block update technique by partially re-using the solution from the previous time-step. Comparison with the FLASH internal multigrid gravity solver shows that tree-based methods provide a competitive alternative, particularly for problems with isolated or mixed boundary conditions. We evaluate several multipole acceptance criteria (MACs) and identify a relatively simple approximate partial error MAC which provides high accuracy at low computational cost. The optical depth estimates are found to agree very well with those of the RADMC-3D radiation transport code, with the tree-solver being much faster. Our algorithm is available in the standard release of the FLASH code in version 4.0 and later.
Enhancing nurses' roles to improve quality and efficiency of non-medical cardiac stress tests.
Bernhardt, Lizelle; Ross, Lisa; Greaves, Claire
Myocardial perfusion imaging (MPI) is a test that aids the diagnosis of coronary heart disease, of which pharmacological stress is a key component. An increase in demand had resulted in a 42 week waiting time for MPI in Leicester. This article looks at how implementing non-medically led stress tests reduced this waiting list. It discusses the obstacles involved and the measures needed to make the service a success.
Hybrid MPI+OpenMP Programming of an Overset CFD Solver and Performance Investigations
NASA Technical Reports Server (NTRS)
Djomehri, M. Jahed; Jin, Haoqiang H.; Biegel, Bryan (Technical Monitor)
2002-01-01
This report describes a two level parallelization of a Computational Fluid Dynamic (CFD) solver with multi-zone overset structured grids. The approach is based on a hybrid MPI+OpenMP programming model suitable for shared memory and clusters of shared memory machines. The performance investigations of the hybrid application on an SGI Origin2000 (O2K) machine is reported using medium and large scale test problems.
Indirect spectrophotometric determination of trace cyanide with cationic porphyrins.
Ishii, H; Kohata, K
1991-05-01
Three highly sensitive methods for the determination of cyanide have been developed, based on the fact that the complexation of silver ions with three cationic porphyrins, 5,10,15,20-tetrakis-(1-methyl-2-pyridinio)porphine [T(2-MPy)P], 5,10,15,20-tetrakis(1-methyl-3-pyridinio)porphine [T(3-MPy)P] and 5,10,15,20-tetrakis(1-methyl-4-pyridinio)porphine [T(4-MPy)P], in alkaline media is inhibited by cyanide and the decrease in absorbance of the silver(II) complex is proportional to the cyanide concentration. Sensitivities of the procedures developed are 0.133, 0.126 and 0.234 ng/cm(2), respectively for an absorbance of 0.001. Cadmium(II), copper(II), mercury(II), zinc(II), iodide and sulfide interfere with the cyanide determination. One of the proposed methods was applied to the determination of cyanide in waste-water samples, with satisfactory results.
Potential Application of a Graphical Processing Unit to Parallel Computations in the NUBEAM Code
NASA Astrophysics Data System (ADS)
Payne, J.; McCune, D.; Prater, R.
2010-11-01
NUBEAM is a comprehensive computational Monte Carlo based model for neutral beam injection (NBI) in tokamaks. NUBEAM computes NBI-relevant profiles in tokamak plasmas by tracking the deposition and the slowing of fast ions. At the core of NUBEAM are vector calculations used to track fast ions. These calculations have recently been parallelized to run on MPI clusters. However, cost and interlink bandwidth limit the ability to fully parallelize NUBEAM on an MPI cluster. Recent implementation of double precision capabilities for Graphical Processing Units (GPUs) presents a cost effective and high performance alternative or complement to MPI computation. Commercially available graphics cards can achieve up to 672 GFLOPS double precision and can handle hundreds of thousands of threads. The ability to execute at least one thread per particle simultaneously could significantly reduce the execution time and the statistical noise of NUBEAM. Progress on implementation on a GPU will be presented.
Kieffer, Philip J; Williams, Jarred M; Shepard, Molly K; Giguère, Steeve; Epstein, Kira L
2018-01-01
The objectives of the study were to: i) determine baseline microvascular perfusion indices (MPI) and assess their repeatability in healthy horses under general anesthesia, and ii) compare the MPIs of 3 microvascular beds (oral mucosa, colonic serosa, and rectal mucosa). Healthy adult horses were anesthetized and sidestream dark field microscopy was used to collect video loops of the oral mucosa, rectal mucosa, and colonic serosa under normotensive conditions without cardiovascular support drugs; videos were later analyzed to produce MPIs. Baseline MPI values were determined for each site, which included the total vessel density (TVD), perfused vessel density (PVD), portion perfused vessels (PPV), and microcirculatory flow index (MFI). Differences in MPIs between microvascular beds were not statistically significant. Repeatability of the measurements varied for each MPI. In particular, the site of sampling had a profound effect on the repeatability of the PPV measurements and should be considered in future studies.
Kieffer, Philip J.; Williams, Jarred M.; Shepard, Molly K.; Giguère, Steeve; Epstein, Kira L.
2018-01-01
The objectives of the study were to: i) determine baseline microvascular perfusion indices (MPI) and assess their repeatability in healthy horses under general anesthesia, and ii) compare the MPIs of 3 microvascular beds (oral mucosa, colonic serosa, and rectal mucosa). Healthy adult horses were anesthetized and sidestream dark field microscopy was used to collect video loops of the oral mucosa, rectal mucosa, and colonic serosa under normotensive conditions without cardiovascular support drugs; videos were later analyzed to produce MPIs. Baseline MPI values were determined for each site, which included the total vessel density (TVD), perfused vessel density (PVD), portion perfused vessels (PPV), and microcirculatory flow index (MFI). Differences in MPIs between microvascular beds were not statistically significant. Repeatability of the measurements varied for each MPI. In particular, the site of sampling had a profound effect on the repeatability of the PPV measurements and should be considered in future studies. PMID:29382969
Review of progress in magnetic particle inspection
NASA Astrophysics Data System (ADS)
Eisenmann, David J.; Enyart, Darrel; Lo, Chester; Brasche, Lisa
2014-02-01
Magnetic particle inspection (MPI) has been widely utilized for decades, and sees considerable use in the aerospace industry with a majority of the steel parts being inspected with MPI at some point in the lifecycle. Typical aircraft locations inspected are landing gear, engine components, attachment hardware, and doors. In spite of its numerous applications the method remains poorly understood, and there are many aspects of that method which would benefit from in-depth study. This shortcoming is due to the fact that MPI combines the complicated nature of electromagnetics, metallurgical material effects, fluid-particle motion dynamics, and physiological human factors into a single inspection. To promote understanding of the intricate method issues that affect sensitivity, or to assist with the revision of industry specifications and standards, research studies will be prioritized through the guidance of a panel of industry experts, using an approach which has worked successfully in the past to guide fluorescent penetrant inspection (FPI) research efforts.
Overview of SCIAMACHY validation: 2002 2004
NASA Astrophysics Data System (ADS)
Piters, A. J. M.; Bramstedt, K.; Lambert, J.-C.; Kirchhoff, B.
2005-08-01
SCIAMACHY, on board Envisat, is now in operation for almost three years. This UV/visible/NIR spectrometer measures the solar irradiance, the earthshine radiance scattered at nadir and from the limb, and the attenuation of solar radiation by the atmosphere during sunrise and sunset, from 240 to 2380 nm and at moderate spectral resolution. Vertical columns and profiles of a variety of atmospheric constituents are inferred from the SCIAMACHY radiometric measurements by dedicated retrieval algorithms. With the support of ESA and several international partners, a methodical SCIAMACHY validation programme has been developed jointly by Germany, the Netherlands and Belgium (the three instrument providing countries) to face complex requirements in terms of measured species, altitude range, spatial and temporal scales, geophysical states and intended scientific applications. This summary paper describes the approach adopted to address those requirements. The actual validation of the operational SCIAMACHY processors established at DLR on behalf of ESA has been hampered by data distribution and processor problems. Since first data releases in summer 2002, operational processors were upgraded regularly and some data products - level-1b spectra, level-2 O3, NO2, BrO and clouds data - have improved significantly. Validation results summarised in this paper conclude that for limited periods and geographical domains they can already be used for atmospheric research. Nevertheless, remaining processor problems cause major errors preventing from scientific usability in other periods and domains. Untied to the constraints of operational processing, seven scientific institutes (BIRA-IASB, IFE, IUP-Heidelberg, KNMI, MPI, SAO and SRON) have developed their own retrieval algorithms and generated SCIAMACHY data products, together addressing nearly all targeted constituents. Most of the UV-visible data products (both columns and profiles) already have acceptable, if not excellent, quality. Several near-infrared column products are still in development but they have already demonstrated their potential for a variety of applications. In any case, scientific users are advised to read carefully validation reports before using the data. It is required and anticipated that SCIAMACHY validation will continue throughout instrument lifetime and beyond. The actual amount of work will obviously depend on funding considerations.
Influence of reconstruction algorithms on image quality in SPECT myocardial perfusion imaging.
Davidsson, Anette; Olsson, Eva; Engvall, Jan; Gustafsson, Agnetha
2017-11-01
We investigated if image- and diagnostic quality in SPECT MPI could be maintained despite a reduced acquisition time adding Depth Dependent Resolution Recovery (DDRR) for image reconstruction. Images were compared with filtered back projection (FBP) and iterative reconstruction using Ordered Subsets Expectation Maximization with (IRAC) and without (IRNC) attenuation correction (AC). Stress- and rest imaging for 15 min was performed on 21 subjects with a dual head gamma camera (Infinia Hawkeye; GE Healthcare), ECG-gating with 8 frames/cardiac cycle and a low-dose CT-scan. A 9 min acquisition was generated using five instead of eight gated frames and was reconstructed with DDRR, with (IRACRR) and without AC (IRNCRR) as well as with FBP. Three experienced nuclear medicine specialists visually assessed anonymized images according to eight criteria on a four point scale, three related to image quality and five to diagnostic confidence. Statistical analysis was performed using Visual Grading Regression (VGR). Observer confidence in statements on image quality was highest for the images that were reconstructed using DDRR (P<0·01 compared to FBP). Iterative reconstruction without DDRR was not superior to FBP. Interobserver variability was significant for statements on image quality (P<0·05) but lower in the diagnostic statements on ischemia and scar. The confidence in assessing ischemia and scar was not different between the reconstruction techniques (P = n.s.). SPECT MPI collected in 9 min, reconstructed with DDRR and AC, produced better image quality than the standard procedure. The observers expressed the highest diagnostic confidence in the DDRR reconstruction. © 2016 Scandinavian Society of Clinical Physiology and Nuclear Medicine. Published by John Wiley & Sons Ltd.
ConnectX2 In niBand Management Queues: New support for Network Of oaded
DOE Office of Scientific and Technical Information (OSTI.GOV)
Graham, Richard L; Poole, Stephen W; Shamis, Pavel
2010-01-01
This paper introduces the newly developed InfiniBand (IB) Management Queue capability, used by the Host Channel Adapter (HCA) to manage network task data flow dependancies, and progress the communications associated with such flows. These tasks include sends, receives, and the newly supported wait task, and are scheduled by the HCA based on a data dependency description provided by the user. This functionality is supported by the ConnectX-2 HCA, and provides the means for delegating collective communication management and progress to the HCA, also known as collective communication offload. This provides a means for overlapping collective communications managed by the HCAmore » and computation on the Central Processing Unit (CPU), thus making it possible to reduce the impact of system noise on parallel applications using collective operations. This paper further describes how this new capability can be used to implement scalable Message Passing Interface (MPI) collective operations, describing the high level details of how this new capability is used to implement the MPI Barrier collective operation, focusing on the latency sensitive performance aspects of this new capability. This paper concludes with small scale benchmark experiments comparing implementations of the barrier collective operation, using the new network offload capabilities, with established point-to-point based implementations of these same algorithms, which manage the data flow using the central processing unit. These early results demonstrate the promise this new capability provides to improve the scalability of high-performance applications using collective communications. The latency of the HCA based implementation of the barrier is similar to that of the best performing point-to-point based implementation managed by the central processing unit, starting to outperform these as the number of processes involved in the collective operation increases.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Graham, Richard L; Poole, Stephen W; Shamis, Pavel
2010-01-01
This paper introduces the newly developed Infini-Band (IB) Management Queue capability, used by the Host Channel Adapter (HCA) to manage network task data flow dependancies, and progress the communications associated with such flows. These tasks include sends, receives, and the newly supported wait task, and are scheduled by the HCA based on a data dependency description provided by the user. This functionality is supported by the ConnectX-2 HCA, and provides the means for delegating collective communication management and progress to the HCA, also known as collective communication offload. This provides a means for overlapping collective communications managed by the HCAmore » and computation on the Central Processing Unit (CPU), thus making it possible to reduce the impact of system noise on parallel applications using collective operations. This paper further describes how this new capability can be used to implement scalable Message Passing Interface (MPI) collective operations, describing the high level details of how this new capability is used to implement the MPI Barrier collective operation, focusing on the latency sensitive performance aspects of this new capability. This paper concludes with small scale benchmark experiments comparing implementations of the barrier collective operation, using the new network offload capabilities, with established point-to-point based implementations of these same algorithms, which manage the data flow using the central processing unit. These early results demonstrate the promise this new capability provides to improve the scalability of high performance applications using collective communications. The latency of the HCA based implementation of the barrier is similar to that of the best performing point-to-point based implementation managed by the central processing unit, starting to outperform these as the number of processes involved in the collective operation increases.« less
Ice-sheet modelling accelerated by graphics cards
NASA Astrophysics Data System (ADS)
Brædstrup, Christian Fredborg; Damsgaard, Anders; Egholm, David Lundbek
2014-11-01
Studies of glaciers and ice sheets have increased the demand for high performance numerical ice flow models over the past decades. When exploring the highly non-linear dynamics of fast flowing glaciers and ice streams, or when coupling multiple flow processes for ice, water, and sediment, researchers are often forced to use super-computing clusters. As an alternative to conventional high-performance computing hardware, the Graphical Processing Unit (GPU) is capable of massively parallel computing while retaining a compact design and low cost. In this study, we present a strategy for accelerating a higher-order ice flow model using a GPU. By applying the newest GPU hardware, we achieve up to 180× speedup compared to a similar but serial CPU implementation. Our results suggest that GPU acceleration is a competitive option for ice-flow modelling when compared to CPU-optimised algorithms parallelised by the OpenMP or Message Passing Interface (MPI) protocols.
GANDALF - Graphical Astrophysics code for N-body Dynamics And Lagrangian Fluids
NASA Astrophysics Data System (ADS)
Hubber, D. A.; Rosotti, G. P.; Booth, R. A.
2018-01-01
GANDALF is a new hydrodynamics and N-body dynamics code designed for investigating planet formation, star formation and star cluster problems. GANDALF is written in C++, parallelized with both OPENMP and MPI and contains a PYTHON library for analysis and visualization. The code has been written with a fully object-oriented approach to easily allow user-defined implementations of physics modules or other algorithms. The code currently contains implementations of smoothed particle hydrodynamics, meshless finite-volume and collisional N-body schemes, but can easily be adapted to include additional particle schemes. We present in this paper the details of its implementation, results from the test suite, serial and parallel performance results and discuss the planned future development. The code is freely available as an open source project on the code-hosting website github at https://github.com/gandalfcode/gandalf and is available under the GPLv2 license.
Libsharp - spherical harmonic transforms revisited
NASA Astrophysics Data System (ADS)
Reinecke, M.; Seljebotn, D. S.
2013-06-01
We present libsharp, a code library for spherical harmonic transforms (SHTs), which evolved from the libpsht library and addresses several of its shortcomings, such as adding MPI support for distributed memory systems and SHTs of fields with arbitrary spin, but also supporting new developments in CPU instruction sets like the Advanced Vector Extensions (AVX) or fused multiply-accumulate (FMA) instructions. The library is implemented in portable C99 and provides an interface that can be easily accessed from other programming languages such as C++, Fortran, Python, etc. Generally, libsharp's performance is at least on par with that of its predecessor; however, significant improvements were made to the algorithms for scalar SHTs, which are roughly twice as fast when using the same CPU capabilities. The library is available at
A Comparison of Three Programming Models for Adaptive Applications
NASA Technical Reports Server (NTRS)
Shan, Hong-Zhang; Singh, Jaswinder Pal; Oliker, Leonid; Biswa, Rupak; Kwak, Dochan (Technical Monitor)
2000-01-01
We study the performance and programming effort for two major classes of adaptive applications under three leading parallel programming models. We find that all three models can achieve scalable performance on the state-of-the-art multiprocessor machines. The basic parallel algorithms needed for different programming models to deliver their best performance are similar, but the implementations differ greatly, far beyond the fact of using explicit messages versus implicit loads/stores. Compared with MPI and SHMEM, CC-SAS (cache-coherent shared address space) provides substantial ease of programming at the conceptual and program orchestration level, which often leads to the performance gain. However it may also suffer from the poor spatial locality of physically distributed shared data on large number of processors. Our CC-SAS implementation of the PARMETIS partitioner itself runs faster than in the other two programming models, and generates more balanced result for our application.
Raytracing and Direct-Drive Targets
NASA Astrophysics Data System (ADS)
Schmitt, Andrew J.; Bates, Jason; Fyfe, David; Eimerl, David
2013-10-01
Accurate simulation of the effects of laser imprinting and drive asymmetries in directly driven targets requires the ability to distinguish between raytrace noise and the intensity structure produced by the spatial and temporal incoherence of optical smoothing. We have developed and implemented a smoother raytrace algorithm for our mpi-parallel radiation hydrodynamics code, FAST3D. The underlying approach is to connect the rays into either sheets (in 2D) or volume-enclosing chunks (in 3D) so that the absorbed energy distribution continuously covers the propagation area illuminated by the laser. We will describe the status and show the different scalings encountered in 2D and 3D problems as the computational size, parallelization strategy, and number of rays is varied. Finally, we show results using the method in current NIKE experimental target simulations and in proposed symmetric and polar direct-drive target designs. Supported by US DoE/NNSA.
LAMMPS strong scaling performance optimization on Blue Gene/Q
DOE Office of Scientific and Technical Information (OSTI.GOV)
Coffman, Paul; Jiang, Wei; Romero, Nichols A.
2014-11-12
LAMMPS "Large-scale Atomic/Molecular Massively Parallel Simulator" is an open-source molecular dynamics package from Sandia National Laboratories. Significant performance improvements in strong-scaling and time-to-solution for this application on IBM's Blue Gene/Q have been achieved through computational optimizations of the OpenMP versions of the short-range Lennard-Jones term of the CHARMM force field and the long-range Coulombic interaction implemented with the PPPM (particle-particle-particle mesh) algorithm, enhanced by runtime parameter settings controlling thread utilization. Additionally, MPI communication performance improvements were made to the PPPM calculation by re-engineering the parallel 3D FFT to use MPICH collectives instead of point-to-point. Performance testing was done using anmore » 8.4-million atom simulation scaling up to 16 racks on the Mira system at Argonne Leadership Computing Facility (ALCF). Speedups resulting from this effort were in some cases over 2x.« less
Strong scaling of general-purpose molecular dynamics simulations on GPUs
NASA Astrophysics Data System (ADS)
Glaser, Jens; Nguyen, Trung Dac; Anderson, Joshua A.; Lui, Pak; Spiga, Filippo; Millan, Jaime A.; Morse, David C.; Glotzer, Sharon C.
2015-07-01
We describe a highly optimized implementation of MPI domain decomposition in a GPU-enabled, general-purpose molecular dynamics code, HOOMD-blue (Anderson and Glotzer, 2013). Our approach is inspired by a traditional CPU-based code, LAMMPS (Plimpton, 1995), but is implemented within a code that was designed for execution on GPUs from the start (Anderson et al., 2008). The software supports short-ranged pair force and bond force fields and achieves optimal GPU performance using an autotuning algorithm. We are able to demonstrate equivalent or superior scaling on up to 3375 GPUs in Lennard-Jones and dissipative particle dynamics (DPD) simulations of up to 108 million particles. GPUDirect RDMA capabilities in recent GPU generations provide better performance in full double precision calculations. For a representative polymer physics application, HOOMD-blue 1.0 provides an effective GPU vs. CPU node speed-up of 12.5 ×.
Implementation of Parallel Dynamic Simulation on Shared-Memory vs. Distributed-Memory Environments
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jin, Shuangshuang; Chen, Yousu; Wu, Di
2015-12-09
Power system dynamic simulation computes the system response to a sequence of large disturbance, such as sudden changes in generation or load, or a network short circuit followed by protective branch switching operation. It consists of a large set of differential and algebraic equations, which is computational intensive and challenging to solve using single-processor based dynamic simulation solution. High-performance computing (HPC) based parallel computing is a very promising technology to speed up the computation and facilitate the simulation process. This paper presents two different parallel implementations of power grid dynamic simulation using Open Multi-processing (OpenMP) on shared-memory platform, and Messagemore » Passing Interface (MPI) on distributed-memory clusters, respectively. The difference of the parallel simulation algorithms and architectures of the two HPC technologies are illustrated, and their performances for running parallel dynamic simulation are compared and demonstrated.« less