Sample records for self-consistent parallel multi

  1. Self-balanced modulation and magnetic rebalancing method for parallel multilevel inverters

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Li, Hui; Shi, Yanjun

    A self-balanced modulation method and a closed-loop magnetic flux rebalancing control method for parallel multilevel inverters. The combination of the two methods provides for balancing of the magnetic flux of the inter-cell transformers (ICTs) of the parallel multilevel inverters without deteriorating the quality of the output voltage. In various embodiments a parallel multi-level inverter modulator is provide including a multi-channel comparator to generate a multiplexed digitized ideal waveform for a parallel multi-level inverter and a finite state machine (FSM) module coupled to the parallel multi-channel comparator, the FSM module to receive the multiplexed digitized ideal waveform and to generate amore » pulse width modulated gate-drive signal for each switching device of the parallel multi-level inverter. The system and method provides for optimization of the output voltage spectrum without influence the magnetic balancing.« less

  2. Self-consistent Simulations and Analysis of the Coupled-Bunch Instability for Arbitrary Multi-Bunch Configurations

    DOE PAGES

    Bassi, Gabriele; Blednykh, Alexei; Smalyuk, Victor

    2016-02-24

    A novel algorithm for self-consistent simulations of long-range wakefield effects has been developed and applied to the study of both longitudinal and transverse coupled-bunch instabilities at NSLS-II. The algorithm is implemented in the new parallel tracking code space (self-consistent parallel algorithm for collective effects) discussed in the paper. The code is applicable for accurate beam dynamics simulations in cases where both bunch-to-bunch and intrabunch motions need to be taken into account, such as chromatic head-tail effects on the coupled-bunch instability of a beam with a nonuniform filling pattern, or multibunch and single-bunch effects of a passive higher-harmonic cavity. The numericalmore » simulations have been compared with analytical studies. For a beam with an arbitrary filling pattern, intensity-dependent complex frequency shifts have been derived starting from a system of coupled Vlasov equations. The analytical formulas and numerical simulations confirm that the analysis is reduced to the formulation of an eigenvalue problem based on the known formulas of the complex frequency shifts for the uniform filling pattern case.« less

  3. Fast l₁-SPIRiT compressed sensing parallel imaging MRI: scalable parallel implementation and clinically feasible runtime.

    PubMed

    Murphy, Mark; Alley, Marcus; Demmel, James; Keutzer, Kurt; Vasanawala, Shreyas; Lustig, Michael

    2012-06-01

    We present l₁-SPIRiT, a simple algorithm for auto calibrating parallel imaging (acPI) and compressed sensing (CS) that permits an efficient implementation with clinically-feasible runtimes. We propose a CS objective function that minimizes cross-channel joint sparsity in the wavelet domain. Our reconstruction minimizes this objective via iterative soft-thresholding, and integrates naturally with iterative self-consistent parallel imaging (SPIRiT). Like many iterative magnetic resonance imaging reconstructions, l₁-SPIRiT's image quality comes at a high computational cost. Excessively long runtimes are a barrier to the clinical use of any reconstruction approach, and thus we discuss our approach to efficiently parallelizing l₁-SPIRiT and to achieving clinically-feasible runtimes. We present parallelizations of l₁-SPIRiT for both multi-GPU systems and multi-core CPUs, and discuss the software optimization and parallelization decisions made in our implementation. The performance of these alternatives depends on the processor architecture, the size of the image matrix, and the number of parallel imaging channels. Fundamentally, achieving fast runtime requires the correct trade-off between cache usage and parallelization overheads. We demonstrate image quality via a case from our clinical experimentation, using a custom 3DFT spoiled gradient echo (SPGR) sequence with up to 8× acceleration via Poisson-disc undersampling in the two phase-encoded directions.

  4. Fast ℓ1-SPIRiT Compressed Sensing Parallel Imaging MRI: Scalable Parallel Implementation and Clinically Feasible Runtime

    PubMed Central

    Murphy, Mark; Alley, Marcus; Demmel, James; Keutzer, Kurt; Vasanawala, Shreyas; Lustig, Michael

    2012-01-01

    We present ℓ1-SPIRiT, a simple algorithm for auto calibrating parallel imaging (acPI) and compressed sensing (CS) that permits an efficient implementation with clinically-feasible runtimes. We propose a CS objective function that minimizes cross-channel joint sparsity in the Wavelet domain. Our reconstruction minimizes this objective via iterative soft-thresholding, and integrates naturally with iterative Self-Consistent Parallel Imaging (SPIRiT). Like many iterative MRI reconstructions, ℓ1-SPIRiT’s image quality comes at a high computational cost. Excessively long runtimes are a barrier to the clinical use of any reconstruction approach, and thus we discuss our approach to efficiently parallelizing ℓ1-SPIRiT and to achieving clinically-feasible runtimes. We present parallelizations of ℓ1-SPIRiT for both multi-GPU systems and multi-core CPUs, and discuss the software optimization and parallelization decisions made in our implementation. The performance of these alternatives depends on the processor architecture, the size of the image matrix, and the number of parallel imaging channels. Fundamentally, achieving fast runtime requires the correct trade-off between cache usage and parallelization overheads. We demonstrate image quality via a case from our clinical experimentation, using a custom 3DFT Spoiled Gradient Echo (SPGR) sequence with up to 8× acceleration via poisson-disc undersampling in the two phase-encoded directions. PMID:22345529

  5. Adaptive multi-resolution 3D Hartree-Fock-Bogoliubov solver for nuclear structure

    NASA Astrophysics Data System (ADS)

    Pei, J. C.; Fann, G. I.; Harrison, R. J.; Nazarewicz, W.; Shi, Yue; Thornton, S.

    2014-08-01

    Background: Complex many-body systems, such as triaxial and reflection-asymmetric nuclei, weakly bound halo states, cluster configurations, nuclear fragments produced in heavy-ion fusion reactions, cold Fermi gases, and pasta phases in neutron star crust, are all characterized by large sizes and complex topologies in which many geometrical symmetries characteristic of ground-state configurations are broken. A tool of choice to study such complex forms of matter is an adaptive multi-resolution wavelet analysis. This method has generated much excitement since it provides a common framework linking many diversified methodologies across different fields, including signal processing, data compression, harmonic analysis and operator theory, fractals, and quantum field theory. Purpose: To describe complex superfluid many-fermion systems, we introduce an adaptive pseudospectral method for solving self-consistent equations of nuclear density functional theory in three dimensions, without symmetry restrictions. Methods: The numerical method is based on the multi-resolution and computational harmonic analysis techniques with a multi-wavelet basis. The application of state-of-the-art parallel programming techniques include sophisticated object-oriented templates which parse the high-level code into distributed parallel tasks with a multi-thread task queue scheduler for each multi-core node. The internode communications are asynchronous. The algorithm is variational and is capable of solving coupled complex-geometric systems of equations adaptively, with functional and boundary constraints, in a finite spatial domain of very large size, limited by existing parallel computer memory. For smooth functions, user-defined finite precision is guaranteed. Results: The new adaptive multi-resolution Hartree-Fock-Bogoliubov (HFB) solver madness-hfb is benchmarked against a two-dimensional coordinate-space solver hfb-ax that is based on the B-spline technique and a three-dimensional solver hfodd that is based on the harmonic-oscillator basis expansion. Several examples are considered, including the self-consistent HFB problem for spin-polarized trapped cold fermions and the Skyrme-Hartree-Fock (+BCS) problem for triaxial deformed nuclei. Conclusions: The new madness-hfb framework has many attractive features when applied to nuclear and atomic problems involving many-particle superfluid systems. Of particular interest are weakly bound nuclear configurations close to particle drip lines, strongly elongated and dinuclear configurations such as those present in fission and heavy-ion fusion, and exotic pasta phases that appear in neutron star crust.

  6. Non-Cartesian Parallel Imaging Reconstruction

    PubMed Central

    Wright, Katherine L.; Hamilton, Jesse I.; Griswold, Mark A.; Gulani, Vikas; Seiberlich, Nicole

    2014-01-01

    Non-Cartesian parallel imaging has played an important role in reducing data acquisition time in MRI. The use of non-Cartesian trajectories can enable more efficient coverage of k-space, which can be leveraged to reduce scan times. These trajectories can be undersampled to achieve even faster scan times, but the resulting images may contain aliasing artifacts. Just as Cartesian parallel imaging can be employed to reconstruct images from undersampled Cartesian data, non-Cartesian parallel imaging methods can mitigate aliasing artifacts by using additional spatial encoding information in the form of the non-homogeneous sensitivities of multi-coil phased arrays. This review will begin with an overview of non-Cartesian k-space trajectories and their sampling properties, followed by an in-depth discussion of several selected non-Cartesian parallel imaging algorithms. Three representative non-Cartesian parallel imaging methods will be described, including Conjugate Gradient SENSE (CG SENSE), non-Cartesian GRAPPA, and Iterative Self-Consistent Parallel Imaging Reconstruction (SPIRiT). After a discussion of these three techniques, several potential promising clinical applications of non-Cartesian parallel imaging will be covered. PMID:24408499

  7. Addressable multi-nozzle electrohydrodynamic jet printing with high consistency by multi-level voltage method

    NASA Astrophysics Data System (ADS)

    Pan, Yanqiao; Huang, YongAn; Guo, Lei; Ding, Yajiang; Yin, Zhouping

    2015-04-01

    It is critical and challenging to achieve the individual jetting ability and high consistency in multi-nozzle electrohydrodynamic jet printing (E-jet printing). We proposed multi-level voltage method (MVM) to implement the addressable E-jet printing using multiple parallel nozzles with high consistency. The fabricated multi-nozzle printhead for MVM consists of three parts: PMMA holder, stainless steel capillaries (27G, outer diameter 400 μm) and FR-4 extractor layer. The key of MVM is to control the maximum meniscus electric field on each nozzle. The individual jetting control can be implemented when the rings under the jetting nozzles are 0 kV and the other rings are 0.5 kV. The onset electric field for each nozzle is ˜3.4 kV/mm by numerical simulation. Furthermore, a series of printing experiments are performed to show the advantage of MVM in printing consistency than the "one-voltage method" and "improved E-jet method", by combination with finite element analyses. The good dimension consistency (274μm, 276μm, 280μm) and position consistency of the droplet array on the hydrophobic Si substrate verified the enhancements. It shows that MVM is an effective technique to implement the addressable E-jet printing with multiple parallel nozzles in high consistency.

  8. A hybrid framework of first principles molecular orbital calculations and a three-dimensional integral equation theory for molecular liquids: Multi-center molecular Ornstein-Zernike self-consistent field approach

    NASA Astrophysics Data System (ADS)

    Kido, Kentaro; Kasahara, Kento; Yokogawa, Daisuke; Sato, Hirofumi

    2015-07-01

    In this study, we reported the development of a new quantum mechanics/molecular mechanics (QM/MM)-type framework to describe chemical processes in solution by combining standard molecular-orbital calculations with a three-dimensional formalism of integral equation theory for molecular liquids (multi-center molecular Ornstein-Zernike (MC-MOZ) method). The theoretical procedure is very similar to the 3D-reference interaction site model self-consistent field (RISM-SCF) approach. Since the MC-MOZ method is highly parallelized for computation, the present approach has the potential to be one of the most efficient procedures to treat chemical processes in solution. Benchmark tests to check the validity of this approach were performed for two solute (solute water and formaldehyde) systems and a simple SN2 reaction (Cl- + CH3Cl → ClCH3 + Cl-) in aqueous solution. The results for solute molecular properties and solvation structures obtained by the present approach were in reasonable agreement with those obtained by other hybrid frameworks and experiments. In particular, the results of the proposed approach are in excellent agreements with those of 3D-RISM-SCF.

  9. A hybrid framework of first principles molecular orbital calculations and a three-dimensional integral equation theory for molecular liquids: multi-center molecular Ornstein-Zernike self-consistent field approach.

    PubMed

    Kido, Kentaro; Kasahara, Kento; Yokogawa, Daisuke; Sato, Hirofumi

    2015-07-07

    In this study, we reported the development of a new quantum mechanics/molecular mechanics (QM/MM)-type framework to describe chemical processes in solution by combining standard molecular-orbital calculations with a three-dimensional formalism of integral equation theory for molecular liquids (multi-center molecular Ornstein-Zernike (MC-MOZ) method). The theoretical procedure is very similar to the 3D-reference interaction site model self-consistent field (RISM-SCF) approach. Since the MC-MOZ method is highly parallelized for computation, the present approach has the potential to be one of the most efficient procedures to treat chemical processes in solution. Benchmark tests to check the validity of this approach were performed for two solute (solute water and formaldehyde) systems and a simple SN2 reaction (Cl(-) + CH3Cl → ClCH3 + Cl(-)) in aqueous solution. The results for solute molecular properties and solvation structures obtained by the present approach were in reasonable agreement with those obtained by other hybrid frameworks and experiments. In particular, the results of the proposed approach are in excellent agreements with those of 3D-RISM-SCF.

  10. Structural setting and kinematics of Nubian fault system, SE Western Desert, Egypt: An example of multi-reactivated intraplate strike-slip faults

    NASA Astrophysics Data System (ADS)

    Sakran, Shawky; Said, Said Mohamed

    2018-02-01

    Detailed surface geological mapping and subsurface seismic interpretation have been integrated to unravel the structural style and kinematic history of the Nubian Fault System (NFS). The NFS consists of several E-W Principal Deformation Zones (PDZs) (e.g. Kalabsha fault). Each PDZ is defined by spectacular E-W, WNW and ENE dextral strike-slip faults, NNE sinistral strike-slip faults, NE to ENE folds, and NNW normal faults. Each fault zone has typical self-similar strike-slip architecture comprising multi-scale fault segments. Several multi-scale uplifts and basins were developed at the step-over zones between parallel strike-slip fault segments as a result of local extension or contraction. The NNE faults consist of right-stepping sinistral strike-slip fault segments (e.g. Sin El Kiddab fault). The NNE sinistral faults extend for long distances ranging from 30 to 100 kms and cut one or two E-W PDZs. Two nearly perpendicular strike-slip tectonic regimes are recognized in the NFS; an inactive E-W Late Cretaceous - Early Cenozoic dextral transpression and an active NNE sinistral shear.

  11. Three-Component Reaction Discovery Enabled by Mass Spectrometry of Self-Assembled Monolayers

    PubMed Central

    Montavon, Timothy J.; Li, Jing; Cabrera-Pardo, Jaime R.; Mrksich, Milan; Kozmin, Sergey A.

    2011-01-01

    Multi-component reactions have been extensively employed in many areas of organic chemistry. Despite significant progress, the discovery of such enabling transformations remains challenging. Here, we present the development of a parallel, label-free reaction-discovery platform, which can be used for identification of new multi-component transformations. Our approach is based on the parallel mass spectrometric screening of interfacial chemical reactions on arrays of self-assembled monolayers. This strategy enabled the identification of a simple organic phosphine that can catalyze a previously unknown condensation of siloxy alkynes, aldehydes and amines to produce 3-hydroxy amides with high efficiency and diastereoselectivity. The reaction was further optimized using solution phase methods. PMID:22169871

  12. Shape, zonal winds and gravitational field of Jupiter: a fully self-consistent, multi-layered model

    NASA Astrophysics Data System (ADS)

    Schubert, Gerald; Kong, Dali; Zhang, Keke

    2016-10-01

    We construct a three-dimensional, finite-element, fully self-consistent, multi-layered,non-spheroidal model of Jupiter consisting of an inner core, a metallic electrically conducting dynamo region and an outer molecular electrically insulating envelope. We assume that the Jovian zonal winds are on cylinders parallel to the rotation axis but, due to the effect of magnetic braking, are confined within the outer molecular envelope. Two related calculations are carried out. The first provides an accurate description of the shape and internal density profile of Jupiter; the effect of rotational distortion is not treated as a small perturbation on a spherically symmetric state. This calculation determines the density, size and shape of the inner core, the irregular shape of the 1-bar pressure level, and the internal structure of Jupiter; the full effect of rotational distortion, without the influence of the zonal winds, is accounted for. Our multi-layered model is able to produce the known mass, the known equatorial and polar radii, and the known zonal gravitational coefficient J2 of Jupiter within their error bars; it also yields the coefficients J4 and J6 within about 5% accuracy, and the core equatorial radius 0.09RJ containing 3.73 Earth masses.The second calculation determines the variation of the gravitational field caused solely by the effect of the zonal winds on the rotationally distorted non-spheroidal Jupiter. Four different cases, ranging from a deep wind profile to a very shallow profile, are considered and implications for accurate interpretation of the zonal gravitational coefficients expected from the Juno mission are discussed.

  13. PADF RF localization experiments with multi-agent caged-MAV platforms

    NASA Astrophysics Data System (ADS)

    Barber, Christopher; Gates, Miguel; Selmic, Rastko; Al-Issa, Huthaifa; Ordonez, Raul; Mitra, Atindra

    2011-06-01

    This paper provides a summary of preliminary RF direction finding results generated within an AFOSR funded testbed facility recently developed at Louisiana Tech University. This facility, denoted as the Louisiana Tech University Micro- Aerial Vehicle/Wireless Sensor Network (MAVSeN) Laboratory, has recently acquired a number of state-of-the-art MAV platforms that enable us to analyze, design, and test some of our recent results in the area of multiplatform position-adaptive direction finding (PADF) [1] [2] for localization of RF emitters in challenging embedded multipath environments. Discussions within the segmented sections of this paper include a description of the MAVSeN Laboratory and the preliminary results from the implementation of mobile platforms with the PADF algorithm. This novel approach to multi-platform RF direction finding is based on the investigation of iterative path-loss based (i.e. path loss exponent) metrics estimates that are measured across multiple platforms in order to develop a control law that robotically/intelligently positionally adapt (i.e. self-adjust) the location of each distributed/cooperative platform. The body of this paper provides a summary of our recent results on PADF and includes a discussion on state-of-the-art Sensor Mote Technologies as applied towards the development of sensor-integrated caged-MAV platform for PADF applications. Also, a discussion of recent experimental results that incorporate sample approaches to real-time singleplatform data pruning is included as part of a discussion on potential approaches to refining a basic PADF technique in order to integrate and perform distributed self-sensitivity and self-consistency analysis as part of a PADF technique with distributed robotic/intelligent features. These techniques are extracted in analytical form from a parallel study denoted as "PADF RF Localization Criteria for Multi-Model Scattering Environments". The focus here is on developing and reporting specific approaches to self-sensitivity and self-consistency within this experimental PADF framework via the exploitation of specific single-agent caged-MAV trajectories that are unique to this experiment set.

  14. Multi-LED parallel transmission for long distance underwater VLC system with one SPAD receiver

    NASA Astrophysics Data System (ADS)

    Wang, Chao; Yu, Hong-Yi; Zhu, Yi-Jun; Wang, Tao; Ji, Ya-Wei

    2018-03-01

    In this paper, a multiple light emitting diode (LED) chips parallel transmission (Multi-LED-PT) scheme for underwater visible light communication system with one photon-counting single photon avalanche diode (SPAD) receiver is proposed. As the lamp always consists of multi-LED chips, the data rate could be improved when we drive these multi-LED chips parallel by using the interleaver-division-multiplexing technique. For each chip, the on-off-keying modulation is used to reduce the influence of clipping. Then a serial successive interference cancellation detection algorithm based on ideal Poisson photon-counting channel by the SPAD is proposed. Finally, compared to the SPAD-based direct current-biased optical orthogonal frequency division multiplexing system, the proposed Multi-LED-PT system could improve the error-rate performance and anti-nonlinearity performance significantly under the effects of absorption, scattering and weak turbulence-induced channel fading together.

  15. Transitioning NWChem to the Next Generation of Manycore Machines

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bylaska, Eric J.; Apra, Edoardo; Kowalski, Karol

    The NorthWest Chemistry (NWChem) modeling software is a popular molecular chemistry simulation software that was designed from the start to work on massively parallel processing supercomputers[6, 28, 49]. It contains an umbrella of modules that today includes Self Consistent Field (SCF), second order Mller-Plesset perturbation theory (MP2), Coupled Cluster, multi-conguration selfconsistent eld (MCSCF), selected conguration interaction (CI), tensor contraction engine (TCE) many body methods, density functional theory (DFT), time-dependent density functional theory (TDDFT), real time time-dependent density functional theory, pseudopotential plane-wave density functional theory (PSPW), band structure (BAND), ab initio molecular dynamics, Car-Parrinello molecular dynamics, classical molecular dynamics (MD), QM/MM,more » AIMD/MM, GIAO NMR, COSMO, COSMO-SMD, and RISM solvation models, free energy simulations, reaction path optimization, parallel in time, among other capabilities[ 22]. Moreover new capabilities continue to be added with each new release.« less

  16. Parallel machine architecture for production rule systems

    DOEpatents

    Allen, Jr., John D.; Butler, Philip L.

    1989-01-01

    A parallel processing system for production rule programs utilizes a host processor for storing production rule right hand sides (RHS) and a plurality of rule processors for storing left hand sides (LHS). The rule processors operate in parallel in the recognize phase of the system recognize -Act Cycle to match their respective LHS's against a stored list of working memory elements (WME) in order to find a self consistent set of WME's. The list of WME is dynamically varied during the Act phase of the system in which the host executes or fires rule RHS's for those rules for which a self-consistent set has been found by the rule processors. The host transmits instructions for creating or deleting working memory elements as dictated by the rule firings until the rule processors are unable to find any further self-consistent working memory element sets at which time the production rule system is halted.

  17. ParallelStructure: A R Package to Distribute Parallel Runs of the Population Genetics Program STRUCTURE on Multi-Core Computers

    PubMed Central

    Besnier, Francois; Glover, Kevin A.

    2013-01-01

    This software package provides an R-based framework to make use of multi-core computers when running analyses in the population genetics program STRUCTURE. It is especially addressed to those users of STRUCTURE dealing with numerous and repeated data analyses, and who could take advantage of an efficient script to automatically distribute STRUCTURE jobs among multiple processors. It also consists of additional functions to divide analyses among combinations of populations within a single data set without the need to manually produce multiple projects, as it is currently the case in STRUCTURE. The package consists of two main functions: MPI_structure() and parallel_structure() as well as an example data file. We compared the performance in computing time for this example data on two computer architectures and showed that the use of the present functions can result in several-fold improvements in terms of computation time. ParallelStructure is freely available at https://r-forge.r-project.org/projects/parallstructure/. PMID:23923012

  18. Template-directed atomically precise self-organization of perfectly ordered parallel cerium silicide nanowire arrays on Si(110)-16 × 2 surfaces.

    PubMed

    Hong, Ie-Hong; Liao, Yung-Cheng; Tsai, Yung-Feng

    2013-11-05

    The perfectly ordered parallel arrays of periodic Ce silicide nanowires can self-organize with atomic precision on single-domain Si(110)-16 × 2 surfaces. The growth evolution of self-ordered parallel Ce silicide nanowire arrays is investigated over a broad range of Ce coverages on single-domain Si(110)-16 × 2 surfaces by scanning tunneling microscopy (STM). Three different types of well-ordered parallel arrays, consisting of uniformly spaced and atomically identical Ce silicide nanowires, are self-organized through the heteroepitaxial growth of Ce silicides on a long-range grating-like 16 × 2 reconstruction at the deposition of various Ce coverages. Each atomically precise Ce silicide nanowire consists of a bundle of chains and rows with different atomic structures. The atomic-resolution dual-polarity STM images reveal that the interchain coupling leads to the formation of the registry-aligned chain bundles within individual Ce silicide nanowire. The nanowire width and the interchain coupling can be adjusted systematically by varying the Ce coverage on a Si(110) surface. This natural template-directed self-organization of perfectly regular parallel nanowire arrays allows for the precise control of the feature size and positions within ±0.2 nm over a large area. Thus, it is a promising route to produce parallel nanowire arrays in a straightforward, low-cost, high-throughput process.

  19. Template-directed atomically precise self-organization of perfectly ordered parallel cerium silicide nanowire arrays on Si(110)-16 × 2 surfaces

    PubMed Central

    2013-01-01

    The perfectly ordered parallel arrays of periodic Ce silicide nanowires can self-organize with atomic precision on single-domain Si(110)-16 × 2 surfaces. The growth evolution of self-ordered parallel Ce silicide nanowire arrays is investigated over a broad range of Ce coverages on single-domain Si(110)-16 × 2 surfaces by scanning tunneling microscopy (STM). Three different types of well-ordered parallel arrays, consisting of uniformly spaced and atomically identical Ce silicide nanowires, are self-organized through the heteroepitaxial growth of Ce silicides on a long-range grating-like 16 × 2 reconstruction at the deposition of various Ce coverages. Each atomically precise Ce silicide nanowire consists of a bundle of chains and rows with different atomic structures. The atomic-resolution dual-polarity STM images reveal that the interchain coupling leads to the formation of the registry-aligned chain bundles within individual Ce silicide nanowire. The nanowire width and the interchain coupling can be adjusted systematically by varying the Ce coverage on a Si(110) surface. This natural template-directed self-organization of perfectly regular parallel nanowire arrays allows for the precise control of the feature size and positions within ±0.2 nm over a large area. Thus, it is a promising route to produce parallel nanowire arrays in a straightforward, low-cost, high-throughput process. PMID:24188092

  20. The parallel algorithm for the 2D discrete wavelet transform

    NASA Astrophysics Data System (ADS)

    Barina, David; Najman, Pavel; Kleparnik, Petr; Kula, Michal; Zemcik, Pavel

    2018-04-01

    The discrete wavelet transform can be found at the heart of many image-processing algorithms. Until now, the transform on general-purpose processors (CPUs) was mostly computed using a separable lifting scheme. As the lifting scheme consists of a small number of operations, it is preferred for processing using single-core CPUs. However, considering a parallel processing using multi-core processors, this scheme is inappropriate due to a large number of steps. On such architectures, the number of steps corresponds to the number of points that represent the exchange of data. Consequently, these points often form a performance bottleneck. Our approach appropriately rearranges calculations inside the transform, and thereby reduces the number of steps. In other words, we propose a new scheme that is friendly to parallel environments. When evaluating on multi-core CPUs, we consistently overcome the original lifting scheme. The evaluation was performed on 61-core Intel Xeon Phi and 8-core Intel Xeon processors.

  1. Syndromes of Self-Reported Psychopathology for Ages 18-59 in 29 Societies.

    PubMed

    Ivanova, Masha Y; Achenbach, Thomas M; Rescorla, Leslie A; Tumer, Lori V; Ahmeti-Pronaj, Adelina; Au, Alma; Maese, Carmen Avila; Bellina, Monica; Caldas, J Carlos; Chen, Yi-Chuen; Csemy, Ladislav; da Rocha, Marina M; Decoster, Jeroen; Dobrean, Anca; Ezpeleta, Lourdes; Fontaine, Johnny R J; Funabiki, Yasuko; Guðmundsson, Halldór S; Harder, Valerie S; de la Cabada, Marie Leiner; Leung, Patrick; Liu, Jianghong; Mahr, Safia; Malykh, Sergey; Maras, Jelena Srdanovic; Markovic, Jasminka; Ndetei, David M; Oh, Kyung Ja; Petot, Jean-Michel; Riad, Geylan; Sakarya, Direnc; Samaniego, Virginia C; Sebre, Sandra; Shahini, Mimoza; Silvares, Edwiges; Simulioniene, Roma; Sokoli, Elvisa; Talcott, Joel B; Vazquez, Natalia; Zasepa, Ewa

    2015-06-01

    This study tested the multi-society generalizability of an eight-syndrome assessment model derived from factor analyses of American adults' self-ratings of 120 behavioral, emotional, and social problems. The Adult Self-Report (ASR; Achenbach and Rescorla 2003) was completed by 17,152 18-59-year-olds in 29 societies. Confirmatory factor analyses tested the fit of self-ratings in each sample to the eight-syndrome model. The primary model fit index (Root Mean Square Error of Approximation) showed good model fit for all samples, while secondary indices showed acceptable to good fit. Only 5 (0.06%) of the 8,598 estimated parameters were outside the admissible parameter space. Confidence intervals indicated that sampling fluctuations could account for the deviant parameters. Results thus supported the tested model in societies differing widely in social, political, and economic systems, languages, ethnicities, religions, and geographical regions. Although other items, societies, and analytic methods might yield different results, the findings indicate that adults in very diverse societies were willing and able to rate themselves on the same standardized set of 120 problem items. Moreover, their self-ratings fit an eight-syndrome model previously derived from self-ratings by American adults. The support for the statistically derived syndrome model is consistent with previous findings for parent, teacher, and self-ratings of 1½-18-year-olds in many societies. The ASR and its parallel collateral-report instrument, the Adult Behavior Checklist (ABCL), may offer mental health professionals practical tools for the multi-informant assessment of clinical constructs of adult psychopathology that appear to be meaningful across diverse societies.

  2. Self Organized Multi Agent Swarms (SOMAS) for Network Security Control

    DTIC Science & Technology

    2009-03-01

    Normal hierarchy vs entangled hierarchy 2.5.7 Quantifying Entangledness . While self organization means that the swarm develops a consistent structure of...flexibility due to centralization of control and com- munication. Thus, self organized, entangled hierarchy multi-agent swarms are evolved in this study to...technique. The resulting design exhibits a self organized multi-agent swarm (SOMAS) with entangled hierarchical control and communication through the

  3. Multi-thread parallel algorithm for reconstructing 3D large-scale porous structures

    NASA Astrophysics Data System (ADS)

    Ju, Yang; Huang, Yaohui; Zheng, Jiangtao; Qian, Xu; Xie, Heping; Zhao, Xi

    2017-04-01

    Geomaterials inherently contain many discontinuous, multi-scale, geometrically irregular pores, forming a complex porous structure that governs their mechanical and transport properties. The development of an efficient reconstruction method for representing porous structures can significantly contribute toward providing a better understanding of the governing effects of porous structures on the properties of porous materials. In order to improve the efficiency of reconstructing large-scale porous structures, a multi-thread parallel scheme was incorporated into the simulated annealing reconstruction method. In the method, four correlation functions, which include the two-point probability function, the linear-path functions for the pore phase and the solid phase, and the fractal system function for the solid phase, were employed for better reproduction of the complex well-connected porous structures. In addition, a random sphere packing method and a self-developed pre-conditioning method were incorporated to cast the initial reconstructed model and select independent interchanging pairs for parallel multi-thread calculation, respectively. The accuracy of the proposed algorithm was evaluated by examining the similarity between the reconstructed structure and a prototype in terms of their geometrical, topological, and mechanical properties. Comparisons of the reconstruction efficiency of porous models with various scales indicated that the parallel multi-thread scheme significantly shortened the execution time for reconstruction of a large-scale well-connected porous model compared to a sequential single-thread procedure.

  4. Combining multi-atlas segmentation with brain surface estimation

    NASA Astrophysics Data System (ADS)

    Huo, Yuankai; Carass, Aaron; Resnick, Susan M.; Pham, Dzung L.; Prince, Jerry L.; Landman, Bennett A.

    2016-03-01

    Whole brain segmentation (with comprehensive cortical and subcortical labels) and cortical surface reconstruction are two essential techniques for investigating the human brain. The two tasks are typically conducted independently, however, which leads to spatial inconsistencies and hinders further integrated cortical analyses. To obtain self-consistent whole brain segmentations and surfaces, FreeSurfer segregates the subcortical and cortical segmentations before and after the cortical surface reconstruction. However, this "segmentation to surface to parcellation" strategy has shown limitation in various situations. In this work, we propose a novel "multi-atlas segmentation to surface" method called Multi-atlas CRUISE (MaCRUISE), which achieves self-consistent whole brain segmentations and cortical surfaces by combining multi-atlas segmentation with the cortical reconstruction method CRUISE. To our knowledge, this is the first work that achieves the reliability of state-of-the-art multi-atlas segmentation and labeling methods together with accurate and consistent cortical surface reconstruction. Compared with previous methods, MaCRUISE has three features: (1) MaCRUISE obtains 132 cortical/subcortical labels simultaneously from a single multi-atlas segmentation before reconstructing volume consistent surfaces; (2) Fuzzy tissue memberships are combined with multi-atlas segmentations to address partial volume effects; (3) MaCRUISE reconstructs topologically consistent cortical surfaces by using the sulci locations from multi-atlas segmentation. Two data sets, one consisting of five subjects with expertly traced landmarks and the other consisting of 100 volumes from elderly subjects are used for validation. Compared with CRUISE, MaCRUISE achieves self-consistent whole brain segmentation and cortical reconstruction without compromising on surface accuracy. MaCRUISE is comparably accurate to FreeSurfer while achieving greater robustness across an elderly population.

  5. Combining Multi-atlas Segmentation with Brain Surface Estimation.

    PubMed

    Huo, Yuankai; Carass, Aaron; Resnick, Susan M; Pham, Dzung L; Prince, Jerry L; Landman, Bennett A

    2016-02-27

    Whole brain segmentation (with comprehensive cortical and subcortical labels) and cortical surface reconstruction are two essential techniques for investigating the human brain. The two tasks are typically conducted independently, however, which leads to spatial inconsistencies and hinders further integrated cortical analyses. To obtain self-consistent whole brain segmentations and surfaces, FreeSurfer segregates the subcortical and cortical segmentations before and after the cortical surface reconstruction. However, this "segmentation to surface to parcellation" strategy has shown limitations in various situations. In this work, we propose a novel "multi-atlas segmentation to surface" method called Multi-atlas CRUISE (MaCRUISE), which achieves self-consistent whole brain segmentations and cortical surfaces by combining multi-atlas segmentation with the cortical reconstruction method CRUISE. To our knowledge, this is the first work that achieves the reliability of state-of-the-art multi-atlas segmentation and labeling methods together with accurate and consistent cortical surface reconstruction. Compared with previous methods, MaCRUISE has three features: (1) MaCRUISE obtains 132 cortical/subcortical labels simultaneously from a single multi-atlas segmentation before reconstructing volume consistent surfaces; (2) Fuzzy tissue memberships are combined with multi-atlas segmentations to address partial volume effects; (3) MaCRUISE reconstructs topologically consistent cortical surfaces by using the sulci locations from multi-atlas segmentation. Two data sets, one consisting of five subjects with expertly traced landmarks and the other consisting of 100 volumes from elderly subjects are used for validation. Compared with CRUISE, MaCRUISE achieves self-consistent whole brain segmentation and cortical reconstruction without compromising on surface accuracy. MaCRUISE is comparably accurate to FreeSurfer while achieving greater robustness across an elderly population.

  6. Multiple Facets of Self-Control in Arab Adolescents: Parallel Pathways to Greater Happiness and Less Physical Aggression

    ERIC Educational Resources Information Center

    Gavriel-Fried, Belle; Ronen, Tammie; Agbaria, Qutaiba; Orkibi, Hod; Hamama, Liat

    2018-01-01

    Adolescence is a period of dramatic change that necessitates using skills and strengths to reduce physical aggression and increase happiness. This study examined the multiple facets of self-control skills in achieving both goals simultaneously, in a sample of 248 Arab adolescents in Israel. We conceptualized and tested a new multi-mediator model…

  7. Syndromes of Self-Reported Psychopathology for Ages 18–59 in 29 Societies

    PubMed Central

    Achenbach, Thomas M.; Rescorla, Leslie A.; Tumer, Lori V.; Ahmeti-Pronaj, Adelina; Au, Alma; Maese, Carmen Avila; Bellina, Monica; Caldas, J. Carlos; Chen, Yi-Chuen; Csemy, Ladislav; da Rocha, Marina M.; Decoster, Jeroen; Dobrean, Anca; Ezpeleta, Lourdes; Fontaine, Johnny R. J.; Funabiki, Yasuko; Guðmundsson, Halldór S.; Harder, Valerie s; de la Cabada, Marie Leiner; Leung, Patrick; Liu, Jianghong; Mahr, Safia; Malykh, Sergey; Maras, Jelena Srdanovic; Markovic, Jasminka; Ndetei, David M.; Oh, Kyung Ja; Petot, Jean-Michel; Riad, Geylan; Sakarya, Direnc; Samaniego, Virginia C.; Sebre, Sandra; Shahini, Mimoza; Silvares, Edwiges; Simulioniene, Roma; Sokoli, Elvisa; Talcott, Joel B.; Vazquez, Natalia; Zasepa, Ewa

    2017-01-01

    This study tested the multi-society generalizability of an eight-syndrome assessment model derived from factor analyses of American adults’ self-ratings of 120 behavioral, emotional, and social problems. The Adult Self-Report (ASR; Achenbach and Rescorla 2003) was completed by 17,152 18–59-year-olds in 29 societies. Confirmatory factor analyses tested the fit of self-ratings in each sample to the eight-syndrome model. The primary model fit index (Root Mean Square Error of Approximation) showed good model fit for all samples, while secondary indices showed acceptable to good fit. Only 5 (0.06%) of the 8,598 estimated parameters were outside the admissible parameter space. Confidence intervals indicated that sampling fluctuations could account for the deviant parameters. Results thus supported the tested model in societies differing widely in social, political, and economic systems, languages, ethnicities, religions, and geographical regions. Although other items, societies, and analytic methods might yield different results, the findings indicate that adults in very diverse societies were willing and able to rate themselves on the same standardized set of 120 problem items. Moreover, their self-ratings fit an eight-syndrome model previously derived from self-ratings by American adults. The support for the statistically derived syndrome model is consistent with previous findings for parent, teacher, and self-ratings of 1½–18-year-olds in many societies. The ASR and its parallel collateral-report instrument, the Adult Behavior Checklist (ABCL), may offer mental health professionals practical tools for the multi-informant assessment of clinical constructs of adult psychopathology that appear to be meaningful across diverse societies. PMID:29805197

  8. Communication: An efficient approach to compute state-specific nuclear gradients for a generic state-averaged multi-configuration self consistent field wavefunction.

    PubMed

    Granovsky, Alexander A

    2015-12-21

    We present a new, very efficient semi-numerical approach for the computation of state-specific nuclear gradients of a generic state-averaged multi-configuration self consistent field wavefunction. Our approach eliminates the costly coupled-perturbed multi-configuration Hartree-Fock step as well as the associated integral transformation stage. The details of the implementation within the Firefly quantum chemistry package are discussed and several sample applications are given. The new approach is routinely applicable to geometry optimization of molecular systems with 1000+ basis functions using a standalone multi-core workstation.

  9. Communication: An efficient approach to compute state-specific nuclear gradients for a generic state-averaged multi-configuration self consistent field wavefunction

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Granovsky, Alexander A., E-mail: alex.granovsky@gmail.com

    We present a new, very efficient semi-numerical approach for the computation of state-specific nuclear gradients of a generic state-averaged multi-configuration self consistent field wavefunction. Our approach eliminates the costly coupled-perturbed multi-configuration Hartree-Fock step as well as the associated integral transformation stage. The details of the implementation within the Firefly quantum chemistry package are discussed and several sample applications are given. The new approach is routinely applicable to geometry optimization of molecular systems with 1000+ basis functions using a standalone multi-core workstation.

  10. Self-consistent asset pricing models

    NASA Astrophysics Data System (ADS)

    Malevergne, Y.; Sornette, D.

    2007-08-01

    We discuss the foundations of factor or regression models in the light of the self-consistency condition that the market portfolio (and more generally the risk factors) is (are) constituted of the assets whose returns it is (they are) supposed to explain. As already reported in several articles, self-consistency implies correlations between the return disturbances. As a consequence, the alphas and betas of the factor model are unobservable. Self-consistency leads to renormalized betas with zero effective alphas, which are observable with standard OLS regressions. When the conditions derived from internal consistency are not met, the model is necessarily incomplete, which means that some sources of risk cannot be replicated (or hedged) by a portfolio of stocks traded on the market, even for infinite economies. Analytical derivations and numerical simulations show that, for arbitrary choices of the proxy which are different from the true market portfolio, a modified linear regression holds with a non-zero value αi at the origin between an asset i's return and the proxy's return. Self-consistency also introduces “orthogonality” and “normality” conditions linking the betas, alphas (as well as the residuals) and the weights of the proxy portfolio. Two diagnostics based on these orthogonality and normality conditions are implemented on a basket of 323 assets which have been components of the S&P500 in the period from January 1990 to February 2005. These two diagnostics show interesting departures from dynamical self-consistency starting about 2 years before the end of the Internet bubble. Assuming that the CAPM holds with the self-consistency condition, the OLS method automatically obeys the resulting orthogonality and normality conditions and therefore provides a simple way to self-consistently assess the parameters of the model by using proxy portfolios made only of the assets which are used in the CAPM regressions. Finally, the factor decomposition with the self-consistency condition derives a risk-factor decomposition in the multi-factor case which is identical to the principal component analysis (PCA), thus providing a direct link between model-driven and data-driven constructions of risk factors. This correspondence shows that PCA will therefore suffer from the same limitations as the CAPM and its multi-factor generalization, namely lack of out-of-sample explanatory power and predictability. In the multi-period context, the self-consistency conditions force the betas to be time-dependent with specific constraints.

  11. (Cost)-effectiveness of a multi-component intervention for adults with epilepsy: study protocol of a Dutch randomized controlled trial (ZMILE study).

    PubMed

    Leenen, Loes Am; Wijnen, Ben Fm; de Kinderen, Reina Ja; Majoie, Marian Hjm; van Heugten, Caroline M; Evers, Silvia Maa

    2014-12-24

    In patients with epilepsy, poor adherence to anti-epileptic drugs has been shown to be the most important cause of poorly controlled epilepsy. Furthermore, it has been noted that the quality of life among patients with epilepsy can be improved by counseling and treatments aimed at increasing their self-efficacy and concordance, thus stimulating self-management skills. However, there is a need for evidence on the effectiveness of such programs, especially within epilepsy care. Therefore, we have developed a multi-component intervention (MCI) which combines a self-management/education program with e-Health interventions. Accordingly, the overall objective of this study is to assess the (cost)-effectiveness and feasibility of the MCI, aiming to improve self-efficacy and concordance in patients with epilepsy. A RCT in two parallel groups will be conducted to compare the MCI with a control condition in epilepsy patients. One hundred eligible epilepsy patients will be recruited and allocated to either the intervention or control group. The intervention group will receive the MCI consisting of a self-management/education program of six meetings, including e-Health interventions, and will be followed for 12 months. The control group will receive care as usual and will be followed for 6 months, after which patients will be offered the possibility of participating in the MCI. The study will consist of three parts: 1) a clinical effectiveness study, 2) a cost-effectiveness study, and 3) process evaluation. The primary outcome will be self-efficacy. Secondary outcomes include adherence, side effects, change in seizure severity & frequency, improved quality of life, proactive coping, and societal costs. Outcome assessments will be done using questionnaires at baseline and after 3, 6, 9, and 12 months (last two applicable only for intervention group). In times of budget constraints, MCI could be a valuable addition to the current healthcare provision for epilepsy, as it is expected that higher concordance and self-efficacy will result in reduced use of healthcare resources and an increased QOL. Accordingly, this study is aimed helping patients to be their own provider of health care, shifting epilepsy management from professionals to self-care by patients equipped with appropriate skills and tools. NTR4484 .

  12. Solar flare model atmospheres

    NASA Technical Reports Server (NTRS)

    Hawley, Suzanne L.; Fisher, George H.

    1993-01-01

    Solar flare model atmospheres computed under the assumption of energetic equilibrium in the chromosphere are presented. The models use a static, one-dimensional plane parallel geometry and are designed within a physically self-consistent coronal loop. Assumed flare heating mechanisms include collisions from a flux of non-thermal electrons and x-ray heating of the chromosphere by the corona. The heating by energetic electrons accounts explicitly for variations of the ionized fraction with depth in the atmosphere. X-ray heating of the chromosphere by the corona incorporates a flare loop geometry by approximating distant portions of the loop with a series of point sources, while treating the loop leg closest to the chromospheric footpoint in the plane-parallel approximation. Coronal flare heating leads to increased heat conduction, chromospheric evaporation and subsequent changes in coronal pressure; these effects are included self-consistently in the models. Cooling in the chromosphere is computed in detail for the important optically thick HI, CaII and MgII transitions using the non-LTE prescription in the program MULTI. Hydrogen ionization rates from x-ray photo-ionization and collisional ionization by non-thermal electrons are included explicitly in the rate equations. The models are computed in the 'impulsive' and 'equilibrium' limits, and in a set of intermediate 'evolving' states. The impulsive atmospheres have the density distribution frozen in pre-flare configuration, while the equilibrium models assume the entire atmosphere is in hydrostatic and energetic equilibrium. The evolving atmospheres represent intermediate stages where hydrostatic equilibrium has been established in the chromosphere and corona, but the corona is not yet in energetic equilibrium with the flare heating source. Thus, for example, chromospheric evaporation is still in the process of occurring.

  13. The development of a revised version of multi-center molecular Ornstein-Zernike equation

    NASA Astrophysics Data System (ADS)

    Kido, Kentaro; Yokogawa, Daisuke; Sato, Hirofumi

    2012-04-01

    Ornstein-Zernike (OZ)-type theory is a powerful tool to obtain 3-dimensional solvent distribution around solute molecule. Recently, we proposed multi-center molecular OZ method, which is suitable for parallel computing of 3D solvation structure. The distribution function in this method consists of two components, namely reference and residue parts. Several types of the function were examined as the reference part to investigate the numerical robustness of the method. As the benchmark, the method is applied to water, benzene in aqueous solution and single-walled carbon nanotube in chloroform solution. The results indicate that fully-parallelization is achieved by utilizing the newly proposed reference functions.

  14. Child and Parent Perceptions of Interparental Relationship Conflict Predict Preschool Children’s Adjustment

    PubMed Central

    Clements, Mari L.; Martin, Sarah E.; Randall, David W.; Kane, Karen L.

    2014-01-01

    Interparental relationship conflict has consistently been linked to child adjustment problems, with children’s perceptions of such conflict particularly predictive of their outcomes. Despite mounting evidence that preschool children can provide reliable and valid accounts of family functioning, little research has examined preschool children’s perceptions of interparental conflict. The present study employed a multi-informant design for both reports of interparental conflict (preschool children and their mothers and fathers) and reports of child adjustment (preschool children, their mothers and fathers, and their preschool teachers). Children completed pictorial measures of interparental conflict and of self-esteem that paralleled questionnaires completed by adult respondents. For both child behavior problems and child self-esteem, preschool children’s perceptions of interparental conflict predicted their adjustment. Preschool children’s reports of interparental conflict were significantly associated with their self-esteem and with both parents’ and teachers’ reports of child behavior problems, and this association remained significant when controlling for parents’ reports of interparental conflict. Further, the interaction between parents’ reports and children’s reports of interparental conflict added unique variance to the prediction of preschool teachers’ reports of children’s behavior problems. PMID:25574451

  15. Parallel, multi-stage processing of colors, faces and shapes in macaque inferior temporal cortex

    PubMed Central

    Lafer-Sousa, Rosa; Conway, Bevil R.

    2014-01-01

    Visual-object processing culminates in inferior temporal (IT) cortex. To assess the organization of IT, we measured fMRI responses in alert monkey to achromatic images (faces, fruit, bodies, places) and colored gratings. IT contained multiple color-biased regions, which were typically ventral to face patches and, remarkably, yoked to them, spaced regularly at four locations predicted by known anatomy. Color and face selectivity increased for more anterior regions, indicative of a broad hierarchical arrangement. Responses to non-face shapes were found across IT, but were stronger outside color-biased regions and face patches, consistent with multiple parallel streams. IT also contained multiple coarse eccentricity maps: face patches overlapped central representations; color-biased regions spanned mid-peripheral representations; and place-biased regions overlapped peripheral representations. These results suggest that IT comprises parallel, multi-stage processing networks subject to one organizing principle. PMID:24141314

  16. Kinetic neoclassical transport in the H-mode pedestal

    DOE PAGES

    Battaglia, D. J.; Burrell, K. H.; Chang, C. S.; ...

    2014-07-16

    Multi-species kinetic neoclassical transport through the QH-mode pedestal and scrapeoff layer on DIII-D is calculated using XGC0, a 5D full-f particle-in-cell drift-kinetic solver with self-consistent neutral recycling and sheath potentials. We achieved quantitative agreement between the fluxdriven simulation and the experimental electron density, impurity density and orthogonal measurements of impurity temperature and flow profiles by adding random-walk particle diffusion to the guiding-center drift motion. Furthermore, we computed the radial electric field (Er) that maintains ambipolar transport across flux surfaces and to the wall self-consistently on closed and open magnetic field lines, and is in excellent agreement with experiment. The Ermore » inside the separatrix is the unique solution that balances the outward flux of thermal tail deuterium ions against the outward neoclassical electron flux and inward pinch of impurity and colder deuterium ions. Particle transport in the pedestal is primarily due to anomalous transport, while the ion heat and momentum transport is primarily due to the neoclassical transport. The full-f treatment quantifies the non-Maxwellian energy distributions that describe a number of experimental observations in low-collisionallity pedestals on DIII-D, including intrinsic co-Ip parallel flows in the pedestal, ion temperature anisotropy and large impurity temperatures in the scrape-off layer.« less

  17. Coarse-grained component concurrency in Earth system modeling: parallelizing atmospheric radiative transfer in the GFDL AM3 model using the Flexible Modeling System coupling framework

    NASA Astrophysics Data System (ADS)

    Balaji, V.; Benson, Rusty; Wyman, Bruce; Held, Isaac

    2016-10-01

    Climate models represent a large variety of processes on a variety of timescales and space scales, a canonical example of multi-physics multi-scale modeling. Current hardware trends, such as Graphical Processing Units (GPUs) and Many Integrated Core (MIC) chips, are based on, at best, marginal increases in clock speed, coupled with vast increases in concurrency, particularly at the fine grain. Multi-physics codes face particular challenges in achieving fine-grained concurrency, as different physics and dynamics components have different computational profiles, and universal solutions are hard to come by. We propose here one approach for multi-physics codes. These codes are typically structured as components interacting via software frameworks. The component structure of a typical Earth system model consists of a hierarchical and recursive tree of components, each representing a different climate process or dynamical system. This recursive structure generally encompasses a modest level of concurrency at the highest level (e.g., atmosphere and ocean on different processor sets) with serial organization underneath. We propose to extend concurrency much further by running more and more lower- and higher-level components in parallel with each other. Each component can further be parallelized on the fine grain, potentially offering a major increase in the scalability of Earth system models. We present here first results from this approach, called coarse-grained component concurrency, or CCC. Within the Geophysical Fluid Dynamics Laboratory (GFDL) Flexible Modeling System (FMS), the atmospheric radiative transfer component has been configured to run in parallel with a composite component consisting of every other atmospheric component, including the atmospheric dynamics and all other atmospheric physics components. We will explore the algorithmic challenges involved in such an approach, and present results from such simulations. Plans to achieve even greater levels of coarse-grained concurrency by extending this approach within other components, such as the ocean, will be discussed.

  18. SPIRiT: Iterative Self-consistent Parallel Imaging Reconstruction from Arbitrary k-Space

    PubMed Central

    Lustig, Michael; Pauly, John M.

    2010-01-01

    A new approach to autocalibrating, coil-by-coil parallel imaging reconstruction is presented. It is a generalized reconstruction framework based on self consistency. The reconstruction problem is formulated as an optimization that yields the most consistent solution with the calibration and acquisition data. The approach is general and can accurately reconstruct images from arbitrary k-space sampling patterns. The formulation can flexibly incorporate additional image priors such as off-resonance correction and regularization terms that appear in compressed sensing. Several iterative strategies to solve the posed reconstruction problem in both image and k-space domain are presented. These are based on a projection over convex sets (POCS) and a conjugate gradient (CG) algorithms. Phantom and in-vivo studies demonstrate efficient reconstructions from undersampled Cartesian and spiral trajectories. Reconstructions that include off-resonance correction and nonlinear ℓ1-wavelet regularization are also demonstrated. PMID:20665790

  19. Large Scale Document Inversion using a Multi-threaded Computing System

    PubMed Central

    Jung, Sungbo; Chang, Dar-Jen; Park, Juw Won

    2018-01-01

    Current microprocessor architecture is moving towards multi-core/multi-threaded systems. This trend has led to a surge of interest in using multi-threaded computing devices, such as the Graphics Processing Unit (GPU), for general purpose computing. We can utilize the GPU in computation as a massive parallel coprocessor because the GPU consists of multiple cores. The GPU is also an affordable, attractive, and user-programmable commodity. Nowadays a lot of information has been flooded into the digital domain around the world. Huge volume of data, such as digital libraries, social networking services, e-commerce product data, and reviews, etc., is produced or collected every moment with dramatic growth in size. Although the inverted index is a useful data structure that can be used for full text searches or document retrieval, a large number of documents will require a tremendous amount of time to create the index. The performance of document inversion can be improved by multi-thread or multi-core GPU. Our approach is to implement a linear-time, hash-based, single program multiple data (SPMD), document inversion algorithm on the NVIDIA GPU/CUDA programming platform utilizing the huge computational power of the GPU, to develop high performance solutions for document indexing. Our proposed parallel document inversion system shows 2-3 times faster performance than a sequential system on two different test datasets from PubMed abstract and e-commerce product reviews. CCS Concepts •Information systems➝Information retrieval • Computing methodologies➝Massively parallel and high-performance simulations. PMID:29861701

  20. Large Scale Document Inversion using a Multi-threaded Computing System.

    PubMed

    Jung, Sungbo; Chang, Dar-Jen; Park, Juw Won

    2017-06-01

    Current microprocessor architecture is moving towards multi-core/multi-threaded systems. This trend has led to a surge of interest in using multi-threaded computing devices, such as the Graphics Processing Unit (GPU), for general purpose computing. We can utilize the GPU in computation as a massive parallel coprocessor because the GPU consists of multiple cores. The GPU is also an affordable, attractive, and user-programmable commodity. Nowadays a lot of information has been flooded into the digital domain around the world. Huge volume of data, such as digital libraries, social networking services, e-commerce product data, and reviews, etc., is produced or collected every moment with dramatic growth in size. Although the inverted index is a useful data structure that can be used for full text searches or document retrieval, a large number of documents will require a tremendous amount of time to create the index. The performance of document inversion can be improved by multi-thread or multi-core GPU. Our approach is to implement a linear-time, hash-based, single program multiple data (SPMD), document inversion algorithm on the NVIDIA GPU/CUDA programming platform utilizing the huge computational power of the GPU, to develop high performance solutions for document indexing. Our proposed parallel document inversion system shows 2-3 times faster performance than a sequential system on two different test datasets from PubMed abstract and e-commerce product reviews. •Information systems➝Information retrieval • Computing methodologies➝Massively parallel and high-performance simulations.

  1. Employing Nested OpenMP for the Parallelization of Multi-Zone Computational Fluid Dynamics Applications

    NASA Technical Reports Server (NTRS)

    Ayguade, Eduard; Gonzalez, Marc; Martorell, Xavier; Jost, Gabriele

    2004-01-01

    In this paper we describe the parallelization of the multi-zone code versions of the NAS Parallel Benchmarks employing multi-level OpenMP parallelism. For our study we use the NanosCompiler, which supports nesting of OpenMP directives and provides clauses to control the grouping of threads, load balancing, and synchronization. We report the benchmark results, compare the timings with those of different hybrid parallelization paradigms and discuss OpenMP implementation issues which effect the performance of multi-level parallel applications.

  2. AMITIS: A 3D GPU-Based Hybrid-PIC Model for Space and Plasma Physics

    NASA Astrophysics Data System (ADS)

    Fatemi, Shahab; Poppe, Andrew R.; Delory, Gregory T.; Farrell, William M.

    2017-05-01

    We have developed, for the first time, an advanced modeling infrastructure in space simulations (AMITIS) with an embedded three-dimensional self-consistent grid-based hybrid model of plasma (kinetic ions and fluid electrons) that runs entirely on graphics processing units (GPUs). The model uses NVIDIA GPUs and their associated parallel computing platform, CUDA, developed for general purpose processing on GPUs. The model uses a single CPU-GPU pair, where the CPU transfers data between the system and GPU memory, executes CUDA kernels, and writes simulation outputs on the disk. All computations, including moving particles, calculating macroscopic properties of particles on a grid, and solving hybrid model equations are processed on a single GPU. We explain various computing kernels within AMITIS and compare their performance with an already existing well-tested hybrid model of plasma that runs in parallel using multi-CPU platforms. We show that AMITIS runs ∼10 times faster than the parallel CPU-based hybrid model. We also introduce an implicit solver for computation of Faraday’s Equation, resulting in an explicit-implicit scheme for the hybrid model equation. We show that the proposed scheme is stable and accurate. We examine the AMITIS energy conservation and show that the energy is conserved with an error < 0.2% after 500,000 timesteps, even when a very low number of particles per cell is used.

  3. Calculating Potential Energy Curves with Quantum Monte Carlo

    NASA Astrophysics Data System (ADS)

    Powell, Andrew D.; Dawes, Richard

    2014-06-01

    Quantum Monte Carlo (QMC) is a computational technique that can be applied to the electronic Schrödinger equation for molecules. QMC methods such as Variational Monte Carlo (VMC) and Diffusion Monte Carlo (DMC) have demonstrated the capability of capturing large fractions of the correlation energy, thus suggesting their possible use for high-accuracy quantum chemistry calculations. QMC methods scale particularly well with respect to parallelization making them an attractive consideration in anticipation of next-generation computing architectures which will involve massive parallelization with millions of cores. Due to the statistical nature of the approach, in contrast to standard quantum chemistry methods, uncertainties (error-bars) are associated with each calculated energy. This study focuses on the cost, feasibility and practical application of calculating potential energy curves for small molecules with QMC methods. Trial wave functions were constructed with the multi-configurational self-consistent field (MCSCF) method from GAMESS-US.[1] The CASINO Monte Carlo quantum chemistry package [2] was used for all of the DMC calculations. An overview of our progress in this direction will be given. References: M. W. Schmidt et al. J. Comput. Chem. 14, 1347 (1993). R. J. Needs et al. J. Phys.: Condensed Matter 22, 023201 (2010).

  4. Multi-mode sensor processing on a dynamically reconfigurable massively parallel processor array

    NASA Astrophysics Data System (ADS)

    Chen, Paul; Butts, Mike; Budlong, Brad; Wasson, Paul

    2008-04-01

    This paper introduces a novel computing architecture that can be reconfigured in real time to adapt on demand to multi-mode sensor platforms' dynamic computational and functional requirements. This 1 teraOPS reconfigurable Massively Parallel Processor Array (MPPA) has 336 32-bit processors. The programmable 32-bit communication fabric provides streamlined inter-processor connections with deterministically high performance. Software programmability, scalability, ease of use, and fast reconfiguration time (ranging from microseconds to milliseconds) are the most significant advantages over FPGAs and DSPs. This paper introduces the MPPA architecture, its programming model, and methods of reconfigurability. An MPPA platform for reconfigurable computing is based on a structural object programming model. Objects are software programs running concurrently on hundreds of 32-bit RISC processors and memories. They exchange data and control through a network of self-synchronizing channels. A common application design pattern on this platform, called a work farm, is a parallel set of worker objects, with one input and one output stream. Statically configured work farms with homogeneous and heterogeneous sets of workers have been used in video compression and decompression, network processing, and graphics applications.

  5. Development of multi-dimensional body image scale for malaysian female adolescents

    PubMed Central

    Taib, Mohd Nasir Mohd; Shariff, Zalilah Mohd; Khor, Geok Lin

    2008-01-01

    The present study was conducted to develop a Multi-dimensional Body Image Scale for Malaysian female adolescents. Data were collected among 328 female adolescents from a secondary school in Kuantan district, state of Pahang, Malaysia by using a self-administered questionnaire and anthropometric measurements. The self-administered questionnaire comprised multiple measures of body image, Eating Attitude Test (EAT-26; Garner & Garfinkel, 1979) and Rosenberg Self-esteem Inventory (Rosenberg, 1965). The 152 items from selected multiple measures of body image were examined through factor analysis and for internal consistency. Correlations between Multi-dimensional Body Image Scale and body mass index (BMI), risk of eating disorders and self-esteem were assessed for construct validity. A seven factor model of a 62-item Multi-dimensional Body Image Scale for Malaysian female adolescents with construct validity and good internal consistency was developed. The scale encompasses 1) preoccupation with thinness and dieting behavior, 2) appearance and body satisfaction, 3) body importance, 4) muscle increasing behavior, 5) extreme dieting behavior, 6) appearance importance, and 7) perception of size and shape dimensions. Besides, a multidimensional body image composite score was proposed to screen negative body image risk in female adolescents. The result found body image was correlated with BMI, risk of eating disorders and self-esteem in female adolescents. In short, the present study supports a multi-dimensional concept for body image and provides a new insight into its multi-dimensionality in Malaysian female adolescents with preliminary validity and reliability of the scale. The Multi-dimensional Body Image Scale can be used to identify female adolescents who are potentially at risk of developing body image disturbance through future intervention programs. PMID:20126371

  6. Development of multi-dimensional body image scale for malaysian female adolescents.

    PubMed

    Chin, Yit Siew; Taib, Mohd Nasir Mohd; Shariff, Zalilah Mohd; Khor, Geok Lin

    2008-01-01

    The present study was conducted to develop a Multi-dimensional Body Image Scale for Malaysian female adolescents. Data were collected among 328 female adolescents from a secondary school in Kuantan district, state of Pahang, Malaysia by using a self-administered questionnaire and anthropometric measurements. The self-administered questionnaire comprised multiple measures of body image, Eating Attitude Test (EAT-26; Garner & Garfinkel, 1979) and Rosenberg Self-esteem Inventory (Rosenberg, 1965). The 152 items from selected multiple measures of body image were examined through factor analysis and for internal consistency. Correlations between Multi-dimensional Body Image Scale and body mass index (BMI), risk of eating disorders and self-esteem were assessed for construct validity. A seven factor model of a 62-item Multi-dimensional Body Image Scale for Malaysian female adolescents with construct validity and good internal consistency was developed. The scale encompasses 1) preoccupation with thinness and dieting behavior, 2) appearance and body satisfaction, 3) body importance, 4) muscle increasing behavior, 5) extreme dieting behavior, 6) appearance importance, and 7) perception of size and shape dimensions. Besides, a multidimensional body image composite score was proposed to screen negative body image risk in female adolescents. The result found body image was correlated with BMI, risk of eating disorders and self-esteem in female adolescents. In short, the present study supports a multi-dimensional concept for body image and provides a new insight into its multi-dimensionality in Malaysian female adolescents with preliminary validity and reliability of the scale. The Multi-dimensional Body Image Scale can be used to identify female adolescents who are potentially at risk of developing body image disturbance through future intervention programs.

  7. Multi-walled boron nitride nanotubes as self-excited launchers.

    PubMed

    Li, Yifan; Zhou, Yi; Wu, Yan; Huang, Chengchi; Wang, Long; Zhou, Xuyan; Zhao, Zhenyang; Li, Hui

    2017-07-27

    A self-excited launcher consisting of multi-walled boron nitride nanotubes (BNNTs) has been investigated using molecular dynamics simulation. The results show that, after a period of high frequency oscillation, the innermost BNNT can be spontaneously ejected along its central axis at a relatively fast speed. The launching is caused by the energy transfer between the nanotubes and without absorbing energy from the external environment. Most self-excited launchers could launch their innermost nanotube, although an inappropriate structure of the nanotubes contributes to a blocked or failed launch. In addition, a launch angle corrector and a nanotube receiver associated with a self-excited launcher are also manufactured to precisely control the launch angle and distance of the BNNTs. This study provides the possibility to fabricate and design self-excited launchers using multi-walled nanotubes.

  8. Balancing exploration, uncertainty and computational demands in many objective reservoir optimization

    NASA Astrophysics Data System (ADS)

    Zatarain Salazar, Jazmin; Reed, Patrick M.; Quinn, Julianne D.; Giuliani, Matteo; Castelletti, Andrea

    2017-11-01

    Reservoir operations are central to our ability to manage river basin systems serving conflicting multi-sectoral demands under increasingly uncertain futures. These challenges motivate the need for new solution strategies capable of effectively and efficiently discovering the multi-sectoral tradeoffs that are inherent to alternative reservoir operation policies. Evolutionary many-objective direct policy search (EMODPS) is gaining importance in this context due to its capability of addressing multiple objectives and its flexibility in incorporating multiple sources of uncertainties. This simulation-optimization framework has high potential for addressing the complexities of water resources management, and it can benefit from current advances in parallel computing and meta-heuristics. This study contributes a diagnostic assessment of state-of-the-art parallel strategies for the auto-adaptive Borg Multi Objective Evolutionary Algorithm (MOEA) to support EMODPS. Our analysis focuses on the Lower Susquehanna River Basin (LSRB) system where multiple sectoral demands from hydropower production, urban water supply, recreation and environmental flows need to be balanced. Using EMODPS with different parallel configurations of the Borg MOEA, we optimize operating policies over different size ensembles of synthetic streamflows and evaporation rates. As we increase the ensemble size, we increase the statistical fidelity of our objective function evaluations at the cost of higher computational demands. This study demonstrates how to overcome the mathematical and computational barriers associated with capturing uncertainties in stochastic multiobjective reservoir control optimization, where parallel algorithmic search serves to reduce the wall-clock time in discovering high quality representations of key operational tradeoffs. Our results show that emerging self-adaptive parallelization schemes exploiting cooperative search populations are crucial. Such strategies provide a promising new set of tools for effectively balancing exploration, uncertainty, and computational demands when using EMODPS.

  9. Multi-threading: A new dimension to massively parallel scientific computation

    NASA Astrophysics Data System (ADS)

    Nielsen, Ida M. B.; Janssen, Curtis L.

    2000-06-01

    Multi-threading is becoming widely available for Unix-like operating systems, and the application of multi-threading opens new ways for performing parallel computations with greater efficiency. We here briefly discuss the principles of multi-threading and illustrate the application of multi-threading for a massively parallel direct four-index transformation of electron repulsion integrals. Finally, other potential applications of multi-threading in scientific computing are outlined.

  10. A parallel algorithm for multi-level logic synthesis using the transduction method. M.S. Thesis

    NASA Technical Reports Server (NTRS)

    Lim, Chieng-Fai

    1991-01-01

    The Transduction Method has been shown to be a powerful tool in the optimization of multilevel networks. Many tools such as the SYLON synthesis system (X90), (CM89), (LM90) have been developed based on this method. A parallel implementation is presented of SYLON-XTRANS (XM89) on an eight processor Encore Multimax shared memory multiprocessor. It minimizes multilevel networks consisting of simple gates through parallel pruning, gate substitution, gate merging, generalized gate substitution, and gate input reduction. This implementation, called Parallel TRANSduction (PTRANS), also uses partitioning to break large circuits up and performs inter- and intra-partition dynamic load balancing. With this, good speedups and high processor efficiencies are achievable without sacrificing the resulting circuit quality.

  11. Parallel Execution of Functional Mock-up Units in Buildings Modeling

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ozmen, Ozgur; Nutaro, James J.; New, Joshua Ryan

    2016-06-30

    A Functional Mock-up Interface (FMI) defines a standardized interface to be used in computer simulations to develop complex cyber-physical systems. FMI implementation by a software modeling tool enables the creation of a simulation model that can be interconnected, or the creation of a software library called a Functional Mock-up Unit (FMU). This report describes an FMU wrapper implementation that imports FMUs into a C++ environment and uses an Euler solver that executes FMUs in parallel using Open Multi-Processing (OpenMP). The purpose of this report is to elucidate the runtime performance of the solver when a multi-component system is imported asmore » a single FMU (for the whole system) or as multiple FMUs (for different groups of components as sub-systems). This performance comparison is conducted using two test cases: (1) a simple, multi-tank problem; and (2) a more realistic use case based on the Modelica Buildings Library. In both test cases, the performance gains are promising when each FMU consists of a large number of states and state events that are wrapped in a single FMU. Load balancing is demonstrated to be a critical factor in speeding up parallel execution of multiple FMUs.« less

  12. Parallel multi-join query optimization algorithm for distributed sensor network in the internet of things

    NASA Astrophysics Data System (ADS)

    Zheng, Yan

    2015-03-01

    Internet of things (IoT), focusing on providing users with information exchange and intelligent control, attracts a lot of attention of researchers from all over the world since the beginning of this century. IoT is consisted of large scale of sensor nodes and data processing units, and the most important features of IoT can be illustrated as energy confinement, efficient communication and high redundancy. With the sensor nodes increment, the communication efficiency and the available communication band width become bottle necks. Many research work is based on the instance which the number of joins is less. However, it is not proper to the increasing multi-join query in whole internet of things. To improve the communication efficiency between parallel units in the distributed sensor network, this paper proposed parallel query optimization algorithm based on distribution attributes cost graph. The storage information relations and the network communication cost are considered in this algorithm, and an optimized information changing rule is established. The experimental result shows that the algorithm has good performance, and it would effectively use the resource of each node in the distributed sensor network. Therefore, executive efficiency of multi-join query between different nodes could be improved.

  13. Parallel Lattice Basis Reduction Using a Multi-threaded Schnorr-Euchner LLL Algorithm

    NASA Astrophysics Data System (ADS)

    Backes, Werner; Wetzel, Susanne

    In this paper, we introduce a new parallel variant of the LLL lattice basis reduction algorithm. Our new, multi-threaded algorithm is the first to provide an efficient, parallel implementation of the Schorr-Euchner algorithm for today’s multi-processor, multi-core computer architectures. Experiments with sparse and dense lattice bases show a speed-up factor of about 1.8 for the 2-thread and about factor 3.2 for the 4-thread version of our new parallel lattice basis reduction algorithm in comparison to the traditional non-parallel algorithm.

  14. A Multi-Dimensional Instrument for Evaluating Taiwanese High School Students' Science Learning Self-Efficacy in Relation to Their Approaches to Learning Science

    ERIC Educational Resources Information Center

    Lin, Tzung-Jin; Tsai, Chin-Chung

    2013-01-01

    In the past, students' science learning self-efficacy (SLSE) was usually measured by questionnaires that consisted of only a single scale, which might be insufficient to fully understand their SLSE. In this study, a multi-dimensional instrument, the SLSE instrument, was developed and validated to assess students' SLSE based on the previous…

  15. What is adaptive about adaptive decision making? A parallel constraint satisfaction account.

    PubMed

    Glöckner, Andreas; Hilbig, Benjamin E; Jekel, Marc

    2014-12-01

    There is broad consensus that human cognition is adaptive. However, the vital question of how exactly this adaptivity is achieved has remained largely open. Herein, we contrast two frameworks which account for adaptive decision making, namely broad and general single-mechanism accounts vs. multi-strategy accounts. We propose and fully specify a single-mechanism model for decision making based on parallel constraint satisfaction processes (PCS-DM) and contrast it theoretically and empirically against a multi-strategy account. To achieve sufficiently sensitive tests, we rely on a multiple-measure methodology including choice, reaction time, and confidence data as well as eye-tracking. Results show that manipulating the environmental structure produces clear adaptive shifts in choice patterns - as both frameworks would predict. However, results on the process level (reaction time, confidence), in information acquisition (eye-tracking), and from cross-predicting choice consistently corroborate single-mechanisms accounts in general, and the proposed parallel constraint satisfaction model for decision making in particular. Copyright © 2014 Elsevier B.V. All rights reserved.

  16. Self-Tuning Fully-Connected PID Neural Network System for Distributed Temperature Sensing and Control of Instrument with Multi-Modules.

    PubMed

    Zhang, Zhen; Ma, Cheng; Zhu, Rong

    2016-10-14

    High integration of multi-functional instruments raises a critical issue in temperature control that is challenging due to its spatial-temporal complexity. This paper presents a multi-input multi-output (MIMO) self-tuning temperature sensing and control system for efficiently modulating the temperature environment within a multi-module instrument. The smart system ensures that the internal temperature of the instrument converges to a target without the need of a system model, thus making the control robust. The system consists of a fully-connected proportional-integral-derivative (PID) neural network (FCPIDNN) and an on-line self-tuning module. The experimental results show that the presented system can effectively control the internal temperature under various mission scenarios, in particular, it is able to self-reconfigure upon actuator failure. The system provides a new scheme for a complex and time-variant MIMO control system which can be widely applied for the distributed measurement and control of the environment in instruments, integration electronics, and house constructions.

  17. A Measure of Child and Adolescent Self-Concept and Psychological Adjustment.

    ERIC Educational Resources Information Center

    Bronstein, Phyllis; And Others

    This paper presents a new multi-scale instrument--the Bronstein-Cruz Child/Adolescent Self-Concept and Adjustment Scale, which measures self-concept and psychological adjustment for children and adolescents, aged 10-18 years. It consists of five subscales: Self-Evaluation, Social and Peer Relations, Family Relations, Sense of Mastery, and…

  18. On the Debye-Hückel effect of electric screening

    NASA Astrophysics Data System (ADS)

    Campos, L. M. B. C.; Lau, F. J. P.

    2014-07-01

    The paper considers non-linear self-consistent electric potential equation (Sec. I), due to a cloud made of a single species of electric charges, satisfying a Boltzmann distribution law (Sec. II). Exact solutions are obtained in a simple logarithmic form, in three cases: (Sec. III) spherical radial symmetry; (Sec. IV) plane parallel symmetry; (Sec. V) a special case of azimuthal-cylindrical symmetry. All these solutions, and their transformations (Sec. VI), involve the Debye-Hückel radius; the latter was originally defined from a solution of the linearized self-consistent potential equation. Using an exact solution of the self-consistent potential equation, the distance at which the potential vanishes differs from the Debye-Hückel radius by a factor of √2 . The preceding (Secs. II-VI) simple logarithmic exact solutions of the self-consistent potential equations involve no arbitrary constants, and thus are special or singular integrals not the general integral. The general solution of the self-consistent potential equation is obtained in the plane parallel case (Sec. VII), and it involves two arbitrary constants that can be reduced to one via a translation (Sec. VIII). The plots of dimensionless potential (Figure 1), electric field (Figure 2), charge density (Figure 3), and total charge between ζ and infinity (Figure 4), versus distance normalized to Debye-Hückel radius ζ ≡ z/a, show that (Sec. IX) there is a continuum of solutions, ranging from a charge distribution concentrated inside the Debye-Hückel radius to one spread-out beyond it. The latter case leads to the limiting case of logarithmic potential, and stronger electric field; the former case, of very concentrated charge distribution, leads to a fratricide effect and weaker electric field.

  19. On the Debye–Hückel effect of electric screening

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Campos, L. M. B. C.; Lau, F. J. P.

    2014-07-15

    The paper considers non-linear self-consistent electric potential equation (Sec. I), due to a cloud made of a single species of electric charges, satisfying a Boltzmann distribution law (Sec. II). Exact solutions are obtained in a simple logarithmic form, in three cases: (Sec. III) spherical radial symmetry; (Sec. IV) plane parallel symmetry; (Sec. V) a special case of azimuthal-cylindrical symmetry. All these solutions, and their transformations (Sec. VI), involve the Debye-Hückel radius; the latter was originally defined from a solution of the linearized self-consistent potential equation. Using an exact solution of the self-consistent potential equation, the distance at which the potentialmore » vanishes differs from the Debye-Hückel radius by a factor of √(2). The preceding (Secs. II–VI) simple logarithmic exact solutions of the self-consistent potential equations involve no arbitrary constants, and thus are special or singular integrals not the general integral. The general solution of the self-consistent potential equation is obtained in the plane parallel case (Sec. VII), and it involves two arbitrary constants that can be reduced to one via a translation (Sec. VIII). The plots of dimensionless potential (Figure 1), electric field (Figure 2), charge density (Figure 3), and total charge between ζ and infinity (Figure 4), versus distance normalized to Debye-Hückel radius ζ ≡ z/a, show that (Sec. IX) there is a continuum of solutions, ranging from a charge distribution concentrated inside the Debye-Hückel radius to one spread-out beyond it. The latter case leads to the limiting case of logarithmic potential, and stronger electric field; the former case, of very concentrated charge distribution, leads to a fratricide effect and weaker electric field.« less

  20. Keldysh formalism for multiple parallel worlds

    NASA Astrophysics Data System (ADS)

    Ansari, M.; Nazarov, Y. V.

    2016-03-01

    We present a compact and self-contained review of the recently developed Keldysh formalism for multiple parallel worlds. The formalism has been applied to consistent quantum evaluation of the flows of informational quantities, in particular, to the evaluation of Renyi and Shannon entropy flows. We start with the formulation of the standard and extended Keldysh techniques in a single world in a form convenient for our presentation. We explain the use of Keldysh contours encompassing multiple parallel worlds. In the end, we briefly summarize the concrete results obtained with the method.

  1. Keldysh formalism for multiple parallel worlds

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ansari, M.; Nazarov, Y. V., E-mail: y.v.nazarov@tudelft.nl

    We present a compact and self-contained review of the recently developed Keldysh formalism for multiple parallel worlds. The formalism has been applied to consistent quantum evaluation of the flows of informational quantities, in particular, to the evaluation of Renyi and Shannon entropy flows. We start with the formulation of the standard and extended Keldysh techniques in a single world in a form convenient for our presentation. We explain the use of Keldysh contours encompassing multiple parallel worlds. In the end, we briefly summarize the concrete results obtained with the method.

  2. Consistent cortical reconstruction and multi-atlas brain segmentation.

    PubMed

    Huo, Yuankai; Plassard, Andrew J; Carass, Aaron; Resnick, Susan M; Pham, Dzung L; Prince, Jerry L; Landman, Bennett A

    2016-09-01

    Whole brain segmentation and cortical surface reconstruction are two essential techniques for investigating the human brain. Spatial inconsistences, which can hinder further integrated analyses of brain structure, can result due to these two tasks typically being conducted independently of each other. FreeSurfer obtains self-consistent whole brain segmentations and cortical surfaces. It starts with subcortical segmentation, then carries out cortical surface reconstruction, and ends with cortical segmentation and labeling. However, this "segmentation to surface to parcellation" strategy has shown limitations in various cohorts such as older populations with large ventricles. In this work, we propose a novel "multi-atlas segmentation to surface" method called Multi-atlas CRUISE (MaCRUISE), which achieves self-consistent whole brain segmentations and cortical surfaces by combining multi-atlas segmentation with the cortical reconstruction method CRUISE. A modification called MaCRUISE(+) is designed to perform well when white matter lesions are present. Comparing to the benchmarks CRUISE and FreeSurfer, the surface accuracy of MaCRUISE and MaCRUISE(+) is validated using two independent datasets with expertly placed cortical landmarks. A third independent dataset with expertly delineated volumetric labels is employed to compare segmentation performance. Finally, 200MR volumetric images from an older adult sample are used to assess the robustness of MaCRUISE and FreeSurfer. The advantages of MaCRUISE are: (1) MaCRUISE constructs self-consistent voxelwise segmentations and cortical surfaces, while MaCRUISE(+) is robust to white matter pathology. (2) MaCRUISE achieves more accurate whole brain segmentations than independently conducting the multi-atlas segmentation. (3) MaCRUISE is comparable in accuracy to FreeSurfer (when FreeSurfer does not exhibit global failures) while achieving greater robustness across an older adult population. MaCRUISE has been made freely available in open source. Copyright © 2016 Elsevier Inc. All rights reserved.

  3. Node Resource Manager: A Distributed Computing Software Framework Used for Solving Geophysical Problems

    NASA Astrophysics Data System (ADS)

    Lawry, B. J.; Encarnacao, A.; Hipp, J. R.; Chang, M.; Young, C. J.

    2011-12-01

    With the rapid growth of multi-core computing hardware, it is now possible for scientific researchers to run complex, computationally intensive software on affordable, in-house commodity hardware. Multi-core CPUs (Central Processing Unit) and GPUs (Graphics Processing Unit) are now commonplace in desktops and servers. Developers today have access to extremely powerful hardware that enables the execution of software that could previously only be run on expensive, massively-parallel systems. It is no longer cost-prohibitive for an institution to build a parallel computing cluster consisting of commodity multi-core servers. In recent years, our research team has developed a distributed, multi-core computing system and used it to construct global 3D earth models using seismic tomography. Traditionally, computational limitations forced certain assumptions and shortcuts in the calculation of tomographic models; however, with the recent rapid growth in computational hardware including faster CPU's, increased RAM, and the development of multi-core computers, we are now able to perform seismic tomography, 3D ray tracing and seismic event location using distributed parallel algorithms running on commodity hardware, thereby eliminating the need for many of these shortcuts. We describe Node Resource Manager (NRM), a system we developed that leverages the capabilities of a parallel computing cluster. NRM is a software-based parallel computing management framework that works in tandem with the Java Parallel Processing Framework (JPPF, http://www.jppf.org/), a third party library that provides a flexible and innovative way to take advantage of modern multi-core hardware. NRM enables multiple applications to use and share a common set of networked computers, regardless of their hardware platform or operating system. Using NRM, algorithms can be parallelized to run on multiple processing cores of a distributed computing cluster of servers and desktops, which results in a dramatic speedup in execution time. NRM is sufficiently generic to support applications in any domain, as long as the application is parallelizable (i.e., can be subdivided into multiple individual processing tasks). At present, NRM has been effective in decreasing the overall runtime of several algorithms: 1) the generation of a global 3D model of the compressional velocity distribution in the Earth using tomographic inversion, 2) the calculation of the model resolution matrix, model covariance matrix, and travel time uncertainty for the aforementioned velocity model, and 3) the correlation of waveforms with archival data on a massive scale for seismic event detection. Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-AC04-94AL85000.

  4. Optoelectronic Materials Center

    DTIC Science & Technology

    1991-06-11

    surface - emitting GaAs/AIGaAs vertical - cavity laser (TJ- VCSEL ) incorporating wavelength-resonant...multi-quantum well, vertical cavity surface - emitted laser . This structure consists entirely of undoped epilayers, thus simplifying the problems of... cavity surface - emitting lasers ( VCSELs ) for doubling and for parallel optical data processing. Progress - GaAIAs/GaAs and InGaAs/GaAs RPG- VCSEL

  5. A Framework for Parallel Unstructured Grid Generation for Complex Aerodynamic Simulations

    NASA Technical Reports Server (NTRS)

    Zagaris, George; Pirzadeh, Shahyar Z.; Chrisochoides, Nikos

    2009-01-01

    A framework for parallel unstructured grid generation targeting both shared memory multi-processors and distributed memory architectures is presented. The two fundamental building-blocks of the framework consist of: (1) the Advancing-Partition (AP) method used for domain decomposition and (2) the Advancing Front (AF) method used for mesh generation. Starting from the surface mesh of the computational domain, the AP method is applied recursively to generate a set of sub-domains. Next, the sub-domains are meshed in parallel using the AF method. The recursive nature of domain decomposition naturally maps to a divide-and-conquer algorithm which exhibits inherent parallelism. For the parallel implementation, the Master/Worker pattern is employed to dynamically balance the varying workloads of each task on the set of available CPUs. Performance results by this approach are presented and discussed in detail as well as future work and improvements.

  6. Implementation of Parallel Dynamic Simulation on Shared-Memory vs. Distributed-Memory Environments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jin, Shuangshuang; Chen, Yousu; Wu, Di

    2015-12-09

    Power system dynamic simulation computes the system response to a sequence of large disturbance, such as sudden changes in generation or load, or a network short circuit followed by protective branch switching operation. It consists of a large set of differential and algebraic equations, which is computational intensive and challenging to solve using single-processor based dynamic simulation solution. High-performance computing (HPC) based parallel computing is a very promising technology to speed up the computation and facilitate the simulation process. This paper presents two different parallel implementations of power grid dynamic simulation using Open Multi-processing (OpenMP) on shared-memory platform, and Messagemore » Passing Interface (MPI) on distributed-memory clusters, respectively. The difference of the parallel simulation algorithms and architectures of the two HPC technologies are illustrated, and their performances for running parallel dynamic simulation are compared and demonstrated.« less

  7. Sequential or parallel decomposed processing of two-digit numbers? Evidence from eye-tracking.

    PubMed

    Moeller, Korbinian; Fischer, Martin H; Nuerk, Hans-Christoph; Willmes, Klaus

    2009-02-01

    While reaction time data have shown that decomposed processing of two-digit numbers occurs, there is little evidence about how decomposed processing functions. Poltrock and Schwartz (1984) argued that multi-digit numbers are compared in a sequential digit-by-digit fashion starting at the leftmost digit pair. In contrast, Nuerk and Willmes (2005) favoured parallel processing of the digits constituting a number. These models (i.e., sequential decomposition, parallel decomposition) make different predictions regarding the fixation pattern in a two-digit number magnitude comparison task and can therefore be differentiated by eye fixation data. We tested these models by evaluating participants' eye fixation behaviour while selecting the larger of two numbers. The stimulus set consisted of within-decade comparisons (e.g., 53_57) and between-decade comparisons (e.g., 42_57). The between-decade comparisons were further divided into compatible and incompatible trials (cf. Nuerk, Weger, & Willmes, 2001) and trials with different decade and unit distances. The observed fixation pattern implies that the comparison of two-digit numbers is not executed by sequentially comparing decade and unit digits as proposed by Poltrock and Schwartz (1984) but rather in a decomposed but parallel fashion. Moreover, the present fixation data provide first evidence that digit processing in multi-digit numbers is not a pure bottom-up effect, but is also influenced by top-down factors. Finally, implications for multi-digit number processing beyond the range of two-digit numbers are discussed.

  8. A multi-block adaptive solving technique based on lattice Boltzmann method

    NASA Astrophysics Data System (ADS)

    Zhang, Yang; Xie, Jiahua; Li, Xiaoyue; Ma, Zhenghai; Zou, Jianfeng; Zheng, Yao

    2018-05-01

    In this paper, a CFD parallel adaptive algorithm is self-developed by combining the multi-block Lattice Boltzmann Method (LBM) with Adaptive Mesh Refinement (AMR). The mesh refinement criterion of this algorithm is based on the density, velocity and vortices of the flow field. The refined grid boundary is obtained by extending outward half a ghost cell from the coarse grid boundary, which makes the adaptive mesh more compact and the boundary treatment more convenient. Two numerical examples of the backward step flow separation and the unsteady flow around circular cylinder demonstrate the vortex structure of the cold flow field accurately and specifically.

  9. Multi-Sensor Data Fusion Identification for Shearer Cutting Conditions Based on Parallel Quasi-Newton Neural Networks and the Dempster-Shafer Theory.

    PubMed

    Si, Lei; Wang, Zhongbin; Liu, Xinhua; Tan, Chao; Xu, Jing; Zheng, Kehong

    2015-11-13

    In order to efficiently and accurately identify the cutting condition of a shearer, this paper proposed an intelligent multi-sensor data fusion identification method using the parallel quasi-Newton neural network (PQN-NN) and the Dempster-Shafer (DS) theory. The vibration acceleration signals and current signal of six cutting conditions were collected from a self-designed experimental system and some special state features were extracted from the intrinsic mode functions (IMFs) based on the ensemble empirical mode decomposition (EEMD). In the experiment, three classifiers were trained and tested by the selected features of the measured data, and the DS theory was used to combine the identification results of three single classifiers. Furthermore, some comparisons with other methods were carried out. The experimental results indicate that the proposed method performs with higher detection accuracy and credibility than the competing algorithms. Finally, an industrial application example in the fully mechanized coal mining face was demonstrated to specify the effect of the proposed system.

  10. Assessment of adult psychopathology: meta-analyses and implications of cross-informant correlations.

    PubMed

    Achenbach, Thomas M; Krukowski, Rebecca A; Dumenci, Levent; Ivanova, Masha Y

    2005-05-01

    Assessment of adult psychopathology relies heavily on self-reports. To determine how well self-reports agree with reports by "informants" who know the person being assessed, the authors examined 51,000 articles published over 10 years in 52 peer-reviewed journals for correlations between self-reports and "informants" reports. Qualifying correlations were found in 108 (0.2%) of the articles. When self-reports and informant reports were obtained with parallel instruments, mean cross-informant correlations were .681 for substance use, .428 for internalizing, and .438 for externalizing problems. When based on different instruments, the mean cross-informant correlation was .304. The moderate sizes of the correlations argue for systematically obtaining multi-informant data. National survey findings were used to illustrate practical ways to obtain and use such data.

  11. Laser-self-mixing interferometry for mechatronics applications.

    PubMed

    Ottonelli, Simona; Dabbicco, Maurizio; De Lucia, Francesco; di Vietro, Michela; Scamarcio, Gaetano

    2009-01-01

    We report on the development of an all-interferometric optomechatronic sensor for the detection of multi-degrees-of-freedom displacements of a remote target. The prototype system exploits the self-mixing technique and consists only of a laser head, equipped with six laser sources, and a suitably designed reflective target. The feasibility of the system was validated experimentally for both single or multi-degrees-of-freedom measurements, thus demonstrating a simple and inexpensive alternative to costly and bulky existing systems.

  12. A parallel multi-domain solution methodology applied to nonlinear thermal transport problems in nuclear fuel pins

    DOE PAGES

    Philip, Bobby; Berrill, Mark A.; Allu, Srikanth; ...

    2015-01-26

    We describe an efficient and nonlinearly consistent parallel solution methodology for solving coupled nonlinear thermal transport problems that occur in nuclear reactor applications over hundreds of individual 3D physical subdomains. Efficiency is obtained by leveraging knowledge of the physical domains, the physics on individual domains, and the couplings between them for preconditioning within a Jacobian Free Newton Krylov method. Details of the computational infrastructure that enabled this work, namely the open source Advanced Multi-Physics (AMP) package developed by the authors are described. The details of verification and validation experiments, and parallel performance analysis in weak and strong scaling studies demonstratingmore » the achieved efficiency of the algorithm are presented. Moreover, numerical experiments demonstrate that the preconditioner developed is independent of the number of fuel subdomains in a fuel rod, which is particularly important when simulating different types of fuel rods. Finally, we demonstrate the power of the coupling methodology by considering problems with couplings between surface and volume physics and coupling of nonlinear thermal transport in fuel rods to an external radiation transport code.« less

  13. Self-Reported quality of life in adults with attention-deficit/hyperactivity disorder and executive function impairment treated with lisdexamfetamine dimesylate: a randomized, double-blind, multicenter, placebo-controlled, parallel-group study.

    PubMed

    Adler, Lenard A; Dirks, Bryan; Deas, Patrick; Raychaudhuri, Aparna; Dauphin, Matthew; Saylor, Keith; Weisler, Richard

    2013-10-09

    This study examined the effects of lisdexamfetamine dimesylate (LDX) on quality of life (QOL) in adults with attention-deficit/hyperactivity disorder (ADHD) and clinically significant executive function deficits (EFD). This report highlights QOL findings from a 10-week randomized placebo-controlled trial of LDX (30-70 mg/d) in adults (18-55 years) with ADHD and EFD (Behavior Rating Inventory of EF-Adult, Global Executive Composite [BRIEF-A GEC] ≥65). The primary efficacy measure was the self-reported BRIEF-A; a key secondary measure was self-reported QOL on the Adult ADHD Impact Module (AIM-A). The clinician-completed ADHD Rating Scale version IV (ADHD-RS-IV) with adult prompts and Clinical Global Impressions-Severity (CGI-S) were also employed. The Adult ADHD QoL (AAQoL) was added while the study was in progress. A post hoc analysis examined the subgroup having evaluable results from both AIM-A and AAQoL. Of 161 randomized (placebo, 81; LDX, 80), 159 were included in the safety population. LDX improved AIM-A multi-item domain scores versus placebo; LS mean difference for Performance and Daily Functioning was 21.6 (ES, 0.93, P<.0001); Impact of Symptoms: Daily Interference was 14.9 (ES, 0.62, P<.0001); Impact of Symptoms: Bother/Concern was 13.5 (ES, 0.57, P=.0003); Relationships/Communication was 7.8 (ES, 0.31, P=.0302); Living With ADHD was 9.1 (ES, 0.79, P<.0001); and General Well-Being was 10.8 (ES, 0.70, P<.0001). AAQoL LS mean difference for total score was 21.0; for subscale: Life Productivity was 21.0; Psychological Health was 12.1; Life Outlook was 12.5; and Relationships was 7.3. In a post hoc analysis of participants with both AIM-A and AAQoL scores, AIM-A multi-item subgroup analysis scores numerically improved with LDX, with smaller difference for Impact of Symptoms: Daily Interference. The safety profile of LDX was consistent with amphetamine use in previous studies. Overall, adults with ADHD/EFD exhibited self-reported improvement on QOL, using the AIM-A and AAQoL scales in line with medium/large ES; these improvements were paralleled by improvements in EF and ADHD symptoms. The safety profile of LDX was similar to previous studies. ClinicalTrials.gov, NCT01101022.

  14. Toward a Model Framework of Generalized Parallel Componential Processing of Multi-Symbol Numbers

    ERIC Educational Resources Information Center

    Huber, Stefan; Cornelsen, Sonja; Moeller, Korbinian; Nuerk, Hans-Christoph

    2015-01-01

    In this article, we propose and evaluate a new model framework of parallel componential multi-symbol number processing, generalizing the idea of parallel componential processing of multi-digit numbers to the case of negative numbers by considering the polarity signs similar to single digits. In a first step, we evaluated this account by defining…

  15. Research on control technology of hardware parallelism for marine controlled source electromagnetic transmitter

    NASA Astrophysics Data System (ADS)

    Wang, Meng; Deng, Ming; Luo, Xianhu; Zhao, Qingxian; Chen, Kai; Jing, Jianen

    2018-02-01

    The marine controlled source electromagnetic (CSEM) method has been recognized as an effective exploration method of shallow hydrocarbons around the world. We developed our own underwater marine CSEM transmitter that consisted of many functional modules with various response times. We previously adopted a centralized software-control technology to design the transmitter circuit topological structure. That structure probably generated a control disorder or malfunction. These undesirable conditions could lead to repeated recovery and deployment of the transmitter, which not only consumed time but also affected data continuity and establishment of stable and continuous CSEM field. We developed an instrument design concept named ‘control technology of hardware parallelism’. In this design, a noteworthy innovation of our new technology is to solve the above-mentioned problems at the physical and fundamental levels. We used several self-contained control-units to simultaneously accomplish the predetermined functions of the transmitter. The new solution relies on two technologies: multi-core embedded technology and multi-channel parallel optical-fiber data transmission technology. The first technology depends on many independent microcontrollers. Every microcontroller is only used to achieve a customized function. The second one relies on several multiple optical-fiber transmission channels realized by a complex programmable logic device and two optical-fiber conversion devices, which are used to establish a communication link between the shipboard monitoring and control-unit and underwater transmitter. We have conducted some marine experiments to verify the reliability and stability of the new method. In particular, the new technology used in the transmitter system could help us obtain more useful measured data in a limited time, improve real-time efficiency, and support the establishment of a stable CSEM field.

  16. Biphasic DC measurement approach for enhanced measurement stability and multi-channel sampling of self-sensing multi-functional structural materials doped with carbon-based additives

    NASA Astrophysics Data System (ADS)

    Downey, Austin; D'Alessandro, Antonella; Ubertini, Filippo; Laflamme, Simon; Geiger, Randall

    2017-06-01

    Investigation of multi-functional carbon-based self-sensing structural materials for structural health monitoring applications is a topic of growing interest. These materials are self-sensing in the sense that they can provide measurable electrical outputs corresponding to physical changes such as strain or induced damage. Nevertheless, the development of an appropriate measurement technique for such materials is yet to be achieved, as many results in the literature suggest that these materials exhibit a drift in their output when measured with direct current (DC) methods. In most of the cases, the electrical output is a resistance and the reported drift is an increase in resistance from the time the measurement starts due to material polarization. Alternating current methods seem more appropriate at eliminating the time drift. However, published results show they are not immune to drift. Moreover, the use of multiple impedance measurement devices (LCR meters) does not allow for the simultaneous multi-channel sampling of multi-sectioned self-sensing materials due to signal crosstalk. The capability to simultaneously monitor multiple sections of self-sensing structural materials is needed to deploy these multi-functional materials for structural health monitoring. Here, a biphasic DC measurement approach with a periodic measure/discharge cycle in the form of a square wave sensing current is used to provide consistent, stable resistance measurements for self-sensing structural materials. DC measurements are made during the measurement region of the square wave while material depolarization is obtained during the discharge region of the periodic signal. The proposed technique is experimentally shown to remove the signal drift in a carbon-based self-sensing cementitious material while providing simultaneous multi-channel measurements of a multi-sectioned self-sensing material. The application of the proposed electrical measurement technique appears promising for real-time utilization of self-sensing materials in structural health monitoring.

  17. High-performance parallel analysis of coupled problems for aircraft propulsion

    NASA Technical Reports Server (NTRS)

    Felippa, C. A.; Farhat, C.; Lanteri, S.; Gumaste, U.; Ronaghi, M.

    1994-01-01

    Applications are described of high-performance parallel, computation for the analysis of complete jet engines, considering its multi-discipline coupled problem. The coupled problem involves interaction of structures with gas dynamics, heat conduction and heat transfer in aircraft engines. The methodology issues addressed include: consistent discrete formulation of coupled problems with emphasis on coupling phenomena; effect of partitioning strategies, augmentation and temporal solution procedures; sensitivity of response to problem parameters; and methods for interfacing multiscale discretizations in different single fields. The computer implementation issues addressed include: parallel treatment of coupled systems; domain decomposition and mesh partitioning strategies; data representation in object-oriented form and mapping to hardware driven representation, and tradeoff studies between partitioning schemes and fully coupled treatment.

  18. Maximal clique enumeration with data-parallel primitives

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lessley, Brenton; Perciano, Talita; Mathai, Manish

    The enumeration of all maximal cliques in an undirected graph is a fundamental problem arising in several research areas. We consider maximal clique enumeration on shared-memory, multi-core architectures and introduce an approach consisting entirely of data-parallel operations, in an effort to achieve efficient and portable performance across different architectures. We study the performance of the algorithm via experiments varying over benchmark graphs and architectures. Overall, we observe that our algorithm achieves up to a 33-time speedup and 9-time speedup over state-of-the-art distributed and serial algorithms, respectively, for graphs with higher ratios of maximal cliques to total cliques. Further, we attainmore » additional speedups on a GPU architecture, demonstrating the portable performance of our data-parallel design.« less

  19. GENESIS: new self-consistent models of exoplanetary spectra

    NASA Astrophysics Data System (ADS)

    Gandhi, Siddharth; Madhusudhan, Nikku

    2017-12-01

    We are entering the era of high-precision and high-resolution spectroscopy of exoplanets. Such observations herald the need for robust self-consistent spectral models of exoplanetary atmospheres to investigate intricate atmospheric processes and to make observable predictions. Spectral models of plane-parallel exoplanetary atmospheres exist, mostly adapted from other astrophysical applications, with different levels of sophistication and accuracy. There is a growing need for a new generation of models custom-built for exoplanets and incorporating state-of-the-art numerical methods and opacities. The present work is a step in this direction. Here we introduce GENESIS, a plane-parallel, self-consistent, line-by-line exoplanetary atmospheric modelling code that includes (a) formal solution of radiative transfer using the Feautrier method, (b) radiative-convective equilibrium with temperature correction based on the Rybicki linearization scheme, (c) latest absorption cross-sections, and (d) internal flux and external irradiation, under the assumptions of hydrostatic equilibrium, local thermodynamic equilibrium and thermochemical equilibrium. We demonstrate the code here with cloud-free models of giant exoplanetary atmospheres over a range of equilibrium temperatures, metallicities, C/O ratios and spanning non-irradiated and irradiated planets, with and without thermal inversions. We provide the community with theoretical emergent spectra and pressure-temperature profiles over this range, along with those for several known hot Jupiters. The code can generate self-consistent spectra at high resolution and has the potential to be integrated into general circulation and non-equilibrium chemistry models as it is optimized for efficiency and convergence. GENESIS paves the way for high-fidelity remote sensing of exoplanetary atmospheres at high resolution with current and upcoming observations.

  20. Algorithm implementation on the Navier-Stokes computer

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Krist, S.E.; Zang, T.A.

    1987-03-01

    The Navier-Stokes Computer is a multi-purpose parallel-processing supercomputer which is currently under development at Princeton University. It consists of multiple local memory parallel processors, called Nodes, which are interconnected in a hypercube network. Details of the procedures involved in implementing an algorithm on the Navier-Stokes computer are presented. The particular finite difference algorithm considered in this analysis was developed for simulation of laminar-turbulent transition in wall bounded shear flows. Projected timing results for implementing this algorithm indicate that operation rates in excess of 42 GFLOPS are feasible on a 128 Node machine.

  1. Algorithm implementation on the Navier-Stokes computer

    NASA Technical Reports Server (NTRS)

    Krist, Steven E.; Zang, Thomas A.

    1987-01-01

    The Navier-Stokes Computer is a multi-purpose parallel-processing supercomputer which is currently under development at Princeton University. It consists of multiple local memory parallel processors, called Nodes, which are interconnected in a hypercube network. Details of the procedures involved in implementing an algorithm on the Navier-Stokes computer are presented. The particular finite difference algorithm considered in this analysis was developed for simulation of laminar-turbulent transition in wall bounded shear flows. Projected timing results for implementing this algorithm indicate that operation rates in excess of 42 GFLOPS are feasible on a 128 Node machine.

  2. Laser-Self-Mixing Interferometry for Mechatronics Applications

    PubMed Central

    Ottonelli, Simona; Dabbicco, Maurizio; De Lucia, Francesco; di Vietro, Michela; Scamarcio, Gaetano

    2009-01-01

    We report on the development of an all-interferometric optomechatronic sensor for the detection of multi-degrees-of-freedom displacements of a remote target. The prototype system exploits the self-mixing technique and consists only of a laser head, equipped with six laser sources, and a suitably designed reflective target. The feasibility of the system was validated experimentally for both single or multi-degrees-of-freedom measurements, thus demonstrating a simple and inexpensive alternative to costly and bulky existing systems. PMID:22412324

  3. Design and Implementation of a Multi-Strategy, Collegewide Program of Evaluation and Planning: The Mercy College Self-Study Project.

    ERIC Educational Resources Information Center

    Kraetzer, Mary C.; And Others

    The rationale, strategies, and methods of The Mercy College Self-Study Project are considered, and evaluation instruments are provided. This program of institutional evaluation and planning was initiated in 1980 and consists of: standardized surveys, a 10-year longitudinal (panel) study, and academic department self-studies. Questionnaires…

  4. Self-Efficacy Scale for Weight Loss among Multi-Ethnic Women of Lower Income: A Psychometric Evaluation

    ERIC Educational Resources Information Center

    Latimer, Lara; Walker, Lorraine O.; Kim, Sunghun; Pasch, Keryn E.; Sterling, Bobbie Sue

    2011-01-01

    Objective: This study examined test-retest reliability, internal consistency, and construct and predictive validity of the Physical Activity and Nutrition Self-Efficacy (PANSE) scale, an 11-item instrument to assess weight-loss self-efficacy among postpartum women of lower income. Methods: Seventy-one women completed the PANSE scale and…

  5. Research on Multi - Person Parallel Modeling Method Based on Integrated Model Persistent Storage

    NASA Astrophysics Data System (ADS)

    Qu, MingCheng; Wu, XiangHu; Tao, YongChao; Liu, Ying

    2018-03-01

    This paper mainly studies the multi-person parallel modeling method based on the integrated model persistence storage. The integrated model refers to a set of MDDT modeling graphics system, which can carry out multi-angle, multi-level and multi-stage description of aerospace general embedded software. Persistent storage refers to converting the data model in memory into a storage model and converting the storage model into a data model in memory, where the data model refers to the object model and the storage model is a binary stream. And multi-person parallel modeling refers to the need for multi-person collaboration, the role of separation, and even real-time remote synchronization modeling.

  6. Performance Characteristics of the Multi-Zone NAS Parallel Benchmarks

    NASA Technical Reports Server (NTRS)

    Jin, Haoqiang; VanderWijngaart, Rob F.

    2003-01-01

    We describe a new suite of computational benchmarks that models applications featuring multiple levels of parallelism. Such parallelism is often available in realistic flow computations on systems of grids, but had not previously been captured in bench-marks. The new suite, named NPB Multi-Zone, is extended from the NAS Parallel Benchmarks suite, and involves solving the application benchmarks LU, BT and SP on collections of loosely coupled discretization meshes. The solutions on the meshes are updated independently, but after each time step they exchange boundary value information. This strategy provides relatively easily exploitable coarse-grain parallelism between meshes. Three reference implementations are available: one serial, one hybrid using the Message Passing Interface (MPI) and OpenMP, and another hybrid using a shared memory multi-level programming model (SMP+OpenMP). We examine the effectiveness of hybrid parallelization paradigms in these implementations on three different parallel computers. We also use an empirical formula to investigate the performance characteristics of the multi-zone benchmarks.

  7. The DANTE Boltzmann transport solver: An unstructured mesh, 3-D, spherical harmonics algorithm compatible with parallel computer architectures

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    McGhee, J.M.; Roberts, R.M.; Morel, J.E.

    1997-06-01

    A spherical harmonics research code (DANTE) has been developed which is compatible with parallel computer architectures. DANTE provides 3-D, multi-material, deterministic, transport capabilities using an arbitrary finite element mesh. The linearized Boltzmann transport equation is solved in a second order self-adjoint form utilizing a Galerkin finite element spatial differencing scheme. The core solver utilizes a preconditioned conjugate gradient algorithm. Other distinguishing features of the code include options for discrete-ordinates and simplified spherical harmonics angular differencing, an exact Marshak boundary treatment for arbitrarily oriented boundary faces, in-line matrix construction techniques to minimize memory consumption, and an effective diffusion based preconditioner formore » scattering dominated problems. Algorithm efficiency is demonstrated for a massively parallel SIMD architecture (CM-5), and compatibility with MPP multiprocessor platforms or workstation clusters is anticipated.« less

  8. Risk Perception of HIV/AIDS and Low Self-Control Trait: Explaining Preventative Behaviors Among Iranian University Students

    PubMed Central

    Esmaeilzadeh, Safooreh; Allahverdipour, Hamid; Fathi, Behrouz; Shirzadi, Shayesteh

    2016-01-01

    Background: In spite of developed countries there are progressive trend about HIV/AIDS and its’ aspects of transmission in the low socio-economic societies. The aim of this was to explain the youth's behavior in adopting HIV/AIDS related preventive behaviors in a sample of Iranian university students by emphasizing on fear appeals approaches alongside examining the role of self-control trait for explaining adoption on danger or fear control processes based on Extended Parallel Process Model (EPPM). Methods: A sample of 156 randomly selected university students in Jolfa, Iran was recruited in a predictive cross-sectional study by application of a researcher-designed questionnaire through self-report data collection manner. Sexual high risk behaviors, the EPPM variables, self-control trait, and general self-efficacy were measured as theoretical framework. Results: Findings indicated that 31.3% of participants were in the fear control process versus 68.7% in danger control about HIV/AIDS and also the presence of multi-sex partners and amphetamine consumption amongst the participants. Low self-control trait and low perceived susceptibility significantly were related to having a history of multi-sex partners while high level of self-efficacy significantly increased the probability of condom use. Conclusion: Findings of the study were indicative of the protective role of high level of self-control, perceived susceptibility and self-efficacy factors on youth's high-risk behaviors and their preventative skills as well. PMID:26573026

  9. Multi-arm spectrometer for parallel frequency analysis of radio-wave signals oriented to astronomical observations

    NASA Astrophysics Data System (ADS)

    Shcherbakov, Alexandre S.; Chavez Dagostino, Miguel; Arellanes, Adan Omar; Tepichin Rodriguez, Eduardo

    2017-08-01

    We describe a potential prototype of modern spectrometer based on acousto-optical technique with three parallel optical arms for analysis of radio-wave signals specific to astronomical observations. Each optical arm exhibits original performances to provide parallel multi-band observations with different scales simultaneously. Similar multi-band instrument is able to realize measurements within various scenarios from planetary atmospheres to attractive objects in the distant Universe. The arrangement under development has two novelties. First, each optical arm represents an individual spectrum analyzer with its individual performances. Such an approach is conditioned by exploiting various materials for acousto-optical cells operating within various regimes, frequency ranges, and light wavelengths from independent light sources. Individually produced beam shapers give both the needed incident light polarization and the required apodization for light beam to increase the dynamic range of the system as a whole. After parallel acousto-optical processing, a few data flows from these optical arms are united by the joint CCD matrix on the stage of the combined extremely high-bit rate electronic data processing that provides the system performances as well. The other novelty consists in the usage of various materials for designing wide-aperture acousto-optical cells exhibiting the best performances within each of optical arms. Here, one can mention specifically selected cuts of tellurium dioxide, bastron, and lithium niobate, which overlap selected areas within the frequency range from 40 MHz to 2.0 GHz. Thus one yields the united versatile instrument for comprehensive studies of astronomical objects simultaneously with precise synchronization in various frequency ranges.

  10. Expressing Parallelism with ROOT

    NASA Astrophysics Data System (ADS)

    Piparo, D.; Tejedor, E.; Guiraud, E.; Ganis, G.; Mato, P.; Moneta, L.; Valls Pla, X.; Canal, P.

    2017-10-01

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  11. Expressing Parallelism with ROOT

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Piparo, D.; Tejedor, E.; Guiraud, E.

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module inmore » Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.« less

  12. Suppressing correlations in massively parallel simulations of lattice models

    NASA Astrophysics Data System (ADS)

    Kelling, Jeffrey; Ódor, Géza; Gemming, Sibylle

    2017-11-01

    For lattice Monte Carlo simulations parallelization is crucial to make studies of large systems and long simulation time feasible, while sequential simulations remain the gold-standard for correlation-free dynamics. Here, various domain decomposition schemes are compared, concluding with one which delivers virtually correlation-free simulations on GPUs. Extensive simulations of the octahedron model for 2 + 1 dimensional Kardar-Parisi-Zhang surface growth, which is very sensitive to correlation in the site-selection dynamics, were performed to show self-consistency of the parallel runs and agreement with the sequential algorithm. We present a GPU implementation providing a speedup of about 30 × over a parallel CPU implementation on a single socket and at least 180 × with respect to the sequential reference.

  13. Advancing the Multi-Informant Assessment of Sluggish Cognitive Tempo: Child Self-Report in Relation to Parent and Teacher Ratings of SCT and Impairment.

    PubMed

    Sáez, Belén; Servera, Mateu; Burns, G Leonard; Becker, Stephen P

    2018-04-27

    Despite increasing interest in sluggish cognitive tempo (SCT) in children and advancements in its measurement, little research has examined child self-reported SCT. Child self-report of SCT is important for the multi-informant assessment of SCT. The current study used a large, school-based sample of children and a multi-informant design to examine child self-reported SCT using the Child Concentration Inventory - Version 2 (CCI-2) which was recently revised based on meta-analytic findings and parallels the item content of validated parent and teacher rating scales. The study involved 2142 unique children (ages 8-13 years, 50.51% males). Children (n = 1980) completed measures of SCT, loneliness, and preference for solitude. Mothers (n = 1648), fathers (n = 1358), and teachers (n = 1773) completed measures of SCT, attention-deficit/hyperactivity disorder-IN (ADHD-IN), academic impairment, social impairment, and conflicted shyness. Children's self-reported SCT demonstrated good reliability with the 15 SCT symptoms showing moderate to strong loadings on the SCT factor. The child self-report SCT factor also showed moderate convergent validity with mother, father, and teacher ratings of children's SCT. In addition, higher child-reported SCT predicted greater mother, father, and teacher ratings of children's academic impairment even after controlling for mother, father, and teacher ratings of children's SCT and ADHD-IN. Higher child-rated SCT also predicted greater mother ratings of children's social impairment after controlling for mother ratings of children's SCT and ADHD-IN. The present study provides initial empirical support for the reliability and validity of child-reported SCT as part of the multi-informant assessment of SCT. A key direction for future research includes evaluating the unique contributions of different informants and their utility within specific contexts to guide evidence-based recommendations for assessing SCT.

  14. Adaptive multi-GPU Exchange Monte Carlo for the 3D Random Field Ising Model

    NASA Astrophysics Data System (ADS)

    Navarro, Cristóbal A.; Huang, Wei; Deng, Youjin

    2016-08-01

    This work presents an adaptive multi-GPU Exchange Monte Carlo approach for the simulation of the 3D Random Field Ising Model (RFIM). The design is based on a two-level parallelization. The first level, spin-level parallelism, maps the parallel computation as optimal 3D thread-blocks that simulate blocks of spins in shared memory with minimal halo surface, assuming a constant block volume. The second level, replica-level parallelism, uses multi-GPU computation to handle the simulation of an ensemble of replicas. CUDA's concurrent kernel execution feature is used in order to fill the occupancy of each GPU with many replicas, providing a performance boost that is more notorious at the smallest values of L. In addition to the two-level parallel design, the work proposes an adaptive multi-GPU approach that dynamically builds a proper temperature set free of exchange bottlenecks. The strategy is based on mid-point insertions at the temperature gaps where the exchange rate is most compromised. The extra work generated by the insertions is balanced across the GPUs independently of where the mid-point insertions were performed. Performance results show that spin-level performance is approximately two orders of magnitude faster than a single-core CPU version and one order of magnitude faster than a parallel multi-core CPU version running on 16-cores. Multi-GPU performance is highly convenient under a weak scaling setting, reaching up to 99 % efficiency as long as the number of GPUs and L increase together. The combination of the adaptive approach with the parallel multi-GPU design has extended our possibilities of simulation to sizes of L = 32 , 64 for a workstation with two GPUs. Sizes beyond L = 64 can eventually be studied using larger multi-GPU systems.

  15. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bassi, Gabriele; Blednykh, Alexei; Smalyuk, Victor

    A novel algorithm for self-consistent simulations of long-range wakefield effects has been developed and applied to the study of both longitudinal and transverse coupled-bunch instabilities at NSLS-II. The algorithm is implemented in the new parallel tracking code space (self-consistent parallel algorithm for collective effects) discussed in the paper. The code is applicable for accurate beam dynamics simulations in cases where both bunch-to-bunch and intrabunch motions need to be taken into account, such as chromatic head-tail effects on the coupled-bunch instability of a beam with a nonuniform filling pattern, or multibunch and single-bunch effects of a passive higher-harmonic cavity. The numericalmore » simulations have been compared with analytical studies. For a beam with an arbitrary filling pattern, intensity-dependent complex frequency shifts have been derived starting from a system of coupled Vlasov equations. The analytical formulas and numerical simulations confirm that the analysis is reduced to the formulation of an eigenvalue problem based on the known formulas of the complex frequency shifts for the uniform filling pattern case.« less

  16. Multi-threaded parallel simulation of non-local non-linear problems in ultrashort laser pulse propagation in the presence of plasma

    NASA Astrophysics Data System (ADS)

    Baregheh, Mandana; Mezentsev, Vladimir; Schmitz, Holger

    2011-06-01

    We describe a parallel multi-threaded approach for high performance modelling of wide class of phenomena in ultrafast nonlinear optics. Specific implementation has been performed using the highly parallel capabilities of a programmable graphics processor.

  17. Multi-Sensor Data Fusion Identification for Shearer Cutting Conditions Based on Parallel Quasi-Newton Neural Networks and the Dempster-Shafer Theory

    PubMed Central

    Si, Lei; Wang, Zhongbin; Liu, Xinhua; Tan, Chao; Xu, Jing; Zheng, Kehong

    2015-01-01

    In order to efficiently and accurately identify the cutting condition of a shearer, this paper proposed an intelligent multi-sensor data fusion identification method using the parallel quasi-Newton neural network (PQN-NN) and the Dempster-Shafer (DS) theory. The vibration acceleration signals and current signal of six cutting conditions were collected from a self-designed experimental system and some special state features were extracted from the intrinsic mode functions (IMFs) based on the ensemble empirical mode decomposition (EEMD). In the experiment, three classifiers were trained and tested by the selected features of the measured data, and the DS theory was used to combine the identification results of three single classifiers. Furthermore, some comparisons with other methods were carried out. The experimental results indicate that the proposed method performs with higher detection accuracy and credibility than the competing algorithms. Finally, an industrial application example in the fully mechanized coal mining face was demonstrated to specify the effect of the proposed system. PMID:26580620

  18. Simulating electron wave dynamics in graphene superlattices exploiting parallel processing advantages

    NASA Astrophysics Data System (ADS)

    Rodrigues, Manuel J.; Fernandes, David E.; Silveirinha, Mário G.; Falcão, Gabriel

    2018-01-01

    This work introduces a parallel computing framework to characterize the propagation of electron waves in graphene-based nanostructures. The electron wave dynamics is modeled using both "microscopic" and effective medium formalisms and the numerical solution of the two-dimensional massless Dirac equation is determined using a Finite-Difference Time-Domain scheme. The propagation of electron waves in graphene superlattices with localized scattering centers is studied, and the role of the symmetry of the microscopic potential in the electron velocity is discussed. The computational methodologies target the parallel capabilities of heterogeneous multi-core CPU and multi-GPU environments and are built with the OpenCL parallel programming framework which provides a portable, vendor agnostic and high throughput-performance solution. The proposed heterogeneous multi-GPU implementation achieves speedup ratios up to 75x when compared to multi-thread and multi-core CPU execution, reducing simulation times from several hours to a couple of minutes.

  19. Parallel group independent component analysis for massive fMRI data sets.

    PubMed

    Chen, Shaojie; Huang, Lei; Qiu, Huitong; Nebel, Mary Beth; Mostofsky, Stewart H; Pekar, James J; Lindquist, Martin A; Eloyan, Ani; Caffo, Brian S

    2017-01-01

    Independent component analysis (ICA) is widely used in the field of functional neuroimaging to decompose data into spatio-temporal patterns of co-activation. In particular, ICA has found wide usage in the analysis of resting state fMRI (rs-fMRI) data. Recently, a number of large-scale data sets have become publicly available that consist of rs-fMRI scans from thousands of subjects. As a result, efficient ICA algorithms that scale well to the increased number of subjects are required. To address this problem, we propose a two-stage likelihood-based algorithm for performing group ICA, which we denote Parallel Group Independent Component Analysis (PGICA). By utilizing the sequential nature of the algorithm and parallel computing techniques, we are able to efficiently analyze data sets from large numbers of subjects. We illustrate the efficacy of PGICA, which has been implemented in R and is freely available through the Comprehensive R Archive Network, through simulation studies and application to rs-fMRI data from two large multi-subject data sets, consisting of 301 and 779 subjects respectively.

  20. [Series: Medical Applications of the PHITS Code (2): Acceleration by Parallel Computing].

    PubMed

    Furuta, Takuya; Sato, Tatsuhiko

    2015-01-01

    Time-consuming Monte Carlo dose calculation becomes feasible owing to the development of computer technology. However, the recent development is due to emergence of the multi-core high performance computers. Therefore, parallel computing becomes a key to achieve good performance of software programs. A Monte Carlo simulation code PHITS contains two parallel computing functions, the distributed-memory parallelization using protocols of message passing interface (MPI) and the shared-memory parallelization using open multi-processing (OpenMP) directives. Users can choose the two functions according to their needs. This paper gives the explanation of the two functions with their advantages and disadvantages. Some test applications are also provided to show their performance using a typical multi-core high performance workstation.

  1. Plasma and energetic particle structure of a collisionless quasi-parallel shock

    NASA Technical Reports Server (NTRS)

    Kennel, C. F.; Scarf, F. L.; Coroniti, F. V.; Russell, C. T.; Smith, E. J.; Wenzel, K. P.; Reinhard, R.; Sanderson, T. R.; Feldman, W. C.; Parks, G. K.

    1983-01-01

    The quasi-parallel interplanetary shock of November 11-12, 1978 from both the collisionless shock and energetic particle points of view were studied using measurements of the interplanetary magnetic and electric fields, solar wind electrons, plasma and MHD waves, and intermediate and high energy ions obtained on ISEE-1, -2, and -3. The interplanetary environment through which the shock was propagating when it encountered the three spacecraft was characterized; the observations of this shock are documented and current theories of quasi-parallel shock structure and particle acceleration are tested. These observations tend to confirm present self consistent theories of first order Fermi acceleration by shocks and of collisionless shock dissipation involving firehouse instability.

  2. Nonlinear Dynamics of a Multistage Gear Transmission System with Multi-Clearance

    NASA Astrophysics Data System (ADS)

    Xiang, Ling; Zhang, Yue; Gao, Nan; Hu, Aijun; Xing, Jingtang

    The nonlinear torsional model of a multistage gear transmission system which consists of a planetary gear and two parallel gear stages is established with time-varying meshing stiffness, comprehensive gear error and multi-clearance. The nonlinear dynamic responses are analyzed by applying the reference of backlash bifurcation parameters. The motions of the system on the change of backlash are identified through global bifurcation diagram, largest Lyapunov exponent (LLE), FFT spectra, Poincaré maps, the phase diagrams and time series. The numerical results demonstrate that the system exhibits rich features of nonlinear dynamics such as the periodic motion, nonperiodic states and chaotic states. It is found that the sun-planet backlash has more complex effect on the system than the ring-planet backlash. The motions of the system with backlash of parallel gear are diverse including some different multi-periodic motions. Furthermore, the state of the system can change from chaos into quasi-periodic behavior, which means that the dynamic behavior of the system is composed of more stable components with the increase of the backlash. Correspondingly, the parameters of the system should be designed properly and controlled timely for better operation and enhancing the life of the system.

  3. Path lumping: An efficient algorithm to identify metastable path channels for conformational dynamics of multi-body systems

    NASA Astrophysics Data System (ADS)

    Meng, Luming; Sheong, Fu Kit; Zeng, Xiangze; Zhu, Lizhe; Huang, Xuhui

    2017-07-01

    Constructing Markov state models from large-scale molecular dynamics simulation trajectories is a promising approach to dissect the kinetic mechanisms of complex chemical and biological processes. Combined with transition path theory, Markov state models can be applied to identify all pathways connecting any conformational states of interest. However, the identified pathways can be too complex to comprehend, especially for multi-body processes where numerous parallel pathways with comparable flux probability often coexist. Here, we have developed a path lumping method to group these parallel pathways into metastable path channels for analysis. We define the similarity between two pathways as the intercrossing flux between them and then apply the spectral clustering algorithm to lump these pathways into groups. We demonstrate the power of our method by applying it to two systems: a 2D-potential consisting of four metastable energy channels and the hydrophobic collapse process of two hydrophobic molecules. In both cases, our algorithm successfully reveals the metastable path channels. We expect this path lumping algorithm to be a promising tool for revealing unprecedented insights into the kinetic mechanisms of complex multi-body processes.

  4. Accelerating global optimization of aerodynamic shapes using a new surrogate-assisted parallel genetic algorithm

    NASA Astrophysics Data System (ADS)

    Ebrahimi, Mehdi; Jahangirian, Alireza

    2017-12-01

    An efficient strategy is presented for global shape optimization of wing sections with a parallel genetic algorithm. Several computational techniques are applied to increase the convergence rate and the efficiency of the method. A variable fidelity computational evaluation method is applied in which the expensive Navier-Stokes flow solver is complemented by an inexpensive multi-layer perceptron neural network for the objective function evaluations. A population dispersion method that consists of two phases, of exploration and refinement, is developed to improve the convergence rate and the robustness of the genetic algorithm. Owing to the nature of the optimization problem, a parallel framework based on the master/slave approach is used. The outcomes indicate that the method is able to find the global optimum with significantly lower computational time in comparison to the conventional genetic algorithm.

  5. Dynamical Generation of Quasi-Stationary Alfvenic Double Layers and Charge Holes and Unified Theory of Quasi-Static and Alfvenic Auroral Arc Formation

    NASA Astrophysics Data System (ADS)

    Song, Y.; Lysak, R. L.

    2015-12-01

    Parallel E-fields play a crucial role for the acceleration of charged particles, creating discrete aurorae. However, once the parallel electric fields are produced, they will disappear right away, unless the electric fields can be continuously generated and sustained for a fairly long time. Thus, the crucial question in auroral physics is how to generate such a powerful and self-sustained parallel electric fields which can effectively accelerate charge particles to high energy during a fairly long time. We propose that nonlinear interaction of incident and reflected Alfven wave packets in inhomogeneous auroral acceleration region can produce quasi-stationary non-propagating electromagnetic plasma structures, such as Alfvenic double layers (DLs) and Charge Holes. Such Alfvenic quasi-static structures often constitute powerful high energy particle accelerators. The Alfvenic DL consists of localized self-sustained powerful electrostatic electric fields nested in a low density cavity and surrounded by enhanced magnetic and mechanical stresses. The enhanced magnetic and velocity fields carrying the free energy serve as a local dynamo, which continuously create the electrostatic parallel electric field for a fairly long time. The generated parallel electric fields will deepen the seed low density cavity, which then further quickly boosts the stronger parallel electric fields creating both Alfvenic and quasi-static discrete aurorae. The parallel electrostatic electric field can also cause ion outflow, perpendicular ion acceleration and heating, and may excite Auroral Kilometric Radiation.

  6. Options for Parallelizing a Planning and Scheduling Algorithm

    NASA Technical Reports Server (NTRS)

    Clement, Bradley J.; Estlin, Tara A.; Bornstein, Benjamin D.

    2011-01-01

    Space missions have a growing interest in putting multi-core processors onboard spacecraft. For many missions processing power significantly slows operations. We investigate how continual planning and scheduling algorithms can exploit multi-core processing and outline different potential design decisions for a parallelized planning architecture. This organization of choices and challenges helps us with an initial design for parallelizing the CASPER planning system for a mesh multi-core processor. This work extends that presented at another workshop with some preliminary results.

  7. The self-assembly of particles with isotropic interactions: Using DNA coated colloids to create designer nanomaterials

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Thompson, R. B.; Dion, S.; Konigslow, K. von

    Self-consistent field theory equations are presented that are suitable for use as a coarse-grained model for DNA coated colloids, polymer-grafted nanoparticles and other systems with approximately isotropic interactions. The equations are generalized for arbitrary numbers of chemically distinct colloids. The advantages and limitations of such a coarse-grained approach for DNA coated colloids are discussed, as are similarities with block copolymer self-assembly. In particular, preliminary results for three species self-assembly are presented that parallel results from a two dimensional ABC triblock copolymer phase. The possibility of incorporating crystallization, dynamics, inverse statistical mechanics and multiscale modelling techniques are discussed.

  8. Verification of Electromagnetic Physics Models for Parallel Computing Architectures in the GeantV Project

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Amadio, G.; et al.

    An intensive R&D and programming effort is required to accomplish new challenges posed by future experimental high-energy particle physics (HEP) programs. The GeantV project aims to narrow the gap between the performance of the existing HEP detector simulation software and the ideal performance achievable, exploiting latest advances in computing technology. The project has developed a particle detector simulation prototype capable of transporting in parallel particles in complex geometries exploiting instruction level microparallelism (SIMD and SIMT), task-level parallelism (multithreading) and high-level parallelism (MPI), leveraging both the multi-core and the many-core opportunities. We present preliminary verification results concerning the electromagnetic (EM) physicsmore » models developed for parallel computing architectures within the GeantV project. In order to exploit the potential of vectorization and accelerators and to make the physics model effectively parallelizable, advanced sampling techniques have been implemented and tested. In this paper we introduce a set of automated statistical tests in order to verify the vectorized models by checking their consistency with the corresponding Geant4 models and to validate them against experimental data.« less

  9. GENESIS: a hybrid-parallel and multi-scale molecular dynamics simulator with enhanced sampling algorithms for biomolecular and cellular simulations.

    PubMed

    Jung, Jaewoon; Mori, Takaharu; Kobayashi, Chigusa; Matsunaga, Yasuhiro; Yoda, Takao; Feig, Michael; Sugita, Yuji

    2015-07-01

    GENESIS (Generalized-Ensemble Simulation System) is a new software package for molecular dynamics (MD) simulations of macromolecules. It has two MD simulators, called ATDYN and SPDYN. ATDYN is parallelized based on an atomic decomposition algorithm for the simulations of all-atom force-field models as well as coarse-grained Go-like models. SPDYN is highly parallelized based on a domain decomposition scheme, allowing large-scale MD simulations on supercomputers. Hybrid schemes combining OpenMP and MPI are used in both simulators to target modern multicore computer architectures. Key advantages of GENESIS are (1) the highly parallel performance of SPDYN for very large biological systems consisting of more than one million atoms and (2) the availability of various REMD algorithms (T-REMD, REUS, multi-dimensional REMD for both all-atom and Go-like models under the NVT, NPT, NPAT, and NPγT ensembles). The former is achieved by a combination of the midpoint cell method and the efficient three-dimensional Fast Fourier Transform algorithm, where the domain decomposition space is shared in real-space and reciprocal-space calculations. Other features in SPDYN, such as avoiding concurrent memory access, reducing communication times, and usage of parallel input/output files, also contribute to the performance. We show the REMD simulation results of a mixed (POPC/DMPC) lipid bilayer as a real application using GENESIS. GENESIS is released as free software under the GPLv2 licence and can be easily modified for the development of new algorithms and molecular models. WIREs Comput Mol Sci 2015, 5:310-323. doi: 10.1002/wcms.1220.

  10. Multi-GPU parallel algorithm design and analysis for improved inversion of probability tomography with gravity gradiometry data

    NASA Astrophysics Data System (ADS)

    Hou, Zhenlong; Huang, Danian

    2017-09-01

    In this paper, we make a study on the inversion of probability tomography (IPT) with gravity gradiometry data at first. The space resolution of the results is improved by multi-tensor joint inversion, depth weighting matrix and the other methods. Aiming at solving the problems brought by the big data in the exploration, we present the parallel algorithm and the performance analysis combining Compute Unified Device Architecture (CUDA) with Open Multi-Processing (OpenMP) based on Graphics Processing Unit (GPU) accelerating. In the test of the synthetic model and real data from Vinton Dome, we get the improved results. It is also proved that the improved inversion algorithm is effective and feasible. The performance of parallel algorithm we designed is better than the other ones with CUDA. The maximum speedup could be more than 200. In the performance analysis, multi-GPU speedup and multi-GPU efficiency are applied to analyze the scalability of the multi-GPU programs. The designed parallel algorithm is demonstrated to be able to process larger scale of data and the new analysis method is practical.

  11. Parallel Monte Carlo transport modeling in the context of a time-dependent, three-dimensional multi-physics code

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Procassini, R.J.

    1997-12-31

    The fine-scale, multi-space resolution that is envisioned for accurate simulations of complex weapons systems in three spatial dimensions implies flop-rate and memory-storage requirements that will only be obtained in the near future through the use of parallel computational techniques. Since the Monte Carlo transport models in these simulations usually stress both of these computational resources, they are prime candidates for parallelization. The MONACO Monte Carlo transport package, which is currently under development at LLNL, will utilize two types of parallelism within the context of a multi-physics design code: decomposition of the spatial domain across processors (spatial parallelism) and distribution ofmore » particles in a given spatial subdomain across additional processors (particle parallelism). This implementation of the package will utilize explicit data communication between domains (message passing). Such a parallel implementation of a Monte Carlo transport model will result in non-deterministic communication patterns. The communication of particles between subdomains during a Monte Carlo time step may require a significant level of effort to achieve a high parallel efficiency.« less

  12. Transitioning NWChem to the Next Generation of Manycore Machines

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bylaska, Eric J.; Apra, E; Kowalski, Karol

    The NorthWest chemistry (NWChem) modeling software is a popular molecular chemistry simulation software that was designed from the start to work on massively parallel processing supercomputers [1-3]. It contains an umbrella of modules that today includes self-consistent eld (SCF), second order Møller-Plesset perturbation theory (MP2), coupled cluster (CC), multiconguration self-consistent eld (MCSCF), selected conguration interaction (CI), tensor contraction engine (TCE) many body methods, density functional theory (DFT), time-dependent density functional theory (TDDFT), real-time time-dependent density functional theory, pseudopotential plane-wave density functional theory (PSPW), band structure (BAND), ab initio molecular dynamics (AIMD), Car-Parrinello molecular dynamics (MD), classical MD, hybrid quantum mechanicsmore » molecular mechanics (QM/MM), hybrid ab initio molecular dynamics molecular mechanics (AIMD/MM), gauge independent atomic orbital nuclear magnetic resonance (GIAO NMR), conductor like screening solvation model (COSMO), conductor-like screening solvation model based on density (COSMO-SMD), and reference interaction site model (RISM) solvation models, free energy simulations, reaction path optimization, parallel in time, among other capabilities [4]. Moreover, new capabilities continue to be added with each new release.« less

  13. A Sparse Self-Consistent Field Algorithm and Its Parallel Implementation: Application to Density-Functional-Based Tight Binding.

    PubMed

    Scemama, Anthony; Renon, Nicolas; Rapacioli, Mathias

    2014-06-10

    We present an algorithm and its parallel implementation for solving a self-consistent problem as encountered in Hartree-Fock or density functional theory. The algorithm takes advantage of the sparsity of matrices through the use of local molecular orbitals. The implementation allows one to exploit efficiently modern symmetric multiprocessing (SMP) computer architectures. As a first application, the algorithm is used within the density-functional-based tight binding method, for which most of the computational time is spent in the linear algebra routines (diagonalization of the Fock/Kohn-Sham matrix). We show that with this algorithm (i) single point calculations on very large systems (millions of atoms) can be performed on large SMP machines, (ii) calculations involving intermediate size systems (1000-100 000 atoms) are also strongly accelerated and can run efficiently on standard servers, and (iii) the error on the total energy due to the use of a cutoff in the molecular orbital coefficients can be controlled such that it remains smaller than the SCF convergence criterion.

  14. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wright, J. C.; Bonoli, P. T.; Schmidt, A. E.

    Lower hybrid (LH) waves ({omega}{sub ci}<<{omega}<<{omega}{sub ce}, where {omega}{sub i,e}{identical_to}Z{sub i,e}eB/m{sub i,e}c) have the attractive property of damping strongly via electron Landau resonance on relatively fast tail electrons and consequently are well-suited to driving current. Established modeling techniques use Wentzel-Kramers-Brillouin (WKB) expansions with self-consistent non-Maxwellian distributions. Higher order WKB expansions have shown some effects on the parallel wave number evolution and consequently on the damping due to diffraction [G. Pereverzev, Nucl. Fusion 32, 1091 (1991)]. A massively parallel version of the TORIC full wave electromagnetic field solver valid in the LH range of frequencies has been developed [J. C. Wrightmore » et al., Comm. Comp. Phys. 4, 545 (2008)] and coupled to an electron Fokker-Planck solver CQL3D[R. W. Harvey and M. G. McCoy, in Proceedings of the IAEA Technical Committee Meeting, Montreal, 1992 (IAEA Institute of Physics Publishing, Vienna, 1993), USDOC/NTIS Document No. DE93002962, pp. 489-526] in order to self-consistently evolve nonthermal electron distributions characteristic of LH current drive (LHCD) experiments in devices such as Alcator C-Mod and ITER (B{sub 0}{approx_equal}5 T, n{sub e0}{approx_equal}1x10{sup 20} m{sup -3}). These simulations represent the first ever self-consistent simulations of LHCD utilizing both a full wave and Fokker-Planck calculation in toroidal geometry.« less

  15. Parallel processing architecture for H.264 deblocking filter on multi-core platforms

    NASA Astrophysics Data System (ADS)

    Prasad, Durga P.; Sonachalam, Sekar; Kunchamwar, Mangesh K.; Gunupudi, Nageswara Rao

    2012-03-01

    Massively parallel computing (multi-core) chips offer outstanding new solutions that satisfy the increasing demand for high resolution and high quality video compression technologies such as H.264. Such solutions not only provide exceptional quality but also efficiency, low power, and low latency, previously unattainable in software based designs. While custom hardware and Application Specific Integrated Circuit (ASIC) technologies may achieve lowlatency, low power, and real-time performance in some consumer devices, many applications require a flexible and scalable software-defined solution. The deblocking filter in H.264 encoder/decoder poses difficult implementation challenges because of heavy data dependencies and the conditional nature of the computations. Deblocking filter implementations tend to be fixed and difficult to reconfigure for different needs. The ability to scale up for higher quality requirements such as 10-bit pixel depth or a 4:2:2 chroma format often reduces the throughput of a parallel architecture designed for lower feature set. A scalable architecture for deblocking filtering, created with a massively parallel processor based solution, means that the same encoder or decoder will be deployed in a variety of applications, at different video resolutions, for different power requirements, and at higher bit-depths and better color sub sampling patterns like YUV, 4:2:2, or 4:4:4 formats. Low power, software-defined encoders/decoders may be implemented using a massively parallel processor array, like that found in HyperX technology, with 100 or more cores and distributed memory. The large number of processor elements allows the silicon device to operate more efficiently than conventional DSP or CPU technology. This software programing model for massively parallel processors offers a flexible implementation and a power efficiency close to that of ASIC solutions. This work describes a scalable parallel architecture for an H.264 compliant deblocking filter for multi core platforms such as HyperX technology. Parallel techniques such as parallel processing of independent macroblocks, sub blocks, and pixel row level are examined in this work. The deblocking architecture consists of a basic cell called deblocking filter unit (DFU) and dependent data buffer manager (DFM). The DFU can be used in several instances, catering to different performance needs the DFM serves the data required for the different number of DFUs, and also manages all the neighboring data required for future data processing of DFUs. This approach achieves the scalability, flexibility, and performance excellence required in deblocking filters.

  16. Self-consistent DFT +U method for real-space time-dependent density functional theory calculations

    NASA Astrophysics Data System (ADS)

    Tancogne-Dejean, Nicolas; Oliveira, Micael J. T.; Rubio, Angel

    2017-12-01

    We implemented various DFT+U schemes, including the Agapito, Curtarolo, and Buongiorno Nardelli functional (ACBN0) self-consistent density-functional version of the DFT +U method [Phys. Rev. X 5, 011006 (2015), 10.1103/PhysRevX.5.011006] within the massively parallel real-space time-dependent density functional theory (TDDFT) code octopus. We further extended the method to the case of the calculation of response functions with real-time TDDFT+U and to the description of noncollinear spin systems. The implementation is tested by investigating the ground-state and optical properties of various transition-metal oxides, bulk topological insulators, and molecules. Our results are found to be in good agreement with previously published results for both the electronic band structure and structural properties. The self-consistent calculated values of U and J are also in good agreement with the values commonly used in the literature. We found that the time-dependent extension of the self-consistent DFT+U method yields improved optical properties when compared to the empirical TDDFT+U scheme. This work thus opens a different theoretical framework to address the nonequilibrium properties of correlated systems.

  17. Breaking CFD Bottlenecks in Gas-Turbine Flow-Path Design

    NASA Technical Reports Server (NTRS)

    Davis, Roger L.; Dannenhoffer, John F., III; Clark, John P.

    2010-01-01

    New ideas are forthcoming to break existing bottlenecks in using CFD during design. CAD-based automated grid generation. Multi-disciplinary use of embedded, overset grids to eliminate complex gridding problems. Use of time-averaged detached-eddy simulations as norm instead of "steady" RANS to include effects of self-excited unsteadiness. Combined GPU/Core parallel computing to provide over an order of magnitude increase in performance/price ratio. Gas-turbine applications are shown here but these ideas can be used for other Air Force, Navy, and NASA applications.

  18. Study protocol of the CAREST-trial: a randomised controlled trial on the (cost-) effectiveness of a CBT-based online self-help training for fear of cancer recurrence in women with curatively treated breast cancer.

    PubMed

    van Helmondt, Sanne Jasperine; van der Lee, Marije Liesbeth; de Vries, Jolanda

    2016-07-25

    One of the most prevalent long-term consequences of surviving breast cancer is fear of cancer recurrence (FCR), which is associated with higher (mental) healthcare costs and lower surveillance rates. The majority of breast cancer survivors report a need for professional help in dealing with FCR. An easy-accessible and cost-effective evidence-based psychological intervention for reducing FCR is lacking. In the current study an online self-help training to reduce FCR will be evaluated. In addition, the secondary aim of this study is to identify factors that predict whether women can benefit from the online self-help training or not. A multi-centre, parallel-groups, randomised controlled trial will be conducted to evaluate the (cost-) effectiveness of the CAREST-trial. A sample of 454 women with curatively treated breast cancer will be recruited from 8 hospitals in the Netherlands. Participants will be randomised to the intervention or usual care group (1:1). Self-report measures will be completed at baseline, 3 (post-intervention), 9, and 24 months. Primary outcome is FCR severity; secondary outcomes are healthcare costs, health status, and psychological distress. The online tailored self-help training "Less fear after cancer" is based on cognitive behavioural therapy and consists of 2 basic modules (psycho-education; basic principles of cognitive behavioural therapy) and 4 optional modules (rumination; action; relaxation; reassurance) to choose from. Each module consists of an informative part (texts, videos, audio files) and a practical part (exercises). For every patient, the intervention will be available for three months. Personal online support by an e-mail coach is available. Online self-help training may be an easy-accessible and cost-effective treatment to reduce the impact of FCR at an early stage in a large group of breast cancer survivors. A strength is the 24 months follow-up period in the health economic evaluation. The results of the study will provide information on the possible strengths and benefits of online self-help training for FCR in breast cancer survivors. This study is registered at the Netherlands Trial Register ( NTR4119 , date registered: August 15, 2013).

  19. Functional consistency across two behavioural modalities: fire-setting and self-harm in female special hospital patients.

    PubMed

    Miller, Sarah; Fritzon, Katarina

    2007-01-01

    Fire-setting and self-harm behaviours among women in high security special hospitals may be understood using Shye's Action System Theory (AST) in which four functional modes are recognized: 'adaptive', 'expressive', 'integrative', and 'conservative'. To test for relationships between different forms of fire-setting and self-harm behaviours and AST modes among women in special hospital, and for consistency within modes across the two behaviours. Clinical case files evidencing both fire-setting and self-harm behaviours (n = 50) were analysed for content, focusing on incident characteristics. A total of 29 fire-setting and 22 self-harm variables were analysed using Smallest Space Analysis (SSA). Chi-square and Spearman's rho (rho) analyses were used to determine functional consistency across behavioural modes. Most women showed one predominant AST mode in fire-setting (n = 39) and self-harm (n = 35). Significant positive correlations were found between integrative and adaptive modes of functioning. The lack of correlation between conservative and expressive modes reflects the differing behaviours used in each activity. Despite this, significant cross-tabulations revealed that each woman had parallel fire-setting and self-harm styles. Findings suggest that, for some women, setting fires and self harm fulfil a similar underlying function. Support is given to AST as a way of furthering understanding of damaging behaviours, whether self- or other-inflicted. Copyright 2007 John Wiley & Sons, Ltd.

  20. A Multi-Level Parallelization Concept for High-Fidelity Multi-Block Solvers

    NASA Technical Reports Server (NTRS)

    Hatay, Ferhat F.; Jespersen, Dennis C.; Guruswamy, Guru P.; Rizk, Yehia M.; Byun, Chansup; Gee, Ken; VanDalsem, William R. (Technical Monitor)

    1997-01-01

    The integration of high-fidelity Computational Fluid Dynamics (CFD) analysis tools with the industrial design process benefits greatly from the robust implementations that are transportable across a wide range of computer architectures. In the present work, a hybrid domain-decomposition and parallelization concept was developed and implemented into the widely-used NASA multi-block Computational Fluid Dynamics (CFD) packages implemented in ENSAERO and OVERFLOW. The new parallel solver concept, PENS (Parallel Euler Navier-Stokes Solver), employs both fine and coarse granularity in data partitioning as well as data coalescing to obtain the desired load-balance characteristics on the available computer platforms. This multi-level parallelism implementation itself introduces no changes to the numerical results, hence the original fidelity of the packages are identically preserved. The present implementation uses the Message Passing Interface (MPI) library for interprocessor message passing and memory accessing. By choosing an appropriate combination of the available partitioning and coalescing capabilities only during the execution stage, the PENS solver becomes adaptable to different computer architectures from shared-memory to distributed-memory platforms with varying degrees of parallelism. The PENS implementation on the IBM SP2 distributed memory environment at the NASA Ames Research Center obtains 85 percent scalable parallel performance using fine-grain partitioning of single-block CFD domains using up to 128 wide computational nodes. Multi-block CFD simulations of complete aircraft simulations achieve 75 percent perfect load-balanced executions using data coalescing and the two levels of parallelism. SGI PowerChallenge, SGI Origin 2000, and a cluster of workstations are the other platforms where the robustness of the implementation is tested. The performance behavior on the other computer platforms with a variety of realistic problems will be included as this on-going study progresses.

  1. [Effects of a Multi-disciplinary Approached, Empowerment Theory Based Self-management Intervention in Older Adults with Chronic Illness].

    PubMed

    Park, Chorong; Song, Misoon; Cho, Belong; Lim, Jaeyoung; Song, Wook; Chang, Heekyung; Park, Yeon-Hwan

    2015-04-01

    The purpose of this study was to develop a multi-disciplinary self-management intervention based on empowerment theory and to evaluate the effectiveness of the intervention for older adults with chronic illness. A randomized controlled trial design was used with 43 Korean older adults with chronic illness (Experimental group=22, Control group=21). The intervention consisted of two phases: (1) 8-week multi-disciplinary, team guided, group-based health education, exercise session, and individual empowerment counseling, (2) 16-week self-help group activities including weekly exercise and group discussion to maintain acquired self-management skills and problem-solving skills. Baseline, 8-week, and 24-week assessments measured health empowerment, exercise self-efficacy, physical activity, and physical function. Health empowerment, physical activity, and physical function in the experimental group increased significantly compared to the control group over time. Exercise self-efficacy significantly increased in experimental group over time but there was no significant difference between the two groups. The self-management program based on empowerment theory improved health empowerment, physical activity, and physical function in older adults. The study finding suggests that a health empowerment strategy may be an effective approach for older adults with multiple chronic illnesses in terms of achieving a sense of control over their chronic illness and actively engaging self-management.

  2. Hybrid Model of Inhomogeneous Solar Wind Plasma Heating by Alfven Wave Spectrum: Parametric Studies

    NASA Technical Reports Server (NTRS)

    Ofman, L.

    2010-01-01

    Observations of the solar wind plasma at 0.3 AU and beyond show that a turbulent spectrum of magnetic fluctuations is present. Remote sensing observations of the corona indicate that heavy ions are hotter than protons and their temperature is anisotropic (T(sub perpindicular / T(sub parallel) >> 1). We study the heating and the acceleration of multi-ion plasma in the solar wind by a turbulent spectrum of Alfvenic fluctuations using a 2-D hybrid numerical model. In the hybrid model the protons and heavy ions are treated kinetically as particles, while the electrons are included as neutralizing background fluid. This is the first two-dimensional hybrid parametric study of the solar wind plasma that includes an input turbulent wave spectrum guided by observation with inhomogeneous background density. We also investigate the effects of He++ ion beams in the inhomogeneous background plasma density on the heating of the solar wind plasma. The 2-D hybrid model treats parallel and oblique waves, together with cross-field inhomogeneity, self-consistently. We investigate the parametric dependence of the perpendicular heating, and the temperature anisotropy in the H+-He++ solar wind plasma. It was found that the scaling of the magnetic fluctuations power spectrum steepens in the higher-density regions, and the heating is channeled to these regions from the surrounding lower-density plasma due to wave refraction. The model parameters are applicable to the expected solar wind conditions at about 10 solar radii.

  3. Parallel Multi-Step/Multi-Rate Integration of Two-Time Scale Dynamic Systems

    NASA Technical Reports Server (NTRS)

    Chang, Johnny T.; Ploen, Scott R.; Sohl, Garett. A,; Martin, Bryan J.

    2004-01-01

    Increasing demands on the fidelity of simulations for real-time and high-fidelity simulations are stressing the capacity of modern processors. New integration techniques are required that provide maximum efficiency for systems that are parallelizable. However many current techniques make assumptions that are at odds with non-cascadable systems. A new serial multi-step/multi-rate integration algorithm for dual-timescale continuous state systems is presented which applies to these systems, and is extended to a parallel multi-step/multi-rate algorithm. The superior performance of both algorithms is demonstrated through a representative example.

  4. Efficient Parallelization of a Dynamic Unstructured Application on the Tera MTA

    NASA Technical Reports Server (NTRS)

    Oliker, Leonid; Biswas, Rupak

    1999-01-01

    The success of parallel computing in solving real-life computationally-intensive problems relies on their efficient mapping and execution on large-scale multiprocessor architectures. Many important applications are both unstructured and dynamic in nature, making their efficient parallel implementation a daunting task. This paper presents the parallelization of a dynamic unstructured mesh adaptation algorithm using three popular programming paradigms on three leading supercomputers. We examine an MPI message-passing implementation on the Cray T3E and the SGI Origin2OOO, a shared-memory implementation using cache coherent nonuniform memory access (CC-NUMA) of the Origin2OOO, and a multi-threaded version on the newly-released Tera Multi-threaded Architecture (MTA). We compare several critical factors of this parallel code development, including runtime, scalability, programmability, and memory overhead. Our overall results demonstrate that multi-threaded systems offer tremendous potential for quickly and efficiently solving some of the most challenging real-life problems on parallel computers.

  5. Parallelization strategies for continuum-generalized method of moments on the multi-thread systems

    NASA Astrophysics Data System (ADS)

    Bustamam, A.; Handhika, T.; Ernastuti, Kerami, D.

    2017-07-01

    Continuum-Generalized Method of Moments (C-GMM) covers the Generalized Method of Moments (GMM) shortfall which is not as efficient as Maximum Likelihood estimator by using the continuum set of moment conditions in a GMM framework. However, this computation would take a very long time since optimizing regularization parameter. Unfortunately, these calculations are processed sequentially whereas in fact all modern computers are now supported by hierarchical memory systems and hyperthreading technology, which allowing for parallel computing. This paper aims to speed up the calculation process of C-GMM by designing a parallel algorithm for C-GMM on the multi-thread systems. First, parallel regions are detected for the original C-GMM algorithm. There are two parallel regions in the original C-GMM algorithm, that are contributed significantly to the reduction of computational time: the outer-loop and the inner-loop. Furthermore, this parallel algorithm will be implemented with standard shared-memory application programming interface, i.e. Open Multi-Processing (OpenMP). The experiment shows that the outer-loop parallelization is the best strategy for any number of observations.

  6. Self-consistent quasi-static parallel electric field associated with substorm growth phase

    NASA Astrophysics Data System (ADS)

    Le Contel, O.; Pellat, R.; Roux, A.

    2000-06-01

    A new approach is proposed to calculate the self-consistent parallel electric field associated with the response of a plasma to quasi-static electromagnetic perturbations (ωωd<ω and ωd>ω (ωd being the bounce averaged magnetic drift frequency equal to kyvd, where ky is the wave number in the y direction and vd the bounce averaged magnetic drift velocity). The first regime (ωd<ω) corresponds to small particle energy and/or small ky, while the second regime (ωd>ω) is adapted to large energies and/or large ky. In particular, in the limit ωd<ω and |vd|<|uy|, where uy is the diamagnetic velocity proportional to the pressure gradient, we find a parallel electric field proportional to the pressure gradient and directed toward the ionosphere in the dusk sector and toward the equator in the dawn sector. This parallel electric field corresponds to a potential drop of a few hundred volts that can accelerate electrons and produce a differential drift between electrons and ions.

  7. On the utility of graphics cards to perform massively parallel simulation of advanced Monte Carlo methods

    PubMed Central

    Lee, Anthony; Yau, Christopher; Giles, Michael B.; Doucet, Arnaud; Holmes, Christopher C.

    2011-01-01

    We present a case-study on the utility of graphics cards to perform massively parallel simulation of advanced Monte Carlo methods. Graphics cards, containing multiple Graphics Processing Units (GPUs), are self-contained parallel computational devices that can be housed in conventional desktop and laptop computers and can be thought of as prototypes of the next generation of many-core processors. For certain classes of population-based Monte Carlo algorithms they offer massively parallel simulation, with the added advantage over conventional distributed multi-core processors that they are cheap, easily accessible, easy to maintain, easy to code, dedicated local devices with low power consumption. On a canonical set of stochastic simulation examples including population-based Markov chain Monte Carlo methods and Sequential Monte Carlo methods, we nd speedups from 35 to 500 fold over conventional single-threaded computer code. Our findings suggest that GPUs have the potential to facilitate the growth of statistical modelling into complex data rich domains through the availability of cheap and accessible many-core computation. We believe the speedup we observe should motivate wider use of parallelizable simulation methods and greater methodological attention to their design. PMID:22003276

  8. Research in computer science

    NASA Technical Reports Server (NTRS)

    Ortega, J. M.

    1986-01-01

    Various graduate research activities in the field of computer science are reported. Among the topics discussed are: (1) failure probabilities in multi-version software; (2) Gaussian Elimination on parallel computers; (3) three dimensional Poisson solvers on parallel/vector computers; (4) automated task decomposition for multiple robot arms; (5) multi-color incomplete cholesky conjugate gradient methods on the Cyber 205; and (6) parallel implementation of iterative methods for solving linear equations.

  9. Development of slew-rate-limited time-over-threshold (ToT) ASIC for a multi-channel silicon-based ion detector

    NASA Astrophysics Data System (ADS)

    Uenomachi, M.; Orita, T.; Shimazoe, K.; Takahashi, H.; Ikeda, H.; Tsujita, K.; Sekiba, D.

    2018-01-01

    High-resolution Elastic Recoil Detection Analysis (HERDA), which consists of a 90o sector magnetic spectrometer and a position-sensitive detector (PSD), is a method of quantitative hydrogen analysis. In order to increase sensitivity, a HERDA system using a multi-channel silicon-based ion detector has been developed. Here, as a parallel and fast readout circuit from a multi-channel silicon-based ion detector, a slew-rate-limited time-over-threshold (ToT) application-specific integrated circuit (ASIC) was designed, and a new slew-rate-limited ToT method is proposed. The designed ASIC has 48 channels and each channel consists of a preamplifier, a slew-rate-limited shaping amplifier, which makes ToT response linear, and a comparator. The measured equivalent noise charges (ENCs) of the preamplifier, the shaper, and the ToT on no detector capacitance were 253±21, 343±46, and 560±56 electrons RMS, respectively. The spectra from a 241Am source measured using a slew-rate-limited ToT ASIC are also reported.

  10. Extended MHD modeling of tearing-driven magnetic relaxation

    NASA Astrophysics Data System (ADS)

    Sauppe, J. P.; Sovinec, C. R.

    2017-05-01

    Discrete relaxation events in reversed-field pinch relevant configurations are investigated numerically with nonlinear extended magnetohydrodynamic (MHD) modeling, including the Hall term in Ohm's law and first-order ion finite Larmor radius effects. Results show variability among relaxation events, where the Hall dynamo effect may help or impede the MHD dynamo effect in relaxing the parallel current density profile. The competitive behavior arises from multi-helicity conditions where the dominant magnetic fluctuation is relatively small. The resulting changes in parallel current density and parallel flow are aligned in the core, consistent with experimental observations. The analysis of simulation results also confirms that the force density from fluctuation-induced Reynolds stress arises subsequent to the drive from the fluctuation-induced Lorentz force density. Transport of the momentum density is found to be dominated by the fluctuation-induced Maxwell stress over most of the cross section with viscous and gyroviscous contributions being large in the edge region. The findings resolve a discrepancy with respect to the relative orientation of current density and flow relaxation, which had not been realized or investigated in King et al. [Phys. Plasmas 19, 055905 (2012)], where only the magnitude of flow relaxation is actually consistent with experimental results.

  11. MCSCF wave functions for excited states of polar molecules - Application to BeO. [Multi-Configuration Self-Consistent Field

    NASA Technical Reports Server (NTRS)

    Bauschlicher, C. W., Jr.; Yarkony, D. R.

    1980-01-01

    A previously reported multi-configuration self-consistent field (MCSCF) algorithm based on the generalized Brillouin theorem is extended in order to treat the excited states of polar molecules. In particular, the algorithm takes into account the proper treatment of nonorthogonality in the space of single excitations and invokes, when necessary, a constrained optimization procedure to prevent the variational collapse of excited states. In addition, a configuration selection scheme (suitable for use in conjunction with extended configuration interaction methods) is proposed for the MCSCF procedure. The algorithm is used to study the low-lying singlet states of BeO, a system which has not previously been studied using an MCSCF procedure. MCSCF wave functions are obtained for three 1 Sigma + and two 1 Pi states. The 1 Sigma + results are juxtaposed with comparable results for MgO in order to assess the generality of the description presented here.

  12. Parameter estimation in large-scale systems biology models: a parallel and self-adaptive cooperative strategy.

    PubMed

    Penas, David R; González, Patricia; Egea, Jose A; Doallo, Ramón; Banga, Julio R

    2017-01-21

    The development of large-scale kinetic models is one of the current key issues in computational systems biology and bioinformatics. Here we consider the problem of parameter estimation in nonlinear dynamic models. Global optimization methods can be used to solve this type of problems but the associated computational cost is very large. Moreover, many of these methods need the tuning of a number of adjustable search parameters, requiring a number of initial exploratory runs and therefore further increasing the computation times. Here we present a novel parallel method, self-adaptive cooperative enhanced scatter search (saCeSS), to accelerate the solution of this class of problems. The method is based on the scatter search optimization metaheuristic and incorporates several key new mechanisms: (i) asynchronous cooperation between parallel processes, (ii) coarse and fine-grained parallelism, and (iii) self-tuning strategies. The performance and robustness of saCeSS is illustrated by solving a set of challenging parameter estimation problems, including medium and large-scale kinetic models of the bacterium E. coli, bakerés yeast S. cerevisiae, the vinegar fly D. melanogaster, Chinese Hamster Ovary cells, and a generic signal transduction network. The results consistently show that saCeSS is a robust and efficient method, allowing very significant reduction of computation times with respect to several previous state of the art methods (from days to minutes, in several cases) even when only a small number of processors is used. The new parallel cooperative method presented here allows the solution of medium and large scale parameter estimation problems in reasonable computation times and with small hardware requirements. Further, the method includes self-tuning mechanisms which facilitate its use by non-experts. We believe that this new method can play a key role in the development of large-scale and even whole-cell dynamic models.

  13. The design of multi-core DSP parallel model based on message passing and multi-level pipeline

    NASA Astrophysics Data System (ADS)

    Niu, Jingyu; Hu, Jian; He, Wenjing; Meng, Fanrong; Li, Chuanrong

    2017-10-01

    Currently, the design of embedded signal processing system is often based on a specific application, but this idea is not conducive to the rapid development of signal processing technology. In this paper, a parallel processing model architecture based on multi-core DSP platform is designed, and it is mainly suitable for the complex algorithms which are composed of different modules. This model combines the ideas of multi-level pipeline parallelism and message passing, and summarizes the advantages of the mainstream model of multi-core DSP (the Master-Slave model and the Data Flow model), so that it has better performance. This paper uses three-dimensional image generation algorithm to validate the efficiency of the proposed model by comparing with the effectiveness of the Master-Slave and the Data Flow model.

  14. Enhancing Self-Determination in Health: Results of an RCT of the Ask Project, a School-Based Intervention for Adolescents with Intellectual Disability

    ERIC Educational Resources Information Center

    McPherson, Lyn; Ware, Robert S.; Carrington, Suzanne; Lennox, Nicholas

    2017-01-01

    Background: Adolescents with intellectual disability have high levels of unrecognized disease and inadequate health screening/promotion which might be addressed by improving health advocacy skills. Methods: A parallel-group cluster randomized controlled trial was conducted to investigate whether a health intervention package, consisting of…

  15. No-insulation multi-width winding technique for high temperature superconducting magnet.

    PubMed

    Hahn, Seungyong; Kim, Youngjae; Keun Park, Dong; Kim, Kwangmin; Voccio, John P; Bascuñán, Juan; Iwasa, Yukikazu

    2013-10-21

    We present a No-Insulation ( NI ) Multi-Width ( MW ) winding technique for an HTS (high temperature superconductor) magnet consisting of double-pancake (DP) coils. The NI enables an HTS magnet self-protecting and the MW minimizes the detrimental anisotropy in current-carrying capacity of HTS tape by assigning tapes of multiple widths to DP coils within a stack, widest tape to the top and bottom sections and the narrowest in the midplane section. This paper presents fabrication and test results of an NI-MW HTS magnet and demonstrates the unique features of the NI-MW technique: self-protecting and enhanced field performance, unattainable with the conventional technique.

  16. Neoclassical parallel flow calculation in the presence of external parallel momentum sources in Heliotron J

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nishioka, K.; Nakamura, Y.; Nishimura, S.

    A moment approach to calculate neoclassical transport in non-axisymmetric torus plasmas composed of multiple ion species is extended to include the external parallel momentum sources due to unbalanced tangential neutral beam injections (NBIs). The momentum sources that are included in the parallel momentum balance are calculated from the collision operators of background particles with fast ions. This method is applied for the clarification of the physical mechanism of the neoclassical parallel ion flows and the multi-ion species effect on them in Heliotron J NBI plasmas. It is found that parallel ion flow can be determined by the balance between themore » parallel viscosity and the external momentum source in the region where the external source is much larger than the thermodynamic force driven source in the collisional plasmas. This is because the friction between C{sup 6+} and D{sup +} prevents a large difference between C{sup 6+} and D{sup +} flow velocities in such plasmas. The C{sup 6+} flow velocities, which are measured by the charge exchange recombination spectroscopy system, are numerically evaluated with this method. It is shown that the experimentally measured C{sup 6+} impurity flow velocities do not contradict clearly with the neoclassical estimations, and the dependence of parallel flow velocities on the magnetic field ripples is consistent in both results.« less

  17. Possibility of Cooper-pair formation controlled by multi-terminal spin injection

    NASA Astrophysics Data System (ADS)

    Ohnishi, K.; Sakamoto, M.; Ishitaki, M.; Kimura, T.

    2018-03-01

    A multi-terminal lateral spin valve consisting of three ferromagnetic nanopillars on a Cu/Nb bilayer has been fabricated. We investigated the influence of the spin injection on the superconducting properties at the Cu/Nb interface. The non-local spin valve signal exhibits a clear spin insulation signature due to the superconducting gap of the Nb. The magnitude of the spin signal is found to show the probe configuration dependence. From the careful analysis of the bias current dependence, we found the suppression of the superconductivity due to the exchange interaction between the Cooper pair and accumulated spin plays an important role in the multi-terminal spin injections. We also discuss about the possibility of the Cooper-pair formation due to the spin injection from the two injectors with the anti-parallel alignment.

  18. Parent and Self-Report Ratings on the Perceived Levels of Social Vulnerability of Adults with Williams Syndrome.

    PubMed

    Lough, Emma; Fisher, Marisa H

    2016-11-01

    The current study took a multi-informant approach to compare parent to self-report ratings of social vulnerability of adults with Williams syndrome (WS). Participants included 102 pairs of adults with WS and their parents. Parents completed the Social Vulnerability Questionnaire and adults with WS completed an adapted version of the questionnaire. Parents consistently reported higher levels of social vulnerability for their son/daughter than the individual with WS reported, with the exception of emotional abuse. The lower ratings of social vulnerability by adults with WS, compared to their parents, offer new information about their insight into their own vulnerability. These findings highlight the importance of teaching self-awareness as a part of a multi-informant approach to interventions designed to target social vulnerability.

  19. GENESIS: a hybrid-parallel and multi-scale molecular dynamics simulator with enhanced sampling algorithms for biomolecular and cellular simulations

    PubMed Central

    Jung, Jaewoon; Mori, Takaharu; Kobayashi, Chigusa; Matsunaga, Yasuhiro; Yoda, Takao; Feig, Michael; Sugita, Yuji

    2015-01-01

    GENESIS (Generalized-Ensemble Simulation System) is a new software package for molecular dynamics (MD) simulations of macromolecules. It has two MD simulators, called ATDYN and SPDYN. ATDYN is parallelized based on an atomic decomposition algorithm for the simulations of all-atom force-field models as well as coarse-grained Go-like models. SPDYN is highly parallelized based on a domain decomposition scheme, allowing large-scale MD simulations on supercomputers. Hybrid schemes combining OpenMP and MPI are used in both simulators to target modern multicore computer architectures. Key advantages of GENESIS are (1) the highly parallel performance of SPDYN for very large biological systems consisting of more than one million atoms and (2) the availability of various REMD algorithms (T-REMD, REUS, multi-dimensional REMD for both all-atom and Go-like models under the NVT, NPT, NPAT, and NPγT ensembles). The former is achieved by a combination of the midpoint cell method and the efficient three-dimensional Fast Fourier Transform algorithm, where the domain decomposition space is shared in real-space and reciprocal-space calculations. Other features in SPDYN, such as avoiding concurrent memory access, reducing communication times, and usage of parallel input/output files, also contribute to the performance. We show the REMD simulation results of a mixed (POPC/DMPC) lipid bilayer as a real application using GENESIS. GENESIS is released as free software under the GPLv2 licence and can be easily modified for the development of new algorithms and molecular models. WIREs Comput Mol Sci 2015, 5:310–323. doi: 10.1002/wcms.1220 PMID:26753008

  20. Research on the parallel load sharing principle of a novel self-decoupled piezoelectric six-dimensional force sensor.

    PubMed

    Li, Ying-Jun; Yang, Cong; Wang, Gui-Cong; Zhang, Hui; Cui, Huan-Yong; Zhang, Yong-Liang

    2017-09-01

    This paper presents a novel integrated piezoelectric six-dimensional force sensor which can realize dynamic measurement of multi-dimensional space load. Firstly, the composition of the sensor, the spatial layout of force-sensitive components, and measurement principle are analyzed and designed. There is no interference of piezoelectric six-dimensional force sensor in theoretical analysis. Based on the principle of actual work and deformation compatibility coherence, this paper deduces the parallel load sharing principle of the piezoelectric six-dimensional force sensor. The main effect factors which affect the load sharing ratio are obtained. The finite element model of the piezoelectric six-dimensional force sensor is established. In order to verify the load sharing principle of the sensor, a load sharing test device of piezoelectric force sensor is designed and fabricated. The load sharing experimental platform is set up. The experimental results are in accordance with the theoretical analysis and simulation results. The experiments show that the multi-dimensional and heavy force measurement can be realized by the parallel arrangement of the load sharing ring and the force sensitive element in the novel integrated piezoelectric six-dimensional force sensor. The ideal load sharing effect of the sensor can be achieved by appropriate size parameters. This paper has an important guide for the design of the force measuring device according to the load sharing mode. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.

  1. Acceleration of Semiempirical QM/MM Methods through Message Passage Interface (MPI), Hybrid MPI/Open Multiprocessing, and Self-Consistent Field Accelerator Implementations.

    PubMed

    Ojeda-May, Pedro; Nam, Kwangho

    2017-08-08

    The strategy and implementation of scalable and efficient semiempirical (SE) QM/MM methods in CHARMM are described. The serial version of the code was first profiled to identify routines that required parallelization. Afterward, the code was parallelized and accelerated with three approaches. The first approach was the parallelization of the entire QM/MM routines, including the Fock matrix diagonalization routines, using the CHARMM message passage interface (MPI) machinery. In the second approach, two different self-consistent field (SCF) energy convergence accelerators were implemented using density and Fock matrices as targets for their extrapolations in the SCF procedure. In the third approach, the entire QM/MM and MM energy routines were accelerated by implementing the hybrid MPI/open multiprocessing (OpenMP) model in which both the task- and loop-level parallelization strategies were adopted to balance loads between different OpenMP threads. The present implementation was tested on two solvated enzyme systems (including <100 QM atoms) and an S N 2 symmetric reaction in water. The MPI version exceeded existing SE QM methods in CHARMM, which include the SCC-DFTB and SQUANTUM methods, by at least 4-fold. The use of SCF convergence accelerators further accelerated the code by ∼12-35% depending on the size of the QM region and the number of CPU cores used. Although the MPI version displayed good scalability, the performance was diminished for large numbers of MPI processes due to the overhead associated with MPI communications between nodes. This issue was partially overcome by the hybrid MPI/OpenMP approach which displayed a better scalability for a larger number of CPU cores (up to 64 CPUs in the tested systems).

  2. The interaction between fishbone modes and shear Alfvén waves in tokamak plasmas

    NASA Astrophysics Data System (ADS)

    He, Hongda; Liu, Yueqiang; Dong, J. Q.; Hao, G. Z.; Wu, Tingting; He, Zhixiong; Zhao, K.

    2016-05-01

    The resonant interaction between the energetic particle triggered fishbone mode and the shear Alfvén waves is computationally investigated and firmly demonstrated based on a tokamak plasma equilibrium, using the self-consistent MHD-kinetic hybrid code MARS-K (Liu et al 2008 Phys. Plasmas 15 112503). This type of continuum resonance, occurring critically due to the mode’s toroidal rotation in the plasma frame, significantly modifies the eigenmode structure of the fishbone instability, by introducing two large peaks of the perturbed parallel current density near but offside the q  =  1 rational surface (q is the safety factor). The self-consistently computed radial plasma displacement substantially differs from that being assumed in the conventional fishbone theory.

  3. Telescopic multi-resolution augmented reality

    NASA Astrophysics Data System (ADS)

    Jenkins, Jeffrey; Frenchi, Christopher; Szu, Harold

    2014-05-01

    To ensure a self-consistent scaling approximation, the underlying microscopic fluctuation components can naturally influence macroscopic means, which may give rise to emergent observable phenomena. In this paper, we describe a consistent macroscopic (cm-scale), mesoscopic (micron-scale), and microscopic (nano-scale) approach to introduce Telescopic Multi-Resolution (TMR) into current Augmented Reality (AR) visualization technology. We propose to couple TMR-AR by introducing an energy-matter interaction engine framework that is based on known Physics, Biology, Chemistry principles. An immediate payoff of TMR-AR is a self-consistent approximation of the interaction between microscopic observables and their direct effect on the macroscopic system that is driven by real-world measurements. Such an interdisciplinary approach enables us to achieve more than multiple scale, telescopic visualization of real and virtual information but also conducting thought experiments through AR. As a result of the consistency, this framework allows us to explore a large dimensionality parameter space of measured and unmeasured regions. Towards this direction, we explore how to build learnable libraries of biological, physical, and chemical mechanisms. Fusing analytical sensors with TMR-AR libraries provides a robust framework to optimize testing and evaluation through data-driven or virtual synthetic simulations. Visualizing mechanisms of interactions requires identification of observable image features that can indicate the presence of information in multiple spatial and temporal scales of analog data. The AR methodology was originally developed to enhance pilot-training as well as `make believe' entertainment industries in a user-friendly digital environment We believe TMR-AR can someday help us conduct thought experiments scientifically, to be pedagogically visualized in a zoom-in-and-out, consistent, multi-scale approximations.

  4. The effect of anisotropic heat transport on magnetic islands in 3-D configurations

    NASA Astrophysics Data System (ADS)

    Schlutt, M. G.; Hegna, C. C.

    2012-08-01

    An analytic theory of nonlinear pressure-induced magnetic island formation using a boundary layer analysis is presented. This theory extends previous work by including the effects of finite parallel heat transport and is applicable to general three dimensional magnetic configurations. In this work, particular attention is paid to the role of finite parallel heat conduction in the context of pressure-induced island physics. It is found that localized currents that require self-consistent deformation of the pressure profile, such as resistive interchange and bootstrap currents, are attenuated by finite parallel heat conduction when the magnetic islands are sufficiently small. However, these anisotropic effects do not change saturated island widths caused by Pfirsch-Schlüter current effects. Implications for finite pressure-induced island healing are discussed.

  5. Differentiating Adolescent Self-Injury from Adolescent Depression: Possible Implications for Borderline Personality Development

    PubMed Central

    Crowell, Sheila E.; Beauchaine, Theodore P.; Hsiao, Ray C.; Vasilev, Christina A.; Yaptangco, Mona; Linehan, Marsha M.; McCauley, Elizabeth

    2011-01-01

    Self-inflicted injury (SII) in adolescence marks heightened risk for suicide attempts, completed suicide, and adult psychopathology. Although several studies have revealed elevated rates of depression among adolescents who self injure, no one has compared adolescent self injury with adolescent depression on biological, self-, and informant-report markers of vulnerability and risk. Such a comparison may have important implications for treatment, prevention, and developmental models of self injury and borderline personality disorder. We used a multi-method, multi-informant approach to examine how adolescent SII differs from adolescent depression. Self-injuring, depressed, and typical adolescent females (n = 25 per group) and their mothers completed measures of psychopathology and emotion regulation, among others. In addition, we assessed electrodermal responding (EDR), a peripheral biomarker of trait impulsivity. Participants in the SII group (a) scored higher than depressed adolescents on measures of both externalizing psychopathology and emotion dysregulation, and (b) exhibited attenuated EDR, similar to patterns observed among impulsive, externalizing males. Self-injuring adolescents also scored higher on measures of borderline pathology. These findings reveal a coherent pattern of differences between self-injuring and depressed adolescent girls, consistent with theories that SII differs from depression in etiology and developmental course. PMID:22016199

  6. A Review of Lightweight Thread Approaches for High Performance Computing

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Castello, Adrian; Pena, Antonio J.; Seo, Sangmin

    High-level, directive-based solutions are becoming the programming models (PMs) of the multi/many-core architectures. Several solutions relying on operating system (OS) threads perfectly work with a moderate number of cores. However, exascale systems will spawn hundreds of thousands of threads in order to exploit their massive parallel architectures and thus conventional OS threads are too heavy for that purpose. Several lightweight thread (LWT) libraries have recently appeared offering lighter mechanisms to tackle massive concurrency. In order to examine the suitability of LWTs in high-level runtimes, we develop a set of microbenchmarks consisting of commonlyfound patterns in current parallel codes. Moreover, wemore » study the semantics offered by some LWT libraries in order to expose the similarities between different LWT application programming interfaces. This study reveals that a reduced set of LWT functions can be sufficient to cover the common parallel code patterns and that those LWT libraries perform better than OS threads-based solutions in cases where task and nested parallelism are becoming more popular with new architectures.« less

  7. Marine Controlled-Source Electromagnetic 2D Inversion for synthetic models.

    NASA Astrophysics Data System (ADS)

    Liu, Y.; Li, Y.

    2016-12-01

    We present a 2D inverse algorithm for frequency domain marine controlled-source electromagnetic (CSEM) data, which is based on the regularized Gauss-Newton approach. As a forward solver, our parallel adaptive finite element forward modeling program is employed. It is a self-adaptive, goal-oriented grid refinement algorithm in which a finite element analysis is performed on a sequence of refined meshes. The mesh refinement process is guided by a dual error estimate weighting to bias refinement towards elements that affect the solution at the EM receiver locations. With the use of the direct solver (MUMPS), we can effectively compute the electromagnetic fields for multi-sources and parametric sensitivities. We also implement the parallel data domain decomposition approach of Key and Ovall (2011), with the goal of being able to compute accurate responses in parallel for complicated models and a full suite of data parameters typical of offshore CSEM surveys. All minimizations are carried out by using the Gauss-Newton algorithm and model perturbations at each iteration step are obtained by using the Inexact Conjugate Gradient iteration method. Synthetic test inversions are presented.

  8. Application of a self-consistent NEGF procedure to study the coherent transport with phase breaking scattering in low dimensional systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pratap, Surender, E-mail: surender.pratap@pilani.bits-pilani.ac.in; Sarkar, Niladri, E-mail: niladri@pilani.bits-pilani.ac.in

    2016-04-13

    We have studied Quantum Transport with dephasing in Low Dimensional systems. Here, we apply a self-consistent NEGF procedure to study the transport mechanism in low-dimensional systems with phase breaking scatterers. Under this we have determined the transmission coefficient of a very small Multi-Moded Nanowire which is under a small bias potential of few meV. We have calculated the transmission of this device first with no scatterers. Then we have introduced scatterers in the device and calculated the transmission for the device.

  9. Blade row dynamic digital compressor program. Volume 1: J85 clean inlet flow and parallel compressor models

    NASA Technical Reports Server (NTRS)

    Tesch, W. A.; Steenken, W. G.

    1976-01-01

    The results are presented of a one-dimensional dynamic digital blade row compressor model study of a J85-13 engine operating with uniform and with circumferentially distorted inlet flow. Details of the geometry and the derived blade row characteristics used to simulate the clean inlet performance are given. A stability criterion based upon the self developing unsteady internal flows near surge provided an accurate determination of the clean inlet surge line. The basic model was modified to include an arbitrary extent multi-sector parallel compressor configuration for investigating 180 deg 1/rev total pressure, total temperature, and combined total pressure and total temperature distortions. The combined distortions included opposed, coincident, and 90 deg overlapped patterns. The predicted losses in surge pressure ratio matched the measured data trends at all speeds and gave accurate predictions at high corrected speeds where the slope of the speed lines approached the vertical.

  10. Cellular Analogs of Operant Behavior.

    DTIC Science & Technology

    1992-07-31

    confirmed at the behavioral level in parallel studies of hippocampal and intravenous self-administration. The results are consistent with the hypothesis...research is to elucidate the cellular reinforcement mechanisms underlying goal-directed or operant behavior. The specific aim here was to study the... studies and behavioral operant conditioning studies . Most of the research reported was initiated and completed in the present project period. Also

  11. The Construction (Using Multi-Media Techniques) of Certain Modules of a Programmed Course in Astronomy-Space Sciences for NASA Personnel of The Goddard Space Flight Center, Greenbelt, Maryland.

    ERIC Educational Resources Information Center

    Collagan, Robert B.

    This paper describes the development of a self-instructional multi-media course in astronomy-space sciences for non-technical NASA personnel. The course consists of a variety of programed materials including slides, films, film-loops, filmstrips video-tapes and audio-tapes, on concepts of time, space, and matter in our solar system and galaxy.…

  12. Parallelization of interpolation, solar radiation and water flow simulation modules in GRASS GIS using OpenMP

    NASA Astrophysics Data System (ADS)

    Hofierka, Jaroslav; Lacko, Michal; Zubal, Stanislav

    2017-10-01

    In this paper, we describe the parallelization of three complex and computationally intensive modules of GRASS GIS using the OpenMP application programming interface for multi-core computers. These include the v.surf.rst module for spatial interpolation, the r.sun module for solar radiation modeling and the r.sim.water module for water flow simulation. We briefly describe the functionality of the modules and parallelization approaches used in the modules. Our approach includes the analysis of the module's functionality, identification of source code segments suitable for parallelization and proper application of OpenMP parallelization code to create efficient threads processing the subtasks. We document the efficiency of the solutions using the airborne laser scanning data representing land surface in the test area and derived high-resolution digital terrain model grids. We discuss the performance speed-up and parallelization efficiency depending on the number of processor threads. The study showed a substantial increase in computation speeds on a standard multi-core computer while maintaining the accuracy of results in comparison to the output from original modules. The presented parallelization approach showed the simplicity and efficiency of the parallelization of open-source GRASS GIS modules using OpenMP, leading to an increased performance of this geospatial software on standard multi-core computers.

  13. Interleaved EPI diffusion imaging using SPIRiT-based reconstruction with virtual coil compression.

    PubMed

    Dong, Zijing; Wang, Fuyixue; Ma, Xiaodong; Zhang, Zhe; Dai, Erpeng; Yuan, Chun; Guo, Hua

    2018-03-01

    To develop a novel diffusion imaging reconstruction framework based on iterative self-consistent parallel imaging reconstruction (SPIRiT) for multishot interleaved echo planar imaging (iEPI), with computation acceleration by virtual coil compression. As a general approach for autocalibrating parallel imaging, SPIRiT improves the performance of traditional generalized autocalibrating partially parallel acquisitions (GRAPPA) methods in that the formulation with self-consistency is better conditioned, suggesting SPIRiT to be a better candidate in k-space-based reconstruction. In this study, a general SPIRiT framework is adopted to incorporate both coil sensitivity and phase variation information as virtual coils and then is applied to 2D navigated iEPI diffusion imaging. To reduce the reconstruction time when using a large number of coils and shots, a novel shot-coil compression method is proposed for computation acceleration in Cartesian sampling. Simulations and in vivo experiments were conducted to evaluate the performance of the proposed method. Compared with the conventional coil compression, the shot-coil compression achieved higher compression rates with reduced errors. The simulation and in vivo experiments demonstrate that the SPIRiT-based reconstruction outperformed the existing method, realigned GRAPPA, and provided superior images with reduced artifacts. The SPIRiT-based reconstruction with virtual coil compression is a reliable method for high-resolution iEPI diffusion imaging. Magn Reson Med 79:1525-1531, 2018. © 2017 International Society for Magnetic Resonance in Medicine. © 2017 International Society for Magnetic Resonance in Medicine.

  14. CUDAMPF: a multi-tiered parallel framework for accelerating protein sequence search in HMMER on CUDA-enabled GPU.

    PubMed

    Jiang, Hanyu; Ganesan, Narayan

    2016-02-27

    HMMER software suite is widely used for analysis of homologous protein and nucleotide sequences with high sensitivity. The latest version of hmmsearch in HMMER 3.x, utilizes heuristic-pipeline which consists of MSV/SSV (Multiple/Single ungapped Segment Viterbi) stage, P7Viterbi stage and the Forward scoring stage to accelerate homology detection. Since the latest version is highly optimized for performance on modern multi-core CPUs with SSE capabilities, only a few acceleration attempts report speedup. However, the most compute intensive tasks within the pipeline (viz., MSV/SSV and P7Viterbi stages) still stand to benefit from the computational capabilities of massively parallel processors. A Multi-Tiered Parallel Framework (CUDAMPF) implemented on CUDA-enabled GPUs presented here, offers a finer-grained parallelism for MSV/SSV and Viterbi algorithms. We couple SIMT (Single Instruction Multiple Threads) mechanism with SIMD (Single Instructions Multiple Data) video instructions with warp-synchronism to achieve high-throughput processing and eliminate thread idling. We also propose a hardware-aware optimal allocation scheme of scarce resources like on-chip memory and caches in order to boost performance and scalability of CUDAMPF. In addition, runtime compilation via NVRTC available with CUDA 7.0 is incorporated into the presented framework that not only helps unroll innermost loop to yield upto 2 to 3-fold speedup than static compilation but also enables dynamic loading and switching of kernels depending on the query model size, in order to achieve optimal performance. CUDAMPF is designed as a hardware-aware parallel framework for accelerating computational hotspots within the hmmsearch pipeline as well as other sequence alignment applications. It achieves significant speedup by exploiting hierarchical parallelism on single GPU and takes full advantage of limited resources based on their own performance features. In addition to exceeding performance of other acceleration attempts, comprehensive evaluations against high-end CPUs (Intel i5, i7 and Xeon) shows that CUDAMPF yields upto 440 GCUPS for SSV, 277 GCUPS for MSV and 14.3 GCUPS for P7Viterbi all with 100 % accuracy, which translates to a maximum speedup of 37.5, 23.1 and 11.6-fold for MSV, SSV and P7Viterbi respectively. The source code is available at https://github.com/Super-Hippo/CUDAMPF.

  15. A new self-regulated self-excited single-phase induction generator using a squirrel cage three-phase induction machine

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fukami, Tadashi; Imamura, Michinori; Kaburaki, Yuichi

    1995-12-31

    A new single-phase capacitor self-excited induction generator with self-regulating feature is presented. The new generator consists of a squirrel cage three-phase induction machine and three capacitors connected in series and parallel with a single phase load. The voltage regulation of this generator is very small due to the effect of the three capacitors. Moreover, since a Y-connected stator winding is employed, the waveform of the output voltage becomes sinusoidal. In this paper the system configuration and the operating principle of the new generator are explained, and the basic characteristics are also investigated by means of a simple analysis and experimentsmore » with a laboratory machine.« less

  16. Winter precipitation particle size distribution measurement by Multi-Angle Snowflake Camera

    NASA Astrophysics Data System (ADS)

    Huang, Gwo-Jong; Kleinkort, Cameron; Bringi, V. N.; Notaroš, Branislav M.

    2017-12-01

    From the radar meteorology viewpoint, the most important properties for quantitative precipitation estimation of winter events are 3D shape, size, and mass of precipitation particles, as well as the particle size distribution (PSD). In order to measure these properties precisely, optical instruments may be the best choice. The Multi-Angle Snowflake Camera (MASC) is a relatively new instrument equipped with three high-resolution cameras to capture the winter precipitation particle images from three non-parallel angles, in addition to measuring the particle fall speed using two pairs of infrared motion sensors. However, the results from the MASC so far are usually presented as monthly or seasonally, and particle sizes are given as histograms, no previous studies have used the MASC for a single storm study, and no researchers use MASC to measure the PSD. We propose the methodology for obtaining the winter precipitation PSD measured by the MASC, and present and discuss the development, implementation, and application of the new technique for PSD computation based on MASC images. Overall, this is the first study of the MASC-based PSD. We present PSD MASC experiments and results for segments of two snow events to demonstrate the performance of our PSD algorithm. The results show that the self-consistency of the MASC measured single-camera PSDs is good. To cross-validate PSD measurements, we compare MASC mean PSD (averaged over three cameras) with the collocated 2D Video Disdrometer, and observe good agreements of the two sets of results.

  17. Accelerated Fast Spin-Echo Magnetic Resonance Imaging of the Heart Using a Self-Calibrated Split-Echo Approach

    PubMed Central

    Klix, Sabrina; Hezel, Fabian; Fuchs, Katharina; Ruff, Jan; Dieringer, Matthias A.; Niendorf, Thoralf

    2014-01-01

    Purpose Design, validation and application of an accelerated fast spin-echo (FSE) variant that uses a split-echo approach for self-calibrated parallel imaging. Methods For self-calibrated, split-echo FSE (SCSE-FSE), extra displacement gradients were incorporated into FSE to decompose odd and even echo groups which were independently phase encoded to derive coil sensitivity maps, and to generate undersampled data (reduction factor up to R = 3). Reference and undersampled data were acquired simultaneously. SENSE reconstruction was employed. Results The feasibility of SCSE-FSE was demonstrated in phantom studies. Point spread function performance of SCSE-FSE was found to be competitive with traditional FSE variants. The immunity of SCSE-FSE for motion induced mis-registration between reference and undersampled data was shown using a dynamic left ventricular model and cardiac imaging. The applicability of black blood prepared SCSE-FSE for cardiac imaging was demonstrated in healthy volunteers including accelerated multi-slice per breath-hold imaging and accelerated high spatial resolution imaging. Conclusion SCSE-FSE obviates the need of external reference scans for SENSE reconstructed parallel imaging with FSE. SCSE-FSE reduces the risk for mis-registration between reference scans and accelerated acquisitions. SCSE-FSE is feasible for imaging of the heart and of large cardiac vessels but also meets the needs of brain, abdominal and liver imaging. PMID:24728341

  18. SQDFT: Spectral Quadrature method for large-scale parallel O(N) Kohn-Sham calculations at high temperature

    NASA Astrophysics Data System (ADS)

    Suryanarayana, Phanish; Pratapa, Phanisri P.; Sharma, Abhiraj; Pask, John E.

    2018-03-01

    We present SQDFT: a large-scale parallel implementation of the Spectral Quadrature (SQ) method for O(N) Kohn-Sham Density Functional Theory (DFT) calculations at high temperature. Specifically, we develop an efficient and scalable finite-difference implementation of the infinite-cell Clenshaw-Curtis SQ approach, in which results for the infinite crystal are obtained by expressing quantities of interest as bilinear forms or sums of bilinear forms, that are then approximated by spatially localized Clenshaw-Curtis quadrature rules. We demonstrate the accuracy of SQDFT by showing systematic convergence of energies and atomic forces with respect to SQ parameters to reference diagonalization results, and convergence with discretization to established planewave results, for both metallic and insulating systems. We further demonstrate that SQDFT achieves excellent strong and weak parallel scaling on computer systems consisting of tens of thousands of processors, with near perfect O(N) scaling with system size and wall times as low as a few seconds per self-consistent field iteration. Finally, we verify the accuracy of SQDFT in large-scale quantum molecular dynamics simulations of aluminum at high temperature.

  19. A Multi-Modality CMOS Sensor Array for Cell-Based Assay and Drug Screening.

    PubMed

    Chi, Taiyun; Park, Jong Seok; Butts, Jessica C; Hookway, Tracy A; Su, Amy; Zhu, Chengjie; Styczynski, Mark P; McDevitt, Todd C; Wang, Hua

    2015-12-01

    In this paper, we present a fully integrated multi-modality CMOS cellular sensor array with four sensing modalities to characterize different cell physiological responses, including extracellular voltage recording, cellular impedance mapping, optical detection with shadow imaging and bioluminescence sensing, and thermal monitoring. The sensor array consists of nine parallel pixel groups and nine corresponding signal conditioning blocks. Each pixel group comprises one temperature sensor and 16 tri-modality sensor pixels, while each tri-modality sensor pixel can be independently configured for extracellular voltage recording, cellular impedance measurement (voltage excitation/current sensing), and optical detection. This sensor array supports multi-modality cellular sensing at the pixel level, which enables holistic cell characterization and joint-modality physiological monitoring on the same cellular sample with a pixel resolution of 80 μm × 100 μm. Comprehensive biological experiments with different living cell samples demonstrate the functionality and benefit of the proposed multi-modality sensing in cell-based assay and drug screening.

  20. Psychometrics of an internalized homophobia instrument for men.

    PubMed

    Theodore, John L; Shidlo, Ariel; Zemon, Vance; Foley, Frederick W; Dorfman, David; Dahlman, Karen L; Hamid, Sahira

    2013-01-01

    The Multi-Axial Gay Men's Inventory-Men's Short Version (MAGI-MSV) assesses internalized homophobia via 20 items and 3 dimensions. This study extended the psychometric examination of the MAGI-MSV. The instrument was administered to 228 ethnically diverse HIV-negative gay men seeking counseling in New York City (mean age = 35, age range = 16-70). Following principal axis factoring and parallel analyses, 4 factors emerged and 14 items were retained. The descriptive labels for factors included gay self-assurance and worth, public appearance of homosexuality, and impact of HIV/AIDS on homosexuality. The new, fourth factor was named maladaptive measures to eliminate homosexuality.

  1. No-insulation multi-width winding technique for high temperature superconducting magnet

    PubMed Central

    Hahn, Seungyong; Kim, Youngjae; Keun Park, Dong; Kim, Kwangmin; Voccio, John P.; Bascuñán, Juan; Iwasa, Yukikazu

    2013-01-01

    We present a No-Insulation (NI) Multi-Width (MW) winding technique for an HTS (high temperature superconductor) magnet consisting of double-pancake (DP) coils. The NI enables an HTS magnet self-protecting and the MW minimizes the detrimental anisotropy in current-carrying capacity of HTS tape by assigning tapes of multiple widths to DP coils within a stack, widest tape to the top and bottom sections and the narrowest in the midplane section. This paper presents fabrication and test results of an NI-MW HTS magnet and demonstrates the unique features of the NI-MW technique: self-protecting and enhanced field performance, unattainable with the conventional technique. PMID:24255549

  2. The effect of a multi-component smoking cessation intervention in African American women residing in public housing.

    PubMed

    Andrews, Jeannette O; Felton, Gwen; Ellen Wewers, Mary; Waller, Jennifer; Tingen, Martha

    2007-02-01

    The purpose of this study was to test the effectiveness of a multi-component smoking cessation intervention in African American women residing in public housing. The intervention consisted of: (a) nurse led behavioral/empowerment counseling; (b) nicotine replacement therapy; and, (c) community health workers to enhance smoking self-efficacy, social support, and spiritual well-being. The results showed a 6-month continuous smoking abstinence of 27.5% and 5.7% in the intervention and comparison groups. Changes in social support and smoking self-efficacy over time predicted smoking abstinence, and self-efficacy mediated 6-month smoking abstinence outcomes. Spiritual well-being did not predict or mediate smoking abstinence outcomes. These findings support the use of a nurse/community health worker model to deliver culturally tailored behavioral interventions with marginalized communities.

  3. Multi-Fluid Moment Simulations of Ganymede using the Next-Generation OpenGGCM

    NASA Astrophysics Data System (ADS)

    Wang, L.; Germaschewski, K.; Hakim, A.; Bhattacharjee, A.; Raeder, J.

    2015-12-01

    We coupled the multi-fluid moment code Gkeyll[1,2] to the next-generation OpenGGCM[3], and studied the reconnection dynamics at the Ganymede. This work is part of our effort to tackle the grand challenge of integrating kinetic effects into global fluid models. The multi-fluid moment model integrates kinetic effects in that it can capture crucial kinetic physics like pressure tensor effects by evolving moments of the Vlasov equations for each species. This approach has advantages over previous models: desired kinetic effects, together with other important effects like the Hall effect, are self-consistently embedded in the moment equations, and can be efficiently implemented, while not suffering from severe time-step restriction due to plasma oscillation nor artificial whistler modes. This model also handles multiple ion species naturally, which opens up opportunties in investigating the role of oxygen in magnetospheric reconnection and improved coupling to ionosphere models. In this work, the multi-fluid moment solver in Gkeyll was wrapped as a time-stepping module for the high performance, highly flexible next-generation OpenGGCM. Gkeyll is only used to provide the local plasma solver, while computational aspects like parallelization and boundary conditions are handled entirely by OpenGGCM, including interfacing to other models like ionospheric boundary conditions provided by coupling with CTIM [3]. The coupled code is used to study the dynamics near Ganymede, and the results are compared with MHD and Hall MHD results by Dorelli et al. [4]. Hakim, A. (2008). Journal of Fusion Energy, 27, 36-43. Hakim, A., Loverich, J., & Shumlak, U. (2006). Journal of Computational Physics, 219, 418-442. Raeder, J., Larson, D., Li, W., Kepko, E. L., & Fuller-Rowell, T. (2008). Space Science Reviews, 141(1-4), 535-555. Dorelli, J. C., Glocer, A., Collinson, G., & Tóth, G. (2015). Journal of Geophysical Research: Space Physics, 120.

  4. Toroidal Ampere-Faraday Equations Solved Consistently with the CQL3D Fokker-Planck Time-Evolution

    NASA Astrophysics Data System (ADS)

    Harvey, R. W.; Petrov, Yu. V.

    2013-10-01

    A self-consistent, time-dependent toroidal electric field calculation is a key feature of a complete 3D Fokker-Planck kinetic distribution radial transport code for f(v,theta,rho,t). In the present CQL3D finite-difference model, the electric field E(rho,t) is either prescribed, or iteratively adjusted to obtain prescribed toroidal or parallel currents. We discuss first results of an implementation of the Ampere-Faraday equation for the self-consistent toroidal electric field, as applied to the runaway electron production in tokamaks due to rapid reduction of the plasma temperature as occurs in a plasma disruption. Our previous results assuming a constant current density (Lenz' Law) model showed that prompt ``hot-tail runaways'' dominated ``knock-on'' and Dreicer ``drizzle'' runaways; we will examine modifications due to the more complete Ampere-Faraday solution. Work supported by US DOE under DE-FG02-ER54744.

  5. Are exergames promoting mobility an attractive alternative to conventional self-regulated exercises for elderly people in a rehabilitation setting? Study protocol of a randomized controlled trial.

    PubMed

    Hasselmann, Viviane; Oesch, Peter; Fernandez-Luque, Luis; Bachmann, Stefan

    2015-09-07

    Maintaining mobility in elderly persons has become a primary goal within healthcare services. In older adults, exercise programs significantly reduce the risk of falling and death. Long-lasting and high-intensive multi-component exercises are most effective. In a rehabilitation setting, self-regulated exercises are conventionally taught by physiotherapists, using handouts. However, the adherence of elderly persons to executing these self-administered programs varies considerably. They are often considered tedious and boring, and thus prematurely stopped. The primary aim of this clinical trial is to determine whether elderly persons in a rehabilitation setting show higher adherence to self-regulated training when using exergames than when performing conventional exercises. The second objective is to explore which mode of exercise leads to greater improvement in balance performance. The study consists of a single blind, stratified, randomized control trial with two parallel groups. Once included, study participants will be stratified according to their balance and computer skills and randomly allocated to self-regulated training with conventional exercise programs or with exergames played with the Windows Kinect® sensor and FitBit® pedometer. In both groups, self-administered exercise programs will be taught by experienced physiotherapists and performed at the patient's own discretion during the ten days of intervention. The primary outcome is the performed daily training volume, collected by the participants in a logbook. Secondary outcomes are objective and subjective balance skills measured by an activity tracker and the Fall Efficacy Scale self-administered questionnaire. Both assessments will be performed at pre- and post-intervention. According to the available literature, this study is the first to compare conventional self-regulated exercises with exergames among older patients in a rehabilitation setting. Results of this study will contribute to our understanding of its motivational potential on exercise adherence in elderly persons and provide more insight into the potential effectiveness of exergames promoting mobility. The present clinical study has been registered on ClinicalTrials.gov under the identifier number: NCT02077049. The detailed trial protocol can be accessed online on: NCT02077049.

  6. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kocharovsky, V. V., E-mail: vkochar@physics.tamu.edu; Department of Physics and Astronomy, Texas A&M University, College Station, Texas 77843-4242; Kocharovsky, VI. V.

    Widespread use of a broken-power-law description of the spectra of synchrotron emission of various plasma objects requires an analysis of origin and a proper interpretation of spectral components. We show that, for a self-consistent magnetic configuration in a collisionless plasma, these components may be angle-dependent according to an anisotropic particle momentum distribution and may have no counterparts in a particle energy distribution. That has never been studied analytically and is in contrast to a usual model of synchrotron radiation, assuming an external magnetic field and a particle ensemble with isotropic momentum distribution. We demonstrate that for the wide intervals ofmore » observation angle the power-law spectra and, in particular, the positions and number of spectral breaks may be essentially different for the cases of the self-consistent and not-self-consistent magnetic fields in current structures responsible for the synchrotron radiation of the ensembles of relativistic particles with the multi-power-law energy distributions.« less

  7. The Design and Evaluation of "CAPTools"--A Computer Aided Parallelization Toolkit

    NASA Technical Reports Server (NTRS)

    Yan, Jerry; Frumkin, Michael; Hribar, Michelle; Jin, Haoqiang; Waheed, Abdul; Johnson, Steve; Cross, Jark; Evans, Emyr; Ierotheou, Constantinos; Leggett, Pete; hide

    1998-01-01

    Writing applications for high performance computers is a challenging task. Although writing code by hand still offers the best performance, it is extremely costly and often not very portable. The Computer Aided Parallelization Tools (CAPTools) are a toolkit designed to help automate the mapping of sequential FORTRAN scientific applications onto multiprocessors. CAPTools consists of the following major components: an inter-procedural dependence analysis module that incorporates user knowledge; a 'self-propagating' data partitioning module driven via user guidance; an execution control mask generation and optimization module for the user to fine tune parallel processing of individual partitions; a program transformation/restructuring facility for source code clean up and optimization; a set of browsers through which the user interacts with CAPTools at each stage of the parallelization process; and a code generator supporting multiple programming paradigms on various multiprocessors. Besides describing the rationale behind the architecture of CAPTools, the parallelization process is illustrated via case studies involving structured and unstructured meshes. The programming process and the performance of the generated parallel programs are compared against other programming alternatives based on the NAS Parallel Benchmarks, ARC3D and other scientific applications. Based on these results, a discussion on the feasibility of constructing architectural independent parallel applications is presented.

  8. Simple and robust generation of ultrafast laser pulse trains using polarization-independent parallel-aligned thin films

    NASA Astrophysics Data System (ADS)

    Wang, Andong; Jiang, Lan; Li, Xiaowei; Wang, Zhi; Du, Kun; Lu, Yongfeng

    2018-05-01

    Ultrafast laser pulse temporal shaping has been widely applied in various important applications such as laser materials processing, coherent control of chemical reactions, and ultrafast imaging. However, temporal pulse shaping has been limited to only-in-lab technique due to the high cost, low damage threshold, and polarization dependence. Herein we propose a novel design of ultrafast laser pulse train generation device, which consists of multiple polarization-independent parallel-aligned thin films. Various pulse trains with controllable temporal profile can be generated flexibly by multi-reflections within the splitting films. Compared with other pulse train generation techniques, this method has advantages of compact structure, low cost, high damage threshold and polarization independence. These advantages endow it with high potential for broad utilization in ultrafast applications.

  9. Adding dynamic rules to self-organizing fuzzy systems

    NASA Technical Reports Server (NTRS)

    Buhusi, Catalin V.

    1992-01-01

    This paper develops a Dynamic Self-Organizing Fuzzy System (DSOFS) capable of adding, removing, and/or adapting the fuzzy rules and the fuzzy reference sets. The DSOFS background consists of a self-organizing neural structure with neuron relocation features which will develop a map of the input-output behavior. The relocation algorithm extends the topological ordering concept. Fuzzy rules (neurons) are dynamically added or released while the neural structure learns the pattern. The DSOFS advantages are the automatic synthesis and the possibility of parallel implementation. A high adaptation speed and a reduced number of neurons is needed in order to keep errors under some limits. The computer simulation results are presented in a nonlinear systems modelling application.

  10. Toward a consistent modeling framework to assess multi-sectoral climate impacts.

    PubMed

    Monier, Erwan; Paltsev, Sergey; Sokolov, Andrei; Chen, Y-H Henry; Gao, Xiang; Ejaz, Qudsia; Couzo, Evan; Schlosser, C Adam; Dutkiewicz, Stephanie; Fant, Charles; Scott, Jeffery; Kicklighter, David; Morris, Jennifer; Jacoby, Henry; Prinn, Ronald; Haigh, Martin

    2018-02-13

    Efforts to estimate the physical and economic impacts of future climate change face substantial challenges. To enrich the currently popular approaches to impact analysis-which involve evaluation of a damage function or multi-model comparisons based on a limited number of standardized scenarios-we propose integrating a geospatially resolved physical representation of impacts into a coupled human-Earth system modeling framework. Large internationally coordinated exercises cannot easily respond to new policy targets and the implementation of standard scenarios across models, institutions and research communities can yield inconsistent estimates. Here, we argue for a shift toward the use of a self-consistent integrated modeling framework to assess climate impacts, and discuss ways the integrated assessment modeling community can move in this direction. We then demonstrate the capabilities of such a modeling framework by conducting a multi-sectoral assessment of climate impacts under a range of consistent and integrated economic and climate scenarios that are responsive to new policies and business expectations.

  11. Anisotropic magnetotail equilibrium and convection

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hau, L.N.

    This paper reports on self-consistent two-dimensional equilibria with anisotropic plasma pressure for the Earth's magnetotail. These configurations are obtained by numerically solving the generalized Grad-Shafranov equation, describing anisotropic plasmas with p[parallel] [ne] p[perpendicular], including the Earth's dipolar field. Consistency between these new equilibria and the assumption of steady-state, sunward convection, described by the double-adiabatic laws, is examined. As for the case of isotropic pressure [Erickson and Wolf, 1980], there exists a discrepancy between typical quite-time magnetic field models and the assumption of steady-state double-adiabatic lossless plasma sheet convection. However, unlike that case, this inconsistency cannot be removed by the presencemore » of a weak equatorial normal magnetic field strength in the near tail region: magnetic field configurations of this type produce unreasonably large pressure anisotropies, p[parallel] > p[perpendicular], in the plasma sheet. 16 refs., 5 figs.« less

  12. Parallelized multi-graphics processing unit framework for high-speed Gabor-domain optical coherence microscopy.

    PubMed

    Tankam, Patrice; Santhanam, Anand P; Lee, Kye-Sung; Won, Jungeun; Canavesi, Cristina; Rolland, Jannick P

    2014-07-01

    Gabor-domain optical coherence microscopy (GD-OCM) is a volumetric high-resolution technique capable of acquiring three-dimensional (3-D) skin images with histological resolution. Real-time image processing is needed to enable GD-OCM imaging in a clinical setting. We present a parallelized and scalable multi-graphics processing unit (GPU) computing framework for real-time GD-OCM image processing. A parallelized control mechanism was developed to individually assign computation tasks to each of the GPUs. For each GPU, the optimal number of amplitude-scans (A-scans) to be processed in parallel was selected to maximize GPU memory usage and core throughput. We investigated five computing architectures for computational speed-up in processing 1000×1000 A-scans. The proposed parallelized multi-GPU computing framework enables processing at a computational speed faster than the GD-OCM image acquisition, thereby facilitating high-speed GD-OCM imaging in a clinical setting. Using two parallelized GPUs, the image processing of a 1×1×0.6  mm3 skin sample was performed in about 13 s, and the performance was benchmarked at 6.5 s with four GPUs. This work thus demonstrates that 3-D GD-OCM data may be displayed in real-time to the examiner using parallelized GPU processing.

  13. Decoupling Principle Analysis and Development of a Parallel Three-Dimensional Force Sensor

    PubMed Central

    Zhao, Yanzhi; Jiao, Leihao; Weng, Dacheng; Zhang, Dan; Zheng, Rencheng

    2016-01-01

    In the development of the multi-dimensional force sensor, dimension coupling is the ubiquitous factor restricting the improvement of the measurement accuracy. To effectively reduce the influence of dimension coupling on the parallel multi-dimensional force sensor, a novel parallel three-dimensional force sensor is proposed using a mechanical decoupling principle, and the influence of the friction on dimension coupling is effectively reduced by making the friction rolling instead of sliding friction. In this paper, the mathematical model is established by combining with the structure model of the parallel three-dimensional force sensor, and the modeling and analysis of mechanical decoupling are carried out. The coupling degree (ε) of the designed sensor is defined and calculated, and the calculation results show that the mechanical decoupling parallel structure of the sensor possesses good decoupling performance. A prototype of the parallel three-dimensional force sensor was developed, and FEM analysis was carried out. The load calibration and data acquisition experiment system are built, and then calibration experiments were done. According to the calibration experiments, the measurement accuracy is less than 2.86% and the coupling accuracy is less than 3.02%. The experimental results show that the sensor system possesses high measuring accuracy, which provides a basis for the applied research of the parallel multi-dimensional force sensor. PMID:27649194

  14. Comparison of Self-Report Versus Sensor-Based Methods for Measuring the Amount of Upper Limb Activity Outside the Clinic.

    PubMed

    Waddell, Kimberly J; Lang, Catherine E

    2018-03-10

    To compare self-reported with sensor-measured upper limb (UL) performance in daily life for individuals with chronic (≥6mo) UL paresis poststroke. Secondary analysis of participants enrolled in a phase II randomized, parallel, dose-response UL movement trial. This analysis compared the accuracy and consistency between self-reported UL performance and sensor-measured UL performance at baseline and immediately post an 8-week intensive UL task-specific intervention. Outpatient rehabilitation. Community-dwelling individuals with chronic (≥6mo) UL paresis poststroke (N=64). Not applicable. Motor Activity Log amount of use scale and the sensor-derived use ratio from wrist-worn accelerometers. There was a high degree of variability between self-reported UL performance and the sensor-derived use ratio. Using sensor-based values as a reference, 3 distinct categories were identified: accurate reporters (reporting difference ±0.1), overreporters (difference >0.1), and underreporters (difference <-0.1). Five of 64 participants accurately self-reported UL performance at baseline and postintervention. Over half of participants (52%) switched categories from pre-to postintervention (eg, moved from underreporting preintervention to overreporting postintervention). For the consistent reporters, no participant characteristics were found to influence whether someone over- or underreported performance compared with sensor-based assessment. Participants did not consistently or accurately self-report UL performance when compared with the sensor-derived use ratio. Although self-report and sensor-based assessments are moderately associated and appear similar conceptually, these results suggest self-reported UL performance is often not consistent with sensor-measured performance and the measures cannot be used interchangeably. Copyright © 2018 American Congress of Rehabilitation Medicine. Published by Elsevier Inc. All rights reserved.

  15. Smart Hand For Manipulators

    NASA Astrophysics Data System (ADS)

    Fiorini, Paolo

    1987-10-01

    Sensor based, computer controlled end effectors for mechanical arms are receiving more and more attention in the robotics industry, because commonly available grippers are only adequate for simple pick and place tasks. This paper describes the current status of the research at JPL on a smart hand for a Puma 560 robot arm. The hand is a self contained, autonomous system, capable of executing high level commands from a supervisory computer. The mechanism consists of parallel fingers, powered by a DC motor, and controlled by a microprocessor embedded in the hand housing. Special sensors are integrated in the hand for measuring the grasp force of the fingers, and for measuring forces and torques applied between the arm and the surrounding environment. Fingers can be exercised under position, velocity and force control modes. The single-chip microcomputer in the hand executes the tasks of communication, data acquisition and sensor based motor control, with a sample cycle of 2 ms and a transmission rate of 9600 baud. The smart hand described in this paper represents a new development in the area of end effector design because of its multi-functionality and autonomy. It will also be a versatile test bed for experimenting with advanced control schemes for dexterous manipulation.

  16. Genetic Parallel Programming: design and implementation.

    PubMed

    Cheang, Sin Man; Leung, Kwong Sak; Lee, Kin Hong

    2006-01-01

    This paper presents a novel Genetic Parallel Programming (GPP) paradigm for evolving parallel programs running on a Multi-Arithmetic-Logic-Unit (Multi-ALU) Processor (MAP). The MAP is a Multiple Instruction-streams, Multiple Data-streams (MIMD), general-purpose register machine that can be implemented on modern Very Large-Scale Integrated Circuits (VLSIs) in order to evaluate genetic programs at high speed. For human programmers, writing parallel programs is more difficult than writing sequential programs. However, experimental results show that GPP evolves parallel programs with less computational effort than that of their sequential counterparts. It creates a new approach to evolving a feasible problem solution in parallel program form and then serializes it into a sequential program if required. The effectiveness and efficiency of GPP are investigated using a suite of 14 well-studied benchmark problems. Experimental results show that GPP speeds up evolution substantially.

  17. Multi-charge-state molecular dynamics and self-diffusion coefficient in the warm dense matter regime

    NASA Astrophysics Data System (ADS)

    Fu, Yongsheng; Hou, Yong; Kang, Dongdong; Gao, Cheng; Jin, Fengtao; Yuan, Jianmin

    2018-01-01

    We present a multi-ion molecular dynamics (MIMD) simulation and apply it to calculating the self-diffusion coefficients of ions with different charge-states in the warm dense matter (WDM) regime. First, the method is used for the self-consistent calculation of electron structures of different charge-state ions in the ion sphere, with the ion-sphere radii being determined by the plasma density and the ion charges. The ionic fraction is then obtained by solving the Saha equation, taking account of interactions among different charge-state ions in the system, and ion-ion pair potentials are computed using the modified Gordon-Kim method in the framework of temperature-dependent density functional theory on the basis of the electron structures. Finally, MIMD is used to calculate ionic self-diffusion coefficients from the velocity correlation function according to the Green-Kubo relation. A comparison with the results of the average-atom model shows that different statistical processes will influence the ionic diffusion coefficient in the WDM regime.

  18. A general parallel sparse-blocked matrix multiply for linear scaling SCF theory

    NASA Astrophysics Data System (ADS)

    Challacombe, Matt

    2000-06-01

    A general approach to the parallel sparse-blocked matrix-matrix multiply is developed in the context of linear scaling self-consistent-field (SCF) theory. The data-parallel message passing method uses non-blocking communication to overlap computation and communication. The space filling curve heuristic is used to achieve data locality for sparse matrix elements that decay with “separation”. Load balance is achieved by solving the bin packing problem for blocks with variable size.With this new method as the kernel, parallel performance of the simplified density matrix minimization (SDMM) for solution of the SCF equations is investigated for RHF/6-31G ∗∗ water clusters and RHF/3-21G estane globules. Sustained rates above 5.7 GFLOPS for the SDMM have been achieved for (H 2 O) 200 with 95 Origin 2000 processors. Scalability is found to be limited by load imbalance, which increases with decreasing granularity, due primarily to the inhomogeneous distribution of variable block sizes.

  19. Pushing configuration-interaction to the limit: Towards massively parallel MCSCF calculations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Vogiatzis, Konstantinos D.; Ma, Dongxia; Olsen, Jeppe

    A new large-scale parallel multiconfigurational self-consistent field (MCSCF) implementation in the open-source NWChem computational chemistry code is presented. The generalized active space approach is used to partition large configuration interaction (CI) vectors and generate a sufficient number of batches that can be distributed to the available cores. Massively parallel CI calculations with large active spaces can be performed. The new parallel MCSCF implementation is tested for the chromium trimer and for an active space of 20 electrons in 20 orbitals, which can now routinely be performed. Unprecedented CI calculations with an active space of 22 electrons in 22 orbitals formore » the pentacene systems were performed and a single CI iteration calculation with an active space of 24 electrons in 24 orbitals for the chromium tetramer was possible. In conclusion, the chromium tetramer corresponds to a CI expansion of one trillion Slater determinants (914 058 513 424) and is the largest conventional CI calculation attempted up to date.« less

  20. Full Parallel Implementation of an All-Electron Four-Component Dirac-Kohn-Sham Program.

    PubMed

    Rampino, Sergio; Belpassi, Leonardo; Tarantelli, Francesco; Storchi, Loriano

    2014-09-09

    A full distributed-memory implementation of the Dirac-Kohn-Sham (DKS) module of the program BERTHA (Belpassi et al., Phys. Chem. Chem. Phys. 2011, 13, 12368-12394) is presented, where the self-consistent field (SCF) procedure is replicated on all the parallel processes, each process working on subsets of the global matrices. The key feature of the implementation is an efficient procedure for switching between two matrix distribution schemes, one (integral-driven) optimal for the parallel computation of the matrix elements and another (block-cyclic) optimal for the parallel linear algebra operations. This approach, making both CPU-time and memory scalable with the number of processors used, virtually overcomes at once both time and memory barriers associated with DKS calculations. Performance, portability, and numerical stability of the code are illustrated on the basis of test calculations on three gold clusters of increasing size, an organometallic compound, and a perovskite model. The calculations are performed on a Beowulf and a BlueGene/Q system.

  1. Pushing configuration-interaction to the limit: Towards massively parallel MCSCF calculations

    DOE PAGES

    Vogiatzis, Konstantinos D.; Ma, Dongxia; Olsen, Jeppe; ...

    2017-11-14

    A new large-scale parallel multiconfigurational self-consistent field (MCSCF) implementation in the open-source NWChem computational chemistry code is presented. The generalized active space approach is used to partition large configuration interaction (CI) vectors and generate a sufficient number of batches that can be distributed to the available cores. Massively parallel CI calculations with large active spaces can be performed. The new parallel MCSCF implementation is tested for the chromium trimer and for an active space of 20 electrons in 20 orbitals, which can now routinely be performed. Unprecedented CI calculations with an active space of 22 electrons in 22 orbitals formore » the pentacene systems were performed and a single CI iteration calculation with an active space of 24 electrons in 24 orbitals for the chromium tetramer was possible. In conclusion, the chromium tetramer corresponds to a CI expansion of one trillion Slater determinants (914 058 513 424) and is the largest conventional CI calculation attempted up to date.« less

  2. Pushing configuration-interaction to the limit: Towards massively parallel MCSCF calculations

    NASA Astrophysics Data System (ADS)

    Vogiatzis, Konstantinos D.; Ma, Dongxia; Olsen, Jeppe; Gagliardi, Laura; de Jong, Wibe A.

    2017-11-01

    A new large-scale parallel multiconfigurational self-consistent field (MCSCF) implementation in the open-source NWChem computational chemistry code is presented. The generalized active space approach is used to partition large configuration interaction (CI) vectors and generate a sufficient number of batches that can be distributed to the available cores. Massively parallel CI calculations with large active spaces can be performed. The new parallel MCSCF implementation is tested for the chromium trimer and for an active space of 20 electrons in 20 orbitals, which can now routinely be performed. Unprecedented CI calculations with an active space of 22 electrons in 22 orbitals for the pentacene systems were performed and a single CI iteration calculation with an active space of 24 electrons in 24 orbitals for the chromium tetramer was possible. The chromium tetramer corresponds to a CI expansion of one trillion Slater determinants (914 058 513 424) and is the largest conventional CI calculation attempted up to date.

  3. Multi-species ion transport in ICF relevant conditions

    NASA Astrophysics Data System (ADS)

    Vold, Erik; Kagan, Grigory; Simakov, Andrei; Molvig, Kim; Yin, Lin; Albright, Brian

    2017-10-01

    Classical transport theory based on Chapman-Enskog methods provides self consistent approximations for kinetic fluxes of mass, heat and momentum for each ion species in a multi-ion plasma characterized with a small Knudsen number. A numerical method for solving the classic forms of multi-ion transport, self-consistently including heat and species mass fluxes relative to the center of mass, is given in [Kagan-Baalrud, arXiv '16] and similar transport coefficients result from recent derivations [Simakov-Molvig, PoP, '16]. We have implemented a combination of these methods in a standalone test code and in xRage, an adaptive-mesh radiation hydrodynamics code, at LANL. Transport mixing is examined between a DT fuel and a CH capsule shell in ICF conditions. The four ion species develop individual self-similar density profiles under the assumption of P-T equilibrium in 1D and show interesting early time transient pressure and center of mass velocity behavior when P-T equilibrium is not enforced. Some 2D results are explored to better understand the transport mix in combination with convective flow driven by macroscopic fluid instabilities at the fuel-capsule interface. Early transient and some 2D behaviors from the fluid transport are compared to kinetic code results. Work performed under the auspices of the U.S. DOE by the LANS, LLC, Los Alamos National Laboratory under Contract No. DE-AC52-06NA25396. Funding provided by the Advanced Simulation and Computing (ASC) Program.

  4. Review of Data Integrity Models in Multi-Level Security Environments

    DTIC Science & Technology

    2011-02-01

    2: (E-1 extension) Only executions described in a (User, TP, (CDIs)) relation are allowed • E-3: Users must be authenticated before allowing TP... authentication and verification procedures for upgrading the integrity of certain objects. The mechanism used to manage access to objects is primarily...that is, the self-consistency of interdependent data and the consistency of real-world environment data. The prevention of authorised users from making

  5. Data Parallel Bin-Based Indexing for Answering Queries on Multi-Core Architectures

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gosink, Luke; Wu, Kesheng; Bethel, E. Wes

    2009-06-02

    The multi-core trend in CPUs and general purpose graphics processing units (GPUs) offers new opportunities for the database community. The increase of cores at exponential rates is likely to affect virtually every server and client in the coming decade, and presents database management systems with a huge, compelling disruption that will radically change how processing is done. This paper presents a new parallel indexing data structure for answering queries that takes full advantage of the increasing thread-level parallelism emerging in multi-core architectures. In our approach, our Data Parallel Bin-based Index Strategy (DP-BIS) first bins the base data, and then partitionsmore » and stores the values in each bin as a separate, bin-based data cluster. In answering a query, the procedures for examining the bin numbers and the bin-based data clusters offer the maximum possible level of concurrency; each record is evaluated by a single thread and all threads are processed simultaneously in parallel. We implement and demonstrate the effectiveness of DP-BIS on two multi-core architectures: a multi-core CPU and a GPU. The concurrency afforded by DP-BIS allows us to fully utilize the thread-level parallelism provided by each architecture--for example, our GPU-based DP-BIS implementation simultaneously evaluates over 12,000 records with an equivalent number of concurrently executing threads. In comparing DP-BIS's performance across these architectures, we show that the GPU-based DP-BIS implementation requires significantly less computation time to answer a query than the CPU-based implementation. We also demonstrate in our analysis that DP-BIS provides better overall performance than the commonly utilized CPU and GPU-based projection index. Finally, due to data encoding, we show that DP-BIS accesses significantly smaller amounts of data than index strategies that operate solely on a column's base data; this smaller data footprint is critical for parallel processors that possess limited memory resources (e.g., GPUs).« less

  6. Mindfulness based cognitive therapy versus treatment as usual in adults with attention deficit hyperactivity disorder (ADHD).

    PubMed

    Janssen, Lotte; Kan, Cornelis C; Carpentier, Pieter J; Sizoo, Bram; Hepark, Sevket; Grutters, Janneke; Donders, Rogier; Buitelaar, Jan K; Speckens, Anne E M

    2015-09-15

    Adults with attention deficit hyperactivity disorder (ADHD) often present with a lifelong pattern of core symptoms that is associated with impairments of functioning in daily life. This has a substantial personal and economic impact. In clinical practice there is a high need for additional or alternative interventions for existing treatments, usually consisting of pharmacotherapy and/or psycho-education. Although previous studies show preliminary evidence for the effectiveness of mindfulness-based interventions in reducing ADHD symptoms and improving executive functioning, these studies have methodological limitations. This study will take account of these limitations and will examine the effectiveness of Mindfulness Based Cognitive Therapy (MBCT) in further detail. A multi-centre, parallel-group, randomised controlled trial will be conducted in N = 120 adults with ADHD. Patients will be randomised to MBCT in addition to treatment as usual (TAU) or TAU alone. Assessments will take place at baseline and at three, six and nine months after baseline. Primary outcome measure will be severity of ADHD symptoms rated by a blinded clinician. Secondary outcome measures will be self-reported ADHD symptoms, executive functioning, mindfulness skills, self-compassion, positive mental health and general functioning. In addition, a cost-effectiveness analysis will be conducted. This trial will offer valuable information about the clinical and cost-effectiveness of MBCT in addition to TAU compared to TAU alone in adults swith ADHD. ClinicalTrials.gov NCT02463396. Registered 8 June 2015.

  7. A novel constant-force scanning probe incorporating mechanical-magnetic coupled structures.

    PubMed

    Wang, Hongxi; Zhao, Jian; Gao, Renjing; Yang, Yintang

    2011-07-01

    A one-dimensional scanning probe with constant measuring force is designed and fabricated by utilizing the negative stiffness of the magnetic coupled structure, which mainly consists of the magnetic structure, the parallel guidance mechanism, and the pre-stressed spring. Based on the theory of material mechanics and the equivalent surface current model for computing the magnetic force, the analytical model of the scanning probe subjected to multi-forces is established, and the nonlinear relationship between the measuring force and the probe displacement is obtained. The practicability of introducing magnetic coupled structure in the constant-force probe is validated by the consistency of the results in numerical simulation and experiments.

  8. A highly efficient multi-core algorithm for clustering extremely large datasets

    PubMed Central

    2010-01-01

    Background In recent years, the demand for computational power in computational biology has increased due to rapidly growing data sets from microarray and other high-throughput technologies. This demand is likely to increase. Standard algorithms for analyzing data, such as cluster algorithms, need to be parallelized for fast processing. Unfortunately, most approaches for parallelizing algorithms largely rely on network communication protocols connecting and requiring multiple computers. One answer to this problem is to utilize the intrinsic capabilities in current multi-core hardware to distribute the tasks among the different cores of one computer. Results We introduce a multi-core parallelization of the k-means and k-modes cluster algorithms based on the design principles of transactional memory for clustering gene expression microarray type data and categorial SNP data. Our new shared memory parallel algorithms show to be highly efficient. We demonstrate their computational power and show their utility in cluster stability and sensitivity analysis employing repeated runs with slightly changed parameters. Computation speed of our Java based algorithm was increased by a factor of 10 for large data sets while preserving computational accuracy compared to single-core implementations and a recently published network based parallelization. Conclusions Most desktop computers and even notebooks provide at least dual-core processors. Our multi-core algorithms show that using modern algorithmic concepts, parallelization makes it possible to perform even such laborious tasks as cluster sensitivity and cluster number estimation on the laboratory computer. PMID:20370922

  9. MultiBLUP: improved SNP-based prediction for complex traits.

    PubMed

    Speed, Doug; Balding, David J

    2014-09-01

    BLUP (best linear unbiased prediction) is widely used to predict complex traits in plant and animal breeding, and increasingly in human genetics. The BLUP mathematical model, which consists of a single random effect term, was adequate when kinships were measured from pedigrees. However, when genome-wide SNPs are used to measure kinships, the BLUP model implicitly assumes that all SNPs have the same effect-size distribution, which is a severe and unnecessary limitation. We propose MultiBLUP, which extends the BLUP model to include multiple random effects, allowing greatly improved prediction when the random effects correspond to classes of SNPs with distinct effect-size variances. The SNP classes can be specified in advance, for example, based on SNP functional annotations, and we also provide an adaptive procedure for determining a suitable partition of SNPs. We apply MultiBLUP to genome-wide association data from the Wellcome Trust Case Control Consortium (seven diseases), and from much larger studies of celiac disease and inflammatory bowel disease, finding that it consistently provides better prediction than alternative methods. Moreover, MultiBLUP is computationally very efficient; for the largest data set, which includes 12,678 individuals and 1.5 M SNPs, the total analysis can be run on a single desktop PC in less than a day and can be parallelized to run even faster. Tools to perform MultiBLUP are freely available in our software LDAK. © 2014 Speed and Balding; Published by Cold Spring Harbor Laboratory Press.

  10. All-optical regenerator of multi-channel signals.

    PubMed

    Li, Lu; Patki, Pallavi G; Kwon, Young B; Stelmakh, Veronika; Campbell, Brandon D; Annamalai, Muthiah; Lakoba, Taras I; Vasilyev, Michael

    2017-10-12

    One of the main reasons why nonlinear-optical signal processing (regeneration, logic, etc.) has not yet become a practical alternative to electronic processing is that the all-optical elements with nonlinear input-output relationship have remained inherently single-channel devices (just like their electronic counterparts) and, hence, cannot fully utilise the parallel processing potential of optical fibres and amplifiers. The nonlinear input-output transfer function requires strong optical nonlinearity, e.g. self-phase modulation, which, for fundamental reasons, is always accompanied by cross-phase modulation and four-wave mixing. In processing multiple wavelength-division-multiplexing channels, large cross-phase modulation and four-wave mixing crosstalks among the channels destroy signal quality. Here we describe a solution to this problem: an optical signal processor employing a group-delay-managed nonlinear medium where strong self-phase modulation is achieved without such nonlinear crosstalk. We demonstrate, for the first time to our knowledge, simultaneous all-optical regeneration of up to 16 wavelength-division-multiplexing channels by one device. This multi-channel concept can be extended to other nonlinear-optical processing schemes.Nonlinear optical processing devices are not yet fully practical as they are single channel. Here the authors demonstrate all-optical regeneration of up to 16 channels by one device, employing a group-delay-managed nonlinear medium where strong self-phase modulation is achieved without nonlinear inter-channel crosstalk.

  11. MLP: A Parallel Programming Alternative to MPI for New Shared Memory Parallel Systems

    NASA Technical Reports Server (NTRS)

    Taft, James R.

    1999-01-01

    Recent developments at the NASA AMES Research Center's NAS Division have demonstrated that the new generation of NUMA based Symmetric Multi-Processing systems (SMPs), such as the Silicon Graphics Origin 2000, can successfully execute legacy vector oriented CFD production codes at sustained rates far exceeding processing rates possible on dedicated 16 CPU Cray C90 systems. This high level of performance is achieved via shared memory based Multi-Level Parallelism (MLP). This programming approach, developed at NAS and outlined below, is distinct from the message passing paradigm of MPI. It offers parallelism at both the fine and coarse grained level, with communication latencies that are approximately 50-100 times lower than typical MPI implementations on the same platform. Such latency reductions offer the promise of performance scaling to very large CPU counts. The method draws on, but is also distinct from, the newly defined OpenMP specification, which uses compiler directives to support a limited subset of multi-level parallel operations. The NAS MLP method is general, and applicable to a large class of NASA CFD codes.

  12. a Spatiotemporal Aggregation Query Method Using Multi-Thread Parallel Technique Based on Regional Division

    NASA Astrophysics Data System (ADS)

    Liao, S.; Chen, L.; Li, J.; Xiong, W.; Wu, Q.

    2015-07-01

    Existing spatiotemporal database supports spatiotemporal aggregation query over massive moving objects datasets. Due to the large amounts of data and single-thread processing method, the query speed cannot meet the application requirements. On the other hand, the query efficiency is more sensitive to spatial variation then temporal variation. In this paper, we proposed a spatiotemporal aggregation query method using multi-thread parallel technique based on regional divison and implemented it on the server. Concretely, we divided the spatiotemporal domain into several spatiotemporal cubes, computed spatiotemporal aggregation on all cubes using the technique of multi-thread parallel processing, and then integrated the query results. By testing and analyzing on the real datasets, this method has improved the query speed significantly.

  13. Molecular structure of self-assembled chiral nanoribbons and nanotubules revealed in the hydrated state.

    PubMed

    Oda, Reiko; Artzner, Franck; Laguerre, Michel; Huc, Ivan

    2008-11-05

    A detailed molecular organization of racemic 16-2-16 tartrate self-assembled multi-bilayer ribbons in the hydrated state is proposed where 16-2-16 amphiphiles, tartrate ions, and water molecules are all accurately positioned by comparing experimental X-ray powder diffraction and diffraction patterns derived from modeling studies. X-ray diffuse scattering studies show that molecular organization is not fundamentally altered when comparing the flat ribbons of the racemate to chirally twisted or helical ribbons of the pure tartrate enantiomer. Essential features of the three-dimensional molecular organizations of these structures include interdigitation of alkyl chains within each bilayer and well-defined networks of ionic and hydrogen bonds between cations, anions, and water molecules between bilayers. The detailed study of diffraction patterns also indicated that the gemini headgroups are oriented parallel to the long edge of the ribbons. The structure thus possesses a high cohesion and good crystallinity, and for the first time, we could relate the packing of the chiral molecules to the expression of the chirality at a mesoscopic scale. The organization of the ribbons at the molecular level sheds light on a number of their macroscopic features. Among these are the reason why enantiomerically pure 16-2-16 tartrate forms ribbons that consist of exactly two bilayers, and a plausible mechanism by which a chirally twisted or helical shape may emerge from the packing of chiral tartrate ions. Importantly, the distinction between commonly observed helical and twisted morphologies could be related to a subtle symmetry breaking. These results demonstrate that accurately solving the molecular structure of self-assembled soft materials--a process rarely achieved--is within reach, that it is a valid approach to correlate molecular parameters to macroscopic properties, and thus that it offers opportunities to modulate properties through molecular design.

  14. Robust High-Resolution Cloth Using Parallelism, History-Based Collisions and Accurate Friction

    PubMed Central

    Selle, Andrew; Su, Jonathan; Irving, Geoffrey; Fedkiw, Ronald

    2015-01-01

    In this paper we simulate high resolution cloth consisting of up to 2 million triangles which allows us to achieve highly detailed folds and wrinkles. Since the level of detail is also influenced by object collision and self collision, we propose a more accurate model for cloth-object friction. We also propose a robust history-based repulsion/collision framework where repulsions are treated accurately and efficiently on a per time step basis. Distributed memory parallelism is used for both time evolution and collisions and we specifically address Gauss-Seidel ordering of repulsion/collision response. This algorithm is demonstrated by several high-resolution and high-fidelity simulations. PMID:19147895

  15. Vacuum birefringence in strong magnetic fields: (II) Complex refractive index from the lowest Landau level

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hattori, Koichi, E-mail: khattori@yonsei.ac.kr; Itakura, Kazunori, E-mail: kazunori.itakura@kek.jp; Department of Particle and Nuclear Studies, Graduate University for Advanced Studies

    2013-07-15

    We compute the refractive indices of a photon propagating in strong magnetic fields on the basis of the analytic representation of the vacuum polarization tensor obtained in our previous paper. When the external magnetic field is strong enough for the fermion one-loop diagram of the polarization tensor to be approximated by the lowest Landau level, the propagating mode in parallel to the magnetic field is subject to modification: The refractive index deviates from unity and can be very large, and when the photon energy is large enough, the refractive index acquires an imaginary part indicating decay of a photon intomore » a fermion–antifermion pair. We study dependences of the refractive index on the propagating angle and the magnetic-field strength. It is also emphasized that a self-consistent treatment of the equation which defines the refractive index is indispensable for accurate description of the refractive index. This self-consistent treatment physically corresponds to consistently including the effects of back reactions of the distorted Dirac sea in response to the incident photon. -- Highlights: •Vacuum birefringence and photon decay are described by the complex refractive index. •Resummed photon vacuum polarization tensor in the lowest Landau level is used. •Back reactions from the distorted Dirac sea are self-consistently taken into account. •Self-consistent treatment drastically changes structure in photon energy dependence. •Dependences on photon propagation angle and magnetic-field strength are presented.« less

  16. Parallel and fault-tolerant algorithms for hypercube multiprocessors

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Aykanat, C.

    1988-01-01

    Several techniques for increasing the performance of parallel algorithms on distributed-memory message-passing multi-processor systems are investigated. These techniques are effectively implemented for the parallelization of the Scaled Conjugate Gradient (SCG) algorithm on a hypercube connected message-passing multi-processor. Significant performance improvement is achieved by using these techniques. The SCG algorithm is used for the solution phase of an FE modeling system. Almost linear speed-up is achieved, and it is shown that hypercube topology is scalable for an FE class of problem. The SCG algorithm is also shown to be suitable for vectorization, and near supercomputer performance is achieved on a vectormore » hypercube multiprocessor by exploiting both parallelization and vectorization. Fault-tolerance issues for the parallel SCG algorithm and for the hypercube topology are also addressed.« less

  17. Using OpenMP vs. Threading Building Blocks for Medical Imaging on Multi-cores

    NASA Astrophysics Data System (ADS)

    Kegel, Philipp; Schellmann, Maraike; Gorlatch, Sergei

    We compare two parallel programming approaches for multi-core systems: the well-known OpenMP and the recently introduced Threading Building Blocks (TBB) library by Intel®. The comparison is made using the parallelization of a real-world numerical algorithm for medical imaging. We develop several parallel implementations, and compare them w.r.t. programming effort, programming style and abstraction, and runtime performance. We show that TBB requires a considerable program re-design, whereas with OpenMP simple compiler directives are sufficient. While TBB appears to be less appropriate for parallelizing existing implementations, it fosters a good programming style and higher abstraction level for newly developed parallel programs. Our experimental measurements on a dual quad-core system demonstrate that OpenMP slightly outperforms TBB in our implementation.

  18. Parallel Large-Scale Molecular Dynamics Simulation Opens New Perspective to Clarify the Effect of a Porous Structure on the Sintering Process of Ni/YSZ Multiparticles.

    PubMed

    Xu, Jingxiang; Higuchi, Yuji; Ozawa, Nobuki; Sato, Kazuhisa; Hashida, Toshiyuki; Kubo, Momoji

    2017-09-20

    Ni sintering in the Ni/YSZ porous anode of a solid oxide fuel cell changes the porous structure, leading to degradation. Preventing sintering and degradation during operation is a great challenge. Usually, a sintering molecular dynamics (MD) simulation model consisting of two particles on a substrate is used; however, the model cannot reflect the porous structure effect on sintering. In our previous study, a multi-nanoparticle sintering modeling method with tens of thousands of atoms revealed the effect of the particle framework and porosity on sintering. However, the method cannot reveal the effect of the particle size on sintering and the effect of sintering on the change in the porous structure. In the present study, we report a strategy to reveal them in the porous structure by using our multi-nanoparticle modeling method and a parallel large-scale multimillion-atom MD simulator. We used this method to investigate the effect of YSZ particle size and tortuosity on sintering and degradation in the Ni/YSZ anodes. Our parallel large-scale MD simulation showed that the sintering degree decreased as the YSZ particle size decreased. The gas fuel diffusion path, which reflects the overpotential, was blocked by pore coalescence during sintering. The degradation of gas diffusion performance increased as the YSZ particle size increased. Furthermore, the gas diffusion performance was quantified by a tortuosity parameter and an optimal YSZ particle size, which is equal to that of Ni, was found for good diffusion after sintering. These findings cannot be obtained by previous MD sintering studies with tens of thousands of atoms. The present parallel large-scale multimillion-atom MD simulation makes it possible to clarify the effects of the particle size and tortuosity on sintering and degradation.

  19. Parallel transformation of K-SVD solar image denoising algorithm

    NASA Astrophysics Data System (ADS)

    Liang, Youwen; Tian, Yu; Li, Mei

    2017-02-01

    The images obtained by observing the sun through a large telescope always suffered with noise due to the low SNR. K-SVD denoising algorithm can effectively remove Gauss white noise. Training dictionaries for sparse representations is a time consuming task, due to the large size of the data involved and to the complexity of the training algorithms. In this paper, an OpenMP parallel programming language is proposed to transform the serial algorithm to the parallel version. Data parallelism model is used to transform the algorithm. Not one atom but multiple atoms updated simultaneously is the biggest change. The denoising effect and acceleration performance are tested after completion of the parallel algorithm. Speedup of the program is 13.563 in condition of using 16 cores. This parallel version can fully utilize the multi-core CPU hardware resources, greatly reduce running time and easily to transplant in multi-core platform.

  20. Hybrid Parallelism for Volume Rendering on Large-, Multi-, and Many-Core Systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Howison, Mark; Bethel, E. Wes; Childs, Hank

    2012-01-01

    With the computing industry trending towards multi- and many-core processors, we study how a standard visualization algorithm, ray-casting volume rendering, can benefit from a hybrid parallelism approach. Hybrid parallelism provides the best of both worlds: using distributed-memory parallelism across a large numbers of nodes increases available FLOPs and memory, while exploiting shared-memory parallelism among the cores within each node ensures that each node performs its portion of the larger calculation as efficiently as possible. We demonstrate results from weak and strong scaling studies, at levels of concurrency ranging up to 216,000, and with datasets as large as 12.2 trillion cells.more » The greatest benefit from hybrid parallelism lies in the communication portion of the algorithm, the dominant cost at higher levels of concurrency. We show that reducing the number of participants with a hybrid approach significantly improves performance.« less

  1. Observation of the Chiral and Achiral Hexatic Phases of Self-assembled Micellar polymers

    PubMed Central

    Pal, Antara; Kamal, Md. Arif; Raghunathan, V. A.

    2016-01-01

    We report the discovery of a thermodynamically stable line hexatic (N + 6) phase in a three-dimensional (3D) system made up of self-assembled polymer-like micelles of amphiphilic molecules. The experimentally observed phase transition sequence nematic (N)  N + 6  two-dimensional hexagonal (2D-H) is in good agreement with the theoretical predictions. Further, the present study also brings to light the effect of chirality on the N + 6 phase. In the chiral N + 6 phase the bond-orientational order within each “polymer” bundle is found to be twisted about an axis parallel to the average polymer direction. This structure is consistent with the theoretically envisaged Moiré state, thereby providing the first experimental demonstration of the Moiré structure. In addition to confirming the predictions of fundamental theories of two-dimensional melting, these results are relevant in a variety of situations in chemistry, physics and biology, where parallel packing of polymer-like objects are encountered. PMID:27577927

  2. Second International Workshop on Software Engineering and Code Design in Parallel Meteorological and Oceanographic Applications

    NASA Technical Reports Server (NTRS)

    OKeefe, Matthew (Editor); Kerr, Christopher L. (Editor)

    1998-01-01

    This report contains the abstracts and technical papers from the Second International Workshop on Software Engineering and Code Design in Parallel Meteorological and Oceanographic Applications, held June 15-18, 1998, in Scottsdale, Arizona. The purpose of the workshop is to bring together software developers in meteorology and oceanography to discuss software engineering and code design issues for parallel architectures, including Massively Parallel Processors (MPP's), Parallel Vector Processors (PVP's), Symmetric Multi-Processors (SMP's), Distributed Shared Memory (DSM) multi-processors, and clusters. Issues to be discussed include: (1) code architectures for current parallel models, including basic data structures, storage allocation, variable naming conventions, coding rules and styles, i/o and pre/post-processing of data; (2) designing modular code; (3) load balancing and domain decomposition; (4) techniques that exploit parallelism efficiently yet hide the machine-related details from the programmer; (5) tools for making the programmer more productive; and (6) the proliferation of programming models (F--, OpenMP, MPI, and HPF).

  3. Concurrent Probabilistic Simulation of High Temperature Composite Structural Response

    NASA Technical Reports Server (NTRS)

    Abdi, Frank

    1996-01-01

    A computational structural/material analysis and design tool which would meet industry's future demand for expedience and reduced cost is presented. This unique software 'GENOA' is dedicated to parallel and high speed analysis to perform probabilistic evaluation of high temperature composite response of aerospace systems. The development is based on detailed integration and modification of diverse fields of specialized analysis techniques and mathematical models to combine their latest innovative capabilities into a commercially viable software package. The technique is specifically designed to exploit the availability of processors to perform computationally intense probabilistic analysis assessing uncertainties in structural reliability analysis and composite micromechanics. The primary objectives which were achieved in performing the development were: (1) Utilization of the power of parallel processing and static/dynamic load balancing optimization to make the complex simulation of structure, material and processing of high temperature composite affordable; (2) Computational integration and synchronization of probabilistic mathematics, structural/material mechanics and parallel computing; (3) Implementation of an innovative multi-level domain decomposition technique to identify the inherent parallelism, and increasing convergence rates through high- and low-level processor assignment; (4) Creating the framework for Portable Paralleled architecture for the machine independent Multi Instruction Multi Data, (MIMD), Single Instruction Multi Data (SIMD), hybrid and distributed workstation type of computers; and (5) Market evaluation. The results of Phase-2 effort provides a good basis for continuation and warrants Phase-3 government, and industry partnership.

  4. SQDFT: Spectral Quadrature method for large-scale parallel O ( N ) Kohn–Sham calculations at high temperature

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Suryanarayana, Phanish; Pratapa, Phanisri P.; Sharma, Abhiraj

    We present SQDFT: a large-scale parallel implementation of the Spectral Quadrature (SQ) method formore » $$\\mathscr{O}(N)$$ Kohn–Sham Density Functional Theory (DFT) calculations at high temperature. Specifically, we develop an efficient and scalable finite-difference implementation of the infinite-cell Clenshaw–Curtis SQ approach, in which results for the infinite crystal are obtained by expressing quantities of interest as bilinear forms or sums of bilinear forms, that are then approximated by spatially localized Clenshaw–Curtis quadrature rules. We demonstrate the accuracy of SQDFT by showing systematic convergence of energies and atomic forces with respect to SQ parameters to reference diagonalization results, and convergence with discretization to established planewave results, for both metallic and insulating systems. Here, we further demonstrate that SQDFT achieves excellent strong and weak parallel scaling on computer systems consisting of tens of thousands of processors, with near perfect $$\\mathscr{O}(N)$$ scaling with system size and wall times as low as a few seconds per self-consistent field iteration. Finally, we verify the accuracy of SQDFT in large-scale quantum molecular dynamics simulations of aluminum at high temperature.« less

  5. SQDFT: Spectral Quadrature method for large-scale parallel O ( N ) Kohn–Sham calculations at high temperature

    DOE PAGES

    Suryanarayana, Phanish; Pratapa, Phanisri P.; Sharma, Abhiraj; ...

    2017-12-07

    We present SQDFT: a large-scale parallel implementation of the Spectral Quadrature (SQ) method formore » $$\\mathscr{O}(N)$$ Kohn–Sham Density Functional Theory (DFT) calculations at high temperature. Specifically, we develop an efficient and scalable finite-difference implementation of the infinite-cell Clenshaw–Curtis SQ approach, in which results for the infinite crystal are obtained by expressing quantities of interest as bilinear forms or sums of bilinear forms, that are then approximated by spatially localized Clenshaw–Curtis quadrature rules. We demonstrate the accuracy of SQDFT by showing systematic convergence of energies and atomic forces with respect to SQ parameters to reference diagonalization results, and convergence with discretization to established planewave results, for both metallic and insulating systems. Here, we further demonstrate that SQDFT achieves excellent strong and weak parallel scaling on computer systems consisting of tens of thousands of processors, with near perfect $$\\mathscr{O}(N)$$ scaling with system size and wall times as low as a few seconds per self-consistent field iteration. Finally, we verify the accuracy of SQDFT in large-scale quantum molecular dynamics simulations of aluminum at high temperature.« less

  6. Smooth adaptive sliding mode vibration control of a flexible parallel manipulator with multiple smart linkages in modal space

    NASA Astrophysics Data System (ADS)

    Zhang, Quan; Li, Chaodong; Zhang, Jiantao; Zhang, Jianhui

    2017-12-01

    This paper addresses the dynamic model and active vibration control of a rigid-flexible parallel manipulator with three smart links actuated by three linear ultrasonic motors. To suppress the vibration of three flexible intermediate links under high speed and acceleration, multiple Lead Zirconium Titanate (PZT) sensors and actuators are collocated mounted on each link, forming a smart structure which can achieve self-sensing and self-actuating. The dynamic characteristics and equations of the flexible link incorporated with the PZT sensors and actuator are analyzed and formulated. The smooth adaptive sliding mode based active vibration control is proposed to suppress the vibration of the smart links, and the first and second modes of the three links are targeted to be suppressed in modal space to avoid the spillover phenomenon. Simulations and experiments are implemented to validate the effectiveness of the smart structures and the proposed control laws. Experimental results show that the vibration of the first mode around 92 Hz and the second mode around 240 Hz of the three smart links are reduced respectively by 64.98%, 59.47%, 62.28%, and 45.80%, 36.79%, 33.33%, which further verify the multi-mode vibration control ability of the smooth adaptive sliding mode control law.

  7. Vacuum tube operation analysis under multi-harmonic driving and heavy beam loading effect in J-PARC RCS

    NASA Astrophysics Data System (ADS)

    Yamamoto, M.; Nomura, M.; Shimada, T.; Tamura, F.; Hara, K.; Hasegawa, K.; Ohmori, C.; Toda, M.; Yoshii, M.; Schnase, A.

    2016-11-01

    An rf cavity in the J-PARC RCS not only covers the frequency range of a fundamental acceleration pattern but also generates multi-harmonic rf voltage because it has a broadband impedance. However, analyzing the vacuum tube operation in the case of multi-harmonics is very complicated because many variables must be solved in a self-consistent manner. We developed a method to analyze the vacuum tube operation using a well-known formula and which includes the dependence on anode current for some variables. The calculation method is verified with beam tests, and the results indicate that it is efficient under condition of multi-harmonics with a heavy beam loading effect.

  8. Sub-arcsecond observations of the solar X-ray corona

    NASA Technical Reports Server (NTRS)

    Golub, L.; Nystrom, G.; Herant, M.; Kalata, K.; Lovas, I.

    1990-01-01

    Results from a high-resolution multi-layer-coated X-ray imaging telescope, part of the Normal Incidence X-ray Telescope sounding rocket payload are presented. Images of the peak of a two-ribbon flare showed detailed structure within each ribbon, as well as the expected bright arches of emission connecting the ribbons. The number of X-ray bright points is small, consistent with predictions based on the previous solar cycle. Topology of the magnetic structure is complex and highly tangled, implying that the magnetic complexity of the photosphere is paralleled in the corona.

  9. [Self-efficacy and self management of healthy habits in fibromyalgia].

    PubMed

    Pérez-Velasco, María; Peñacoba-Puente, Cecilia

    2015-01-01

    Fibromyalgia is a disorder characterized by general chronic pain, together with other symptoms such as fatigue, sleep disorders, anxiety and depression. To analyze, in FM patients, the effects of a multi-component intervention program (nursing+cognitive-behavioural therapy, focused on improving resting habits, physical exercise, and family relationships, working simultaneously on empowerment and patient self-efficacy. A quasi-experimental design was used following-up 5 women diagnosed with fibromyalgia. An analysis was performed on their daily habits, self-efficacy for chronic pain, pain perception, functional limitation, and affect. The intervention was composed by 8 group sessions: Six of them aimed at health education and self-management of healthy habits (nursing), and two sessions dedicated to increasing self-efficacy (cognitive-behavioural therapy). Follow-up consisted of five individual sessions (nursing) so as to consolidate the newly acquired habits, maintain self-management and self-efficacy based on observing compliance. Statistically significant improvements were observed (pre-, pos-) in habit modification and in self-efficacy, as well as for positive and negative affect. Also, statistically significant differences were found pre-follow up for functional limitation. The role of nursing has to be considered within multi-component programs, in particular during follow-up, for changing habits and for self-efficacy, in response to some of the current limitations of interventions with these patients. Copyright © 2015 Elsevier España, S.L.U. All rights reserved.

  10. Hydraulic Fracture Induced Seismicity During A Multi-Stage Pad Completion in Western Canada: Evidence of Activation of Multiple, Parallel Faults

    NASA Astrophysics Data System (ADS)

    Maxwell, S.; Garrett, D.; Huang, J.; Usher, P.; Mamer, P.

    2017-12-01

    Following reports of injection induced seismicity in the Western Canadian Sedimentary Basin, regulators have imposed seismic monitoring and traffic light protocols for fracturing operations in specific areas. Here we describe a case study in one of these reservoirs, the Montney Shale in NE British Columbia, where induced seismicity was monitored with a local array during multi-stage hydraulic fracture stimulations on several wells from a single drilling pad. Seismicity primarily occurred during the injection time periods, and correlated with periods of high injection rates and wellhead pressures above fracturing pressures. Sequential hydraulic fracture stages were found to progressively activate several parallel, critically-stressed faults, as illuminated by multiple linear hypocenter patterns in the range between Mw 1 and 3. Moment tensor inversion of larger events indicated a double-couple mechanism consistent with the regional strike-slip stress state and the hypocenter lineations. The critically-stressed faults obliquely cross the well paths which were purposely drilled parallel to the minimum principal stress direction. Seismicity on specific faults started and stopped when fracture initiation points of individual injection stages were proximal to the intersection of the fault and well. The distance ranges when the seismicity occurs is consistent with expected hydraulic fracture dimensions, suggesting that the induced fault slip only occurs when a hydraulic fracture grows directly into the fault and the faults are temporarily exposed to significantly elevated fracture pressures during the injection. Some faults crossed multiple wells and the seismicity was found to restart during injection of proximal stages on adjacent wells, progressively expanding the seismogenic zone of the fault. Progressive fault slip is therefore inferred from the seismicity migrating further along the faults during successive injection stages. An accelerometer was also deployed close to the pad operations providing information about the local ground motion at near offsets, although no ground motion was recorded that exceeds the minimum levels requiring mandatory reporting to the regulator.

  11. The Parallel System for Integrating Impact Models and Sectors (pSIMS)

    NASA Technical Reports Server (NTRS)

    Elliott, Joshua; Kelly, David; Chryssanthacopoulos, James; Glotter, Michael; Jhunjhnuwala, Kanika; Best, Neil; Wilde, Michael; Foster, Ian

    2014-01-01

    We present a framework for massively parallel climate impact simulations: the parallel System for Integrating Impact Models and Sectors (pSIMS). This framework comprises a) tools for ingesting and converting large amounts of data to a versatile datatype based on a common geospatial grid; b) tools for translating this datatype into custom formats for site-based models; c) a scalable parallel framework for performing large ensemble simulations, using any one of a number of different impacts models, on clusters, supercomputers, distributed grids, or clouds; d) tools and data standards for reformatting outputs to common datatypes for analysis and visualization; and e) methodologies for aggregating these datatypes to arbitrary spatial scales such as administrative and environmental demarcations. By automating many time-consuming and error-prone aspects of large-scale climate impacts studies, pSIMS accelerates computational research, encourages model intercomparison, and enhances reproducibility of simulation results. We present the pSIMS design and use example assessments to demonstrate its multi-model, multi-scale, and multi-sector versatility.

  12. Computational performance of a smoothed particle hydrodynamics simulation for shared-memory parallel computing

    NASA Astrophysics Data System (ADS)

    Nishiura, Daisuke; Furuichi, Mikito; Sakaguchi, Hide

    2015-09-01

    The computational performance of a smoothed particle hydrodynamics (SPH) simulation is investigated for three types of current shared-memory parallel computer devices: many integrated core (MIC) processors, graphics processing units (GPUs), and multi-core CPUs. We are especially interested in efficient shared-memory allocation methods for each chipset, because the efficient data access patterns differ between compute unified device architecture (CUDA) programming for GPUs and OpenMP programming for MIC processors and multi-core CPUs. We first introduce several parallel implementation techniques for the SPH code, and then examine these on our target computer architectures to determine the most effective algorithms for each processor unit. In addition, we evaluate the effective computing performance and power efficiency of the SPH simulation on each architecture, as these are critical metrics for overall performance in a multi-device environment. In our benchmark test, the GPU is found to produce the best arithmetic performance as a standalone device unit, and gives the most efficient power consumption. The multi-core CPU obtains the most effective computing performance. The computational speed of the MIC processor on Xeon Phi approached that of two Xeon CPUs. This indicates that using MICs is an attractive choice for existing SPH codes on multi-core CPUs parallelized by OpenMP, as it gains computational acceleration without the need for significant changes to the source code.

  13. Multi-line triggering and interdigitated electrode structure for photoconductive semiconductor switches

    DOEpatents

    Mar, Alan [Albuquerque, NM; Zutavern, Fred J [Albuquerque, NM; Loubriel, Guillermo [Albuquerque, NM

    2007-02-06

    An improved photoconductive semiconductor switch comprises multiple-line optical triggering of multiple, high-current parallel filaments between the switch electrodes. The switch can also have a multi-gap, interdigitated electrode for the generation of additional parallel filaments. Multi-line triggering can increase the switch lifetime at high currents by increasing the number of current filaments and reducing the current density at the contact electrodes in a controlled manner. Furthermore, the improved switch can mitigate the degradation of switching conditions with increased number of firings of the switch.

  14. A multi-center randomized controlled trial to compare a self-ligating bracket with a conventional bracket in a UK population: Part 1: Treatment efficiency.

    PubMed

    O'Dywer, Lian; Littlewood, Simon J; Rahman, Shahla; Spencer, R James; Barber, Sophy K; Russell, Joanne S

    2016-01-01

    To use a two-arm parallel trial to compare treatment efficiency between a self-ligating and a conventional preadjusted edgewise appliance system. A prospective multi-center randomized controlled clinical trial was conducted in three hospital orthodontic departments. Subjects were randomly allocated to receive treatment with either a self-ligating (3M SmartClip) or conventional (3M Victory) preadjusted edgewise appliance bracket system using a computer-generated random sequence concealed in opaque envelopes, with stratification for operator and center. Two operators followed a standardized protocol regarding bracket bonding procedure and archwire sequence. Efficiency of each ligation system was assessed by comparing the duration of treatment (months), total number of appointments (scheduled and emergency visits), and number of bracket bond failures. One hundred thirty-eight subjects (mean age 14 years 11 months) were enrolled in the study, of which 135 subjects (97.8%) completed treatment. The mean treatment time and number of visits were 25.12 months and 19.97 visits in the SmartClip group and 25.80 months and 20.37 visits in the Victory group. The overall bond failure rate was 6.6% for the SmartClip and 7.2% for Victory, with a similar debond distribution between the two appliances. No significant differences were found between the bracket systems in any of the outcome measures. No serious harm was observed from either bracket system. There was no clinically significant difference in treatment efficiency between treatment with a self-ligating bracket system and a conventional ligation system.

  15. A space-based climatology of diurnal MLT tidal winds, temperatures and densities from UARS wind measurements

    NASA Astrophysics Data System (ADS)

    Svoboda, Aaron A.; Forbes, Jeffrey M.; Miyahara, Saburo

    2005-11-01

    A self-consistent global tidal climatology, useful for comparing and interpreting radar observations from different locations around the globe, is created from space-based Upper Atmosphere Research Satellite (UARS) horizontal wind measurements. The climatology created includes tidal structures for horizontal winds, temperature and relative density, and is constructed by fitting local (in latitude and height) UARS wind data at 95 km to a set of basis functions called Hough mode extensions (HMEs). These basis functions are numerically computed modifications to Hough modes and are globally self-consistent in wind, temperature, and density. We first demonstrate this self-consistency with a proxy data set from the Kyushu University General Circulation Model, and then use a linear weighted superposition of the HMEs obtained from monthly fits to the UARS data to extrapolate the global, multi-variable tidal structure. A brief explanation of the HMEs’ origin is provided as well as information about a public website that has been set up to make the full extrapolated data sets available.

  16. A new system for parallel drug screening against multiple-resistant HIV mutants based on lentiviral self-inactivating (SIN) vectors and multi-colour analyses

    PubMed Central

    2013-01-01

    Background Despite progress in the development of combined antiretroviral therapies (cART), HIV infection remains a significant challenge for human health. Current problems of cART include multi-drug-resistant virus variants, long-term toxicity and enormous treatment costs. Therefore, the identification of novel effective drugs is urgently needed. Methods We developed a straightforward screening approach for simultaneously evaluating the sensitivity of multiple HIV gag-pol mutants to antiviral drugs in one assay. Our technique is based on multi-colour lentiviral self-inactivating (SIN) LeGO vector technology. Results We demonstrated the successful use of this approach for screening compounds against up to four HIV gag-pol variants (wild-type and three mutants) simultaneously. Importantly, the technique was adapted to Biosafety Level 1 conditions by utilising ecotropic pseudotypes. This allowed upscaling to a large-scale screening protocol exploited by pharmaceutical companies in a successful proof-of-concept experiment. Conclusions The technology developed here facilitates fast screening for anti-HIV activity of individual agents from large compound libraries. Although drugs targeting gag-pol variants were used here, our approach permits screening compounds that target several different, key cellular and viral functions of the HIV life-cycle. The modular principle of the method also allows the easy exchange of various mutations in HIV sequences. In conclusion, the methodology presented here provides a valuable new approach for the identification of novel anti-HIV drugs. PMID:23286882

  17. Scalability and Portability of Two Parallel Implementations of ADI

    NASA Technical Reports Server (NTRS)

    Phung, Thanh; VanderWijngaart, Rob F.

    1994-01-01

    Two domain decompositions for the implementation of the NAS Scalar Penta-diagonal Parallel Benchmark on MIMD systems are investigated, namely transposition and multi-partitioning. Hardware platforms considered are the Intel iPSC/860 and Paragon XP/S-15, and clusters of SGI workstations on ethernet, communicating through PVM. It is found that the multi-partitioning strategy offers the kind of coarse granularity that allows scaling up to hundreds of processors on a massively parallel machine. Moreover, efficiency is retained when the code is ported verbatim (save message passing syntax) to a PVM environment on a modest size cluster of workstations.

  18. A Performance Comparison of the Parallel Preconditioners for Iterative Methods for Large Sparse Linear Systems Arising from Partial Differential Equations on Structured Grids

    NASA Astrophysics Data System (ADS)

    Ma, Sangback

    In this paper we compare various parallel preconditioners such as Point-SSOR (Symmetric Successive OverRelaxation), ILU(0) (Incomplete LU) in the Wavefront ordering, ILU(0) in the Multi-color ordering, Multi-Color Block SOR (Successive OverRelaxation), SPAI (SParse Approximate Inverse) and pARMS (Parallel Algebraic Recursive Multilevel Solver) for solving large sparse linear systems arising from two-dimensional PDE (Partial Differential Equation)s on structured grids. Point-SSOR is well-known, and ILU(0) is one of the most popular preconditioner, but it is inherently serial. ILU(0) in the Wavefront ordering maximizes the parallelism in the natural order, but the lengths of the wave-fronts are often nonuniform. ILU(0) in the Multi-color ordering is a simple way of achieving a parallelism of the order N, where N is the order of the matrix, but its convergence rate often deteriorates as compared to that of natural ordering. We have chosen the Multi-Color Block SOR preconditioner combined with direct sparse matrix solver, since for the Laplacian matrix the SOR method is known to have a nondeteriorating rate of convergence when used with the Multi-Color ordering. By using block version we expect to minimize the interprocessor communications. SPAI computes the sparse approximate inverse directly by least squares method. Finally, ARMS is a preconditioner recursively exploiting the concept of independent sets and pARMS is the parallel version of ARMS. Experiments were conducted for the Finite Difference and Finite Element discretizations of five two-dimensional PDEs with large meshsizes up to a million on an IBM p595 machine with distributed memory. Our matrices are real positive, i. e., their real parts of the eigenvalues are positive. We have used GMRES(m) as our outer iterative method, so that the convergence of GMRES(m) for our test matrices are mathematically guaranteed. Interprocessor communications were done using MPI (Message Passing Interface) primitives. The results show that in general ILU(0) in the Multi-Color ordering ahd ILU(0) in the Wavefront ordering outperform the other methods but for symmetric and nearly symmetric 5-point matrices Multi-Color Block SOR gives the best performance, except for a few cases with a small number of processors.

  19. Topical Meeting on Optical Bistability Held at Rochester, New York on 15-17 June 1983.

    DTIC Science & Technology

    1983-01-01

    distortion of their initial directions of polarization : both of the beams are linearly polarized , with their electric vectors either (i)parallel to...New Zealand. ChSAM aIB ct Multistability, self-oscillation, and chaos in a model for polarization I Chas mnd Optlcal Bltabillty: Blfuraton...second circularly polarized pumping beam has been observed, transition sequence arises that is consistent with recent observ- Sense of response

  20. Stress-Strain Properties of SIFCON in Uniaxial Compression and Tension

    DTIC Science & Technology

    1988-08-01

    direction act as contacting beams whereas fibers aligned parallel to the loading direction act as individual columns . The combination of fiber-to-fiber...applicable to the study of SIFCON. These include such topics as the influence of strain rate on composite behavior, cyclic loading response, fiber-to-matrix...the specimen are shown in Figure 17. The grips consisted of self-clamping steel plates and a universal joint connection to the loading machine which

  1. Pollutant removal in a multi-stage municipal wastewater treatment system comprised of constructed wetlands and a maturation pond, in a temperate climate.

    PubMed

    Rivas, A; Barceló-Quintal, I; Moeller, G E

    2011-01-01

    A multi-stage municipal wastewater treatment system is proposed to comply with Mexican standards for discharge into receiving water bodies. The system is located in Santa Fe de la Laguna, Mexico, an area with a temperate climate. It was designed for 2,700 people equivalent (259.2 m3/d) and consists of a preliminary treatment, a septic tank as well as two modules operating in parallel, each consisting of a horizontal subsurface-flow wetland, a maturation pond and a vertical flow polishing wetland. After two years of operation, on-site research was performed. An efficient biochemical oxygen demand (BOD5) (94-98%), chemical oxygen demand (91-93%), total suspended solids (93-97%), total Kjeldahl nitrogen (56-88%) and fecal coliform (4-5 logs) removal was obtained. Significant phosphorus removal was not accomplished in this study (25-52%). Evapotranspiration was measured in different treatment units. This study demonstrates that during the dry season wastewater treatment by this multi-stage system cannot comply with the limits established by Mexican standards for receiving water bodies type 'C'. However, it has demonstrated the system's potential for less restrictive uses such as agricultural irrigation, recreation and provides the opportunity for wastewater treatment in rural areas without electric energy.

  2. Parallel LC circuit model for multi-band absorption and preliminary design of radiative cooling.

    PubMed

    Feng, Rui; Qiu, Jun; Liu, Linhua; Ding, Weiqiang; Chen, Lixue

    2014-12-15

    We perform a comprehensive analysis of multi-band absorption by exciting magnetic polaritons in the infrared region. According to the independent properties of the magnetic polaritons, we propose a parallel inductance and capacitance(PLC) circuit model to explain and predict the multi-band resonant absorption peaks, which is fully validated by using the multi-sized structure with identical dielectric spacing layer and the multilayer structure with the same strip width. More importantly, we present the application of the PLC circuit model to preliminarily design a radiative cooling structure realized by merging several close peaks together. This omnidirectional and polarization insensitive structure is a good candidate for radiative cooling application.

  3. Investigation of a Multi-Component Intervention Addressing Mathematical Reasoning and Self-Regulation of Behavior for Students with Emotional/Behavioral Disabilities

    ERIC Educational Resources Information Center

    Fisher, Marie B.

    2013-01-01

    For students with Emotional/Behavioral Disabilities (EBD), negative student outcomes are the poorest across disability categories, including high rates of school dropouts, unemployment and incarcerations. Mathematically, students with EBD receiving instruction in special education settings experience practices not consistent with recommendations…

  4. Social support, sense of community in school, and self-efficacy as resources during early adolescence: an integrative model.

    PubMed

    Vieno, Alessio; Santinello, Massimo; Pastore, Massimiliano; Perkins, Douglas D

    2007-03-01

    Influences of different sources of social support (from parents and friends), school sense of community, and self-efficacy on psychosocial well being (as measured by self-reported life satisfaction and psychological symptoms) in early adolescence were investigated in an integrative model. The model was tested using structural equation modeling. Multi-group comparisons were used to estimate differences between sex and age groups. The survey sample was composed of 7,097 students in Northern Italy (51.4% male) divided into three age cohorts (equivalent to 6th, 8th, and 10th grades with median ages of 11, 13, and 15). Findings obtained using SEM were consistent with self-efficacy and school sense of community mediating effects of social support on psychosocial adjustment. The multi-group comparison indicates a need for more complex developmental models and more research on how changing forms of support interact with each other as their effects also change during this important stage of the life. Implications for primary prevention and cross-cultural comparisons are discussed.

  5. Self-Consistent Model of Magnetospheric Ring Current and Propagating Electromagnetic Ion Cyclotron Waves: Waves in Multi-Ion Magnetosphere

    NASA Technical Reports Server (NTRS)

    Khazanov, G. V.; Gamayunov, K. V.; Gallagher, D. L.; Kozyra, J. U.

    2006-01-01

    The further development of a self-consistent theoretical model of interacting ring current ions and electromagnetic ion cyclotron waves (Khazanov et al., 2003) is presented In order to adequately take into account wave propagation and refraction in a multi-ion magnetosphere, we explicitly include the ray tracing equations in our previous self-consistent model and use the general form of the wave kinetic equation. This is a major new feature of the present model and, to the best of our knowledge, the ray tracing equations for the first time are explicitly employed on a global magnetospheric scale in order to self-consistently simulate the spatial, temporal, and spectral evolution of the ring current and of electromagnetic ion cyclotron waves To demonstrate the effects of EMIC wave propagation and refraction on the wave energy distribution and evolution, we simulate the May 1998 storm. The main findings of our simulation can be summarized as follows. First, owing to the density gradient at the plasmapause, the net wave refraction is suppressed, and He+-mode grows preferably at the plasmapause. This result is in total agreement with previous ray tracing studies and is very clearly found in presented B field spectrograms. Second, comparison of global wave distributions with the results from another ring current model (Kozyra et al., 1997) reveals that this new model provides more intense and more highly plasmapause-organized wave distributions during the May 1998 storm period Finally, it is found that He(+)-mode energy distributions are not Gaussian distributions and most important that wave energy can occupy not only the region of generation, i.e., the region of small wave normal angles, but all wave normal angles, including those to near 90 . The latter is extremely crucial for energy transfer to thermal plasmaspheric electrons by resonant Landau damping and subsequent downward heat transport and excitation of stable auroral red arcs.

  6. Self-Consistent Model of Magnetospheric Ring Current and Propagating Electromagnetic Ion Cyclotron Waves. 1; Waves in Multi Ion Magnetosphere

    NASA Technical Reports Server (NTRS)

    Khazanov, G. V.; Gumayunov, K. V.; Gallagher, D. L.; Kozyra, J. U.

    2006-01-01

    The further development of a self-consistent theoretical model of interacting ring current ions and electromagnetic ion cyclotron waves [Khazanov et al., 2003] is presented. In order to adequately take into account the wave propagation and refraction in a multi-ion plasmasphere, we explicitly include the ray tracing equations in our previous self-consistent model and use the general form of the wave kinetic equation. This is a major new feature of the present model and, to the best of our knowledge, the ray tracing equations for the first time are explicitly employed on a global magnetospheric scale in order to self-consistently simulate spatial, temporal, and spectral evolutions of the ring current and electromagnetic ion cyclotron waves. To demonstrate the effects of EMIC wave propagation and refraction on the EMIC wave energy distributions and evolution we simulate the May 1998 storm. The main findings of our simulation can be summarized as follows. First, due to the density gradient at the plasmapause, the net wave refraction is suppressed, and He(+)-mode grows preferably at plasmapause. This result is in a total agreement with the previous ray tracing studies, and very clear observed in presented B-field spectrograms. Second, comparison the global wave distributions with the results from other ring current model [Kozyra et al., 1997] reveals that our model provides more intense and higher plasmapause organized distributions during the May, 1998 storm period. Finally, the found He(+)-mode energy distributions are not Gaussian distributions, and most important that wave energy can occupy not only the region of generation, i. e. the region of small wave normal angles, but the entire wave normal angle region and even only the region near 90 degrees. The latter is extremely crucial for energy transfer to thermal plasmaspheric electrons by resonant Landau damping, and subsequent downward heat transport and excitation of stable auroral red arcs.

  7. STABILITY OF SMALL SELF-INTERSTITIAL CLUSTERS IN TUNGSTEN

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Setyawan, Wahyu; Nandipati, Giridhar; Kurtz, Richard J.

    2015-12-31

    Density functional theory was employed to explore the stability of interstitial clusters in W up to size seven. For each cluster size, the most stable configuration consists of parallel dumbbells. For clusters larger than size three, parallel dumbbells prefer to form in a multilayer fashion, instead of a planar structure. For size-7 clusters, the most stable configuration is a complete octahedron. The binding energy of a [111] dumbbell to the most stable cluster increases with cluster size, namely 2.49, 3.68, 4.76, 4.82, 5.47, and 6.85 eV for clusters of size 1, 2, 3, 4, 5, and 6, respectively. For amore » size-2 cluster, collinear dumbbells are still repulsive at the maximum allowable distance of 13.8 Å (the fifth neighbor along [111]). On the other hand, parallel dumbbells are strongly bound together. Two parallel dumbbells in which the axis-to-axis distance is within a cylindrical radius of 5.2 Å still exhibit a considerable binding of 0.28 eV. The most stable cluster in each size will be used to explore interactions with transmutation products.« less

  8. An embedded multi-core parallel model for real-time stereo imaging

    NASA Astrophysics Data System (ADS)

    He, Wenjing; Hu, Jian; Niu, Jingyu; Li, Chuanrong; Liu, Guangyu

    2018-04-01

    The real-time processing based on embedded system will enhance the application capability of stereo imaging for LiDAR and hyperspectral sensor. The task partitioning and scheduling strategies for embedded multiprocessor system starts relatively late, compared with that for PC computer. In this paper, aimed at embedded multi-core processing platform, a parallel model for stereo imaging is studied and verified. After analyzing the computing amount, throughout capacity and buffering requirements, a two-stage pipeline parallel model based on message transmission is established. This model can be applied to fast stereo imaging for airborne sensors with various characteristics. To demonstrate the feasibility and effectiveness of the parallel model, a parallel software was designed using test flight data, based on the 8-core DSP processor TMS320C6678. The results indicate that the design performed well in workload distribution and had a speed-up ratio up to 6.4.

  9. The self-consistent parallel electric field due to electrostatic ion-cyclotron turbulence in downward auroral-current regions of the Earth's magnetosphere. IV

    NASA Astrophysics Data System (ADS)

    Jasperse, John R.; Basu, Bamandas; Lund, Eric J.; Grossbard, Neil

    2010-06-01

    The physical processes that determine the self-consistent electric field (E∥) parallel to the magnetic field have been an unresolved problem in magnetospheric physics for over 40 years. Recently, a new multimoment fluid theory was developed for inhomogeneous, nonuniformly magnetized plasma in the guiding-center and gyrotropic approximation that includes the effect of electrostatic, turbulent, wave-particle interactions (see Jasperse et al. [Phys. Plasmas 13, 072903 (2006); Jasperse et al., Phys. Plasmas13, 112902 (2006)]). In the present paper and its companion paper [Jasperse et al., Phys. Plasmas 17, 062903 (2010)], which are intended as sequels to the earlier work, a fundamental model for downward, magnetic field-aligned (Birkeland) currents for quasisteady conditions is presented. The model includes the production of electrostatic ion-cyclotron turbulence in the long-range potential region by an electron, bump-on-tail-driven ion-cyclotron instability. Anomalous momentum transfer (anomalous resistivity) by itself is found to produce a very small contribution to E∥; however, the presence of electrostatic, ion-cyclotron turbulence has a very large effect on the altitude dependence of the entire quasisteady solution. Anomalous energy transfer (anomalous heating and cooling) modifies the density, drift, and temperature altitude profiles and hence the generalized parallel-pressure gradients and mirror forces in the electron and ion momentum-balance equations. As a result, |E∥| is enhanced by nearly a factor of 40 compared to its value when turbulence is absent. The space-averaged potential increase associated with the strong double layer at the bottom of the downward-current sheet is estimated using the FAST satellite data and the multimoment fluid theory.

  10. Shingle 2.0: generalising self-consistent and automated domain discretisation for multi-scale geophysical models

    NASA Astrophysics Data System (ADS)

    Candy, Adam S.; Pietrzak, Julie D.

    2018-01-01

    The approaches taken to describe and develop spatial discretisations of the domains required for geophysical simulation models are commonly ad hoc, model- or application-specific, and under-documented. This is particularly acute for simulation models that are flexible in their use of multi-scale, anisotropic, fully unstructured meshes where a relatively large number of heterogeneous parameters are required to constrain their full description. As a consequence, it can be difficult to reproduce simulations, to ensure a provenance in model data handling and initialisation, and a challenge to conduct model intercomparisons rigorously. This paper takes a novel approach to spatial discretisation, considering it much like a numerical simulation model problem of its own. It introduces a generalised, extensible, self-documenting approach to carefully describe, and necessarily fully, the constraints over the heterogeneous parameter space that determine how a domain is spatially discretised. This additionally provides a method to accurately record these constraints, using high-level natural language based abstractions that enable full accounts of provenance, sharing, and distribution. Together with this description, a generalised consistent approach to unstructured mesh generation for geophysical models is developed that is automated, robust and repeatable, quick-to-draft, rigorously verified, and consistent with the source data throughout. This interprets the description above to execute a self-consistent spatial discretisation process, which is automatically validated to expected discrete characteristics and metrics. Library code, verification tests, and examples available in the repository at https://github.com/shingleproject/Shingle. Further details of the project presented at http://shingleproject.org.

  11. GPU accelerated dynamic functional connectivity analysis for functional MRI data.

    PubMed

    Akgün, Devrim; Sakoğlu, Ünal; Esquivel, Johnny; Adinoff, Bryon; Mete, Mutlu

    2015-07-01

    Recent advances in multi-core processors and graphics card based computational technologies have paved the way for an improved and dynamic utilization of parallel computing techniques. Numerous applications have been implemented for the acceleration of computationally-intensive problems in various computational science fields including bioinformatics, in which big data problems are prevalent. In neuroimaging, dynamic functional connectivity (DFC) analysis is a computationally demanding method used to investigate dynamic functional interactions among different brain regions or networks identified with functional magnetic resonance imaging (fMRI) data. In this study, we implemented and analyzed a parallel DFC algorithm based on thread-based and block-based approaches. The thread-based approach was designed to parallelize DFC computations and was implemented in both Open Multi-Processing (OpenMP) and Compute Unified Device Architecture (CUDA) programming platforms. Another approach developed in this study to better utilize CUDA architecture is the block-based approach, where parallelization involves smaller parts of fMRI time-courses obtained by sliding-windows. Experimental results showed that the proposed parallel design solutions enabled by the GPUs significantly reduce the computation time for DFC analysis. Multicore implementation using OpenMP on 8-core processor provides up to 7.7× speed-up. GPU implementation using CUDA yielded substantial accelerations ranging from 18.5× to 157× speed-up once thread-based and block-based approaches were combined in the analysis. Proposed parallel programming solutions showed that multi-core processor and CUDA-supported GPU implementations accelerated the DFC analyses significantly. Developed algorithms make the DFC analyses more practical for multi-subject studies with more dynamic analyses. Copyright © 2015 Elsevier Ltd. All rights reserved.

  12. Coding for parallel execution of hardware-in-the-loop millimeter-wave scene generation models on multicore SIMD processor architectures

    NASA Astrophysics Data System (ADS)

    Olson, Richard F.

    2013-05-01

    Rendering of point scatterer based radar scenes for millimeter wave (mmW) seeker tests in real-time hardware-in-the-loop (HWIL) scene generation requires efficient algorithms and vector-friendly computer architectures for complex signal synthesis. New processor technology from Intel implements an extended 256-bit vector SIMD instruction set (AVX, AVX2) in a multi-core CPU design providing peak execution rates of hundreds of GigaFLOPS (GFLOPS) on one chip. Real world mmW scene generation code can approach peak SIMD execution rates only after careful algorithm and source code design. An effective software design will maintain high computing intensity emphasizing register-to-register SIMD arithmetic operations over data movement between CPU caches or off-chip memories. Engineers at the U.S. Army Aviation and Missile Research, Development and Engineering Center (AMRDEC) applied two basic parallel coding methods to assess new 256-bit SIMD multi-core architectures for mmW scene generation in HWIL. These include use of POSIX threads built on vector library functions and more portable, highlevel parallel code based on compiler technology (e.g. OpenMP pragmas and SIMD autovectorization). Since CPU technology is rapidly advancing toward high processor core counts and TeraFLOPS peak SIMD execution rates, it is imperative that coding methods be identified which produce efficient and maintainable parallel code. This paper describes the algorithms used in point scatterer target model rendering, the parallelization of those algorithms, and the execution performance achieved on an AVX multi-core machine using the two basic parallel coding methods. The paper concludes with estimates for scale-up performance on upcoming multi-core technology.

  13. Using a source-to-source transformation to introduce multi-threading into the AliRoot framework for a parallel event reconstruction

    NASA Astrophysics Data System (ADS)

    Lohn, Stefan B.; Dong, Xin; Carminati, Federico

    2012-12-01

    Chip-Multiprocessors are going to support massive parallelism by many additional physical and logical cores. Improving performance can no longer be obtained by increasing clock-frequency because the technical limits are almost reached. Instead, parallel execution must be used to gain performance. Resources like main memory, the cache hierarchy, bandwidth of the memory bus or links between cores and sockets are not going to be improved as fast. Hence, parallelism can only result into performance gains if the memory usage is optimized and the communication between threads is minimized. Besides concurrent programming has become a domain for experts. Implementing multi-threading is error prone and labor-intensive. A full reimplementation of the whole AliRoot source-code is unaffordable. This paper describes the effort to evaluate the adaption of AliRoot to the needs of multi-threading and to provide the capability of parallel processing by using a semi-automatic source-to-source transformation to address the problems as described before and to provide a straight-forward way of parallelization with almost no interference between threads. This makes the approach simple and reduces the required manual changes in the code. In a first step, unconditional thread-safety will be introduced to bring the original sequential and thread unaware source-code into the position of utilizing multi-threading. Afterwards further investigations have to be performed to point out candidates of classes that are useful to share amongst threads. Then in a second step, the transformation has to change the code to share these classes and finally to verify if there are anymore invalid interferences between threads.

  14. Multi-Objective and Multidisciplinary Design Optimisation (MDO) of UAV Systems using Hierarchical Asynchronous Parallel Evolutionary Algorithms

    DTIC Science & Technology

    2007-09-17

    been proposed; these include a combination of variable fidelity models, parallelisation strategies and hybridisation techniques (Coello, Veldhuizen et...Coello et al (Coello, Veldhuizen et al. 2002). 4.4.2 HIERARCHICAL POPULATION TOPOLOGY A hierarchical population topology, when integrated into...to hybrid parallel Multi-Objective Evolutionary Algorithms (pMOEA) (Cantu-Paz 2000; Veldhuizen , Zydallis et al. 2003); it uses a master slave

  15. Accelerating Pathology Image Data Cross-Comparison on CPU-GPU Hybrid Systems

    PubMed Central

    Wang, Kaibo; Huai, Yin; Lee, Rubao; Wang, Fusheng; Zhang, Xiaodong; Saltz, Joel H.

    2012-01-01

    As an important application of spatial databases in pathology imaging analysis, cross-comparing the spatial boundaries of a huge amount of segmented micro-anatomic objects demands extremely data- and compute-intensive operations, requiring high throughput at an affordable cost. However, the performance of spatial database systems has not been satisfactory since their implementations of spatial operations cannot fully utilize the power of modern parallel hardware. In this paper, we provide a customized software solution that exploits GPUs and multi-core CPUs to accelerate spatial cross-comparison in a cost-effective way. Our solution consists of an efficient GPU algorithm and a pipelined system framework with task migration support. Extensive experiments with real-world data sets demonstrate the effectiveness of our solution, which improves the performance of spatial cross-comparison by over 18 times compared with a parallelized spatial database approach. PMID:23355955

  16. On-board landmark navigation and attitude reference parallel processor system

    NASA Technical Reports Server (NTRS)

    Gilbert, L. E.; Mahajan, D. T.

    1978-01-01

    An approach to autonomous navigation and attitude reference for earth observing spacecraft is described along with the landmark identification technique based on a sequential similarity detection algorithm (SSDA). Laboratory experiments undertaken to determine if better than one pixel accuracy in registration can be achieved consistent with onboard processor timing and capacity constraints are included. The SSDA is implemented using a multi-microprocessor system including synchronization logic and chip library. The data is processed in parallel stages, effectively reducing the time to match the small known image within a larger image as seen by the onboard image system. Shared memory is incorporated in the system to help communicate intermediate results among microprocessors. The functions include finding mean values and summation of absolute differences over the image search area. The hardware is a low power, compact unit suitable to onboard application with the flexibility to provide for different parameters depending upon the environment.

  17. A multi-satellite orbit determination problem in a parallel processing environment

    NASA Technical Reports Server (NTRS)

    Deakyne, M. S.; Anderle, R. J.

    1988-01-01

    The Engineering Orbit Analysis Unit at GE Valley Forge used an Intel Hypercube Parallel Processor to investigate the performance and gain experience of parallel processors with a multi-satellite orbit determination problem. A general study was selected in which major blocks of computation for the multi-satellite orbit computations were used as units to be assigned to the various processors on the Hypercube. Problems encountered or successes achieved in addressing the orbit determination problem would be more likely to be transferable to other parallel processors. The prime objective was to study the algorithm to allow processing of observations later in time than those employed in the state update. Expertise in ephemeris determination was exploited in addressing these problems and the facility used to bring a realism to the study which would highlight the problems which may not otherwise be anticipated. Secondary objectives were to gain experience of a non-trivial problem in a parallel processor environment, to explore the necessary interplay of serial and parallel sections of the algorithm in terms of timing studies, to explore the granularity (coarse vs. fine grain) to discover the granularity limit above which there would be a risk of starvation where the majority of nodes would be idle or under the limit where the overhead associated with splitting the problem may require more work and communication time than is useful.

  18. Optimal Golomb Ruler Sequences Generation for Optical WDM Systems: A Novel Parallel Hybrid Multi-objective Bat Algorithm

    NASA Astrophysics Data System (ADS)

    Bansal, Shonak; Singh, Arun Kumar; Gupta, Neena

    2017-02-01

    In real-life, multi-objective engineering design problems are very tough and time consuming optimization problems due to their high degree of nonlinearities, complexities and inhomogeneity. Nature-inspired based multi-objective optimization algorithms are now becoming popular for solving multi-objective engineering design problems. This paper proposes original multi-objective Bat algorithm (MOBA) and its extended form, namely, novel parallel hybrid multi-objective Bat algorithm (PHMOBA) to generate shortest length Golomb ruler called optimal Golomb ruler (OGR) sequences at a reasonable computation time. The OGRs found their application in optical wavelength division multiplexing (WDM) systems as channel-allocation algorithm to reduce the four-wave mixing (FWM) crosstalk. The performances of both the proposed algorithms to generate OGRs as optical WDM channel-allocation is compared with other existing classical computing and nature-inspired algorithms, including extended quadratic congruence (EQC), search algorithm (SA), genetic algorithms (GAs), biogeography based optimization (BBO) and big bang-big crunch (BB-BC) optimization algorithms. Simulations conclude that the proposed parallel hybrid multi-objective Bat algorithm works efficiently as compared to original multi-objective Bat algorithm and other existing algorithms to generate OGRs for optical WDM systems. The algorithm PHMOBA to generate OGRs, has higher convergence and success rate than original MOBA. The efficiency improvement of proposed PHMOBA to generate OGRs up to 20-marks, in terms of ruler length and total optical channel bandwidth (TBW) is 100 %, whereas for original MOBA is 85 %. Finally the implications for further research are also discussed.

  19. Multi-objective Calibration of DHSVM Based on Hydrologic Key Elements in Jinhua River Basin, East China

    NASA Astrophysics Data System (ADS)

    Pan, S.; Liu, L.; Xu, Y. P.

    2017-12-01

    Abstract: In physically based distributed hydrological model, large number of parameters, representing spatial heterogeneity of watershed and various processes in hydrologic cycle, are involved. For lack of calibration module in Distributed Hydrology Soil Vegetation Model, this study developed a multi-objective calibration module using Epsilon-Dominance Non-Dominated Sorted Genetic Algorithm II (ɛ-NSGAII) and based on parallel computing of Linux cluster for DHSVM (ɛP-DHSVM). In this study, two hydrologic key elements (i.e., runoff and evapotranspiration) are used as objectives in multi-objective calibration of model. MODIS evapotranspiration obtained by SEBAL is adopted to fill the gap of lack of observation for evapotranspiration. The results show that good performance of runoff simulation in single objective calibration cannot ensure good simulation performance of other hydrologic key elements. Self-developed ɛP-DHSVM model can make multi-objective calibration more efficiently and effectively. The running speed can be increased by more than 20-30 times via applying ɛP-DHSVM. In addition, runoff and evapotranspiration can be simulated very well simultaneously by ɛP-DHSVM, with superior values for two efficiency coefficients (0.74 for NS of runoff and 0.79 for NS of evapotranspiration, -10.5% and -8.6% for PBIAS of runoff and evapotranspiration respectively).

  20. Neural simulations on multi-core architectures.

    PubMed

    Eichner, Hubert; Klug, Tobias; Borst, Alexander

    2009-01-01

    Neuroscience is witnessing increasing knowledge about the anatomy and electrophysiological properties of neurons and their connectivity, leading to an ever increasing computational complexity of neural simulations. At the same time, a rather radical change in personal computer technology emerges with the establishment of multi-cores: high-density, explicitly parallel processor architectures for both high performance as well as standard desktop computers. This work introduces strategies for the parallelization of biophysically realistic neural simulations based on the compartmental modeling technique and results of such an implementation, with a strong focus on multi-core architectures and automation, i.e. user-transparent load balancing.

  1. Neural Simulations on Multi-Core Architectures

    PubMed Central

    Eichner, Hubert; Klug, Tobias; Borst, Alexander

    2009-01-01

    Neuroscience is witnessing increasing knowledge about the anatomy and electrophysiological properties of neurons and their connectivity, leading to an ever increasing computational complexity of neural simulations. At the same time, a rather radical change in personal computer technology emerges with the establishment of multi-cores: high-density, explicitly parallel processor architectures for both high performance as well as standard desktop computers. This work introduces strategies for the parallelization of biophysically realistic neural simulations based on the compartmental modeling technique and results of such an implementation, with a strong focus on multi-core architectures and automation, i.e. user-transparent load balancing. PMID:19636393

  2. Multi-Probe SPM using Interference Patterns for a Parallel Nano Imaging

    NASA Astrophysics Data System (ADS)

    Koyama, Hirotaka; Oohira, Fumikazu; Hosogi, Maho; Hashiguchi, Gen

    This paper proposes a new composition of the multi-probe using optical interference patterns for a parallel nano imaging in a large area scanning. We achieved large-scale integration with 50,000 probes fabricated with MEMS technology, and measured the optical interference patterns with CCD, which was difficult in a conventional single scanning probe. In this research, the multi-probes are made of Si3N4 by MEMS process, and, the multi-probes are joined with a Pyrex glass by an anodic bonding. We designed, fabricated, and evaluated the characteristics of the probe. In addition, we changed the probe shape to decrease the warpage of the Si3N4 probe. We used the supercritical drying to avoid stiction of the Si3N4 probe with the glass surface and fabricated 4 types of the probe shapes without stiction. We took some interference patterns by CCD and measured the position of them. We calculate the probe height using the interference displacement and compared the result with the theoretical deflection curve. As a result, these interference patterns matched the theoretical deflection curve. We found that this multi-probe chip using interference patterns is effective in measurement for a parallel nano imaging.

  3. PROPELLER for motion-robust imaging of in vivo mouse abdomen at 9.4 T.

    PubMed

    Teh, Irvin; Golay, Xavier; Larkman, David J

    2010-11-01

    In vivo high-field MRI in the abdomen of small animals is technically challenging because of the small voxel sizes, short T(2) and physiological motion. In standard Cartesian sampling, respiratory and gastrointestinal motion can lead to ghosting artefacts. Although respiratory triggering and navigator echoes can either avoid or compensate for motion, they can lead to variable TRs, require invasive intubation and ventilation, or extend TEs. A self-navigated fast spin echo (FSE)-based periodically rotated overlapping parallel lines with enhanced reconstruction (PROPELLER) acquisition was implemented at 9.4 T to enable high-resolution in vivo MRI of mouse abdomen without the use of additional navigators or triggering. T(2)-weighted FSE-PROPELLER data were compared with single-shot FSE and multi-shot FSE data with and without triggering. Single-shot methods, although rapid and robust to motion, demonstrated strong blurring. Multi-shot FSE data showed better resolution, but suffered from marked blurring in the phase-encoding direction and motion in between shots, leading to ghosting artefacts. When respiratory triggering was used, motion artefacts were largely avoided. However, TRs and acquisition times were lengthened by up to approximately 20%. The PROPELLER data showed a 25% and 61% improvement in signal-to-noise ratio and contrast-to-noise ratio, respectively, compared with multi-shot FSE data, together with a 35% reduction in artefact power. A qualitative comparison between acquisition methods using diffusion-weighted imaging was performed. The results were similar, with the exception that respiratory triggering was unable to exclude major motion artefacts as a result of the sensitisation to motion by the diffusion gradients. The PROPELLER data were of consistently higher quality. Considerations specific to the use of PROPELLER at high field are discussed, including the selection of practical blade widths and the effects on contrast, resolution and artefacts.

  4. The group-based social skills training SOSTA-FRA in children and adolescents with high functioning autism spectrum disorder - study protocol of the randomised, multi-centre controlled SOSTA - net trial

    PubMed Central

    2013-01-01

    Background Group-based social skills training (SST) has repeatedly been recommended as treatment of choice in high-functioning autism spectrum disorder (HFASD). To date, no sufficiently powered randomised controlled trial has been performed to establish efficacy and safety of SST in children and adolescents with HFASD. In this randomised, multi-centre, controlled trial with 220 children and adolescents with HFASD it is hypothesized, that add-on group-based SST using the 12 weeks manualised SOSTA–FRA program will result in improved social responsiveness (measured by the parent rated social responsiveness scale, SRS) compared to treatment as usual (TAU). It is further expected, that parent and self reported anxiety and depressive symptoms will decline and pro-social behaviour will increase in the treatment group. A neurophysiological study in the Frankfurt HFASD subgroup will be performed pre- and post treatment to assess changes in neural function induced by SST versus TAU. Methods/design The SOSTA – net trial is designed as a prospective, randomised, multi-centre, controlled trial with two parallel groups. The primary outcome is change in SRS score directly after the intervention and at 3 months follow-up. Several secondary outcome measures are also obtained. The target sample consists of 220 individuals with ASD, included at the six study centres. Discussion This study is currently one of the largest trials on SST in children and adolescents with HFASD worldwide. Compared to recent randomised controlled studies, our study shows several advantages with regard to in- and exclusion criteria, study methods, and the therapeutic approach chosen, which can be easily implemented in non-university-based clinical settings. Trial registration ISRCTN94863788 – SOSTA – net: Group-based social skills training in children and adolescents with high functioning autism spectrum disorder. PMID:23289935

  5. Modelling of radio frequency sheath and fast wave coupling on the realistic ion cyclotron resonant antenna surroundings and the outer wall

    NASA Astrophysics Data System (ADS)

    Lu, L.; Colas, L.; Jacquot, J.; Després, B.; Heuraux, S.; Faudot, E.; Van Eester, D.; Crombé, K.; Křivská, A.; Noterdaeme, J.-M.; Helou, W.; Hillairet, J.

    2018-03-01

    In order to model the sheath rectification in a realistic geometry over the size of ion cyclotron resonant heating (ICRH) antennas, the self-consistent sheaths and waves for ICH (SSWICH) code couples self-consistently the RF wave propagation and the DC SOL biasing via nonlinear RF and DC sheath boundary conditions applied at plasma/wall interfaces. A first version of SSWICH had 2D (toroidal and radial) geometry, rectangular walls either normal or parallel to the confinement magnetic field B 0 and only included the evanescent slow wave (SW) excited parasitically by the ICRH antenna. The main wave for plasma heating, the fast wave (FW) plays no role on the sheath excitation in this version. A new version of the code, 2D SSWICH-full wave, was developed based on the COMSOL software, to accommodate full RF field polarization and shaped walls tilted with respect to B 0 . SSWICH-full wave simulations have shown the mode conversion of FW into SW occurring at the sharp corners where the boundary shape varies rapidly. It has also evidenced ‘far-field’ sheath oscillations appearing at the shaped walls with a relatively long magnetic connection length to the antenna, that are only accessible to the propagating FW. Joint simulation, conducted by SSWICH-full wave within a multi-2D approach excited using the 3D wave coupling code (RAPLICASOL), has recovered the double-hump poloidal structure measured in the experimental temperature and potential maps when only the SW is modelled. The FW contribution on the potential poloidal structure seems to be affected by the 3D effects, which was ignored in the current stage. Finally, SSWICH-full wave simulation revealed the left-right asymmetry that has been observed extensively in the unbalanced strap feeding experiments, suggesting that the spatial proximity effects in RF sheath excitation, studied for SW only previously, is still important in the vicinity of the wave launcher under full wave polarizations.

  6. A Microelectrode Array with Reproducible Performance Shows Loss of Consistency Following Functionalization with a Self-Assembled 6-Mercapto-1-hexanol Layer.

    PubMed

    Corrigan, Damion K; Vezza, Vincent; Schulze, Holger; Bachmann, Till T; Mount, Andrew R; Walton, Anthony J; Terry, Jonathan G

    2018-06-09

    For analytical applications involving label-free biosensors and multiple measurements, i.e., across an electrode array, it is essential to develop complete sensor systems capable of functionalization and of producing highly consistent responses. To achieve this, a multi-microelectrode device bearing twenty-four equivalent 50 µm diameter Pt disc microelectrodes was designed in an integrated 3-electrode system configuration and then fabricated. Cyclic voltammetry and electrochemical impedance spectroscopy were used for initial electrochemical characterization of the individual working electrodes. These confirmed the expected consistency of performance with a high degree of measurement reproducibility for each microelectrode across the array. With the aim of assessing the potential for production of an enhanced multi-electrode sensor for biomedical use, the working electrodes were then functionalized with 6-mercapto-1-hexanol (MCH). This is a well-known and commonly employed surface modification process, which involves the same principles of thiol attachment chemistry and self-assembled monolayer (SAM) formation commonly employed in the functionalization of electrodes and the formation of biosensors. Following this SAM formation, the reproducibility of the observed electrochemical signal between electrodes was seen to decrease markedly, compromising the ability to achieve consistent analytical measurements from the sensor array following this relatively simple and well-established surface modification. To successfully and consistently functionalize the sensors, it was necessary to dilute the constituent molecules by a factor of ten thousand to support adequate SAM formation on microelectrodes. The use of this multi-electrode device therefore demonstrates in a high throughput manner irreproducibility in the SAM formation process at the higher concentration, even though these electrodes are apparently functionalized simultaneously in the same film formation environment, confirming that the often seen significant electrode-to-electrode variation in label-free SAM biosensing films formed under such conditions is not likely to be due to variation in film deposition conditions, but rather kinetically controlled variation in the SAM layer formation process at these microelectrodes.

  7. A high performance parallel algorithm for 1-D FFT

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Agarwal, R.C.; Gustavson, F.G.; Zubair, M.

    1994-12-31

    In this paper the authors propose a parallel high performance FFT algorithm based on a multi-dimensional formulation. They use this to solve a commonly encountered FFT based kernel on a distributed memory parallel machine, the IBM scalable parallel system, SP1. The kernel requires a forward FFT computation of an input sequence, multiplication of the transformed data by a coefficient array, and finally an inverse FFT computation of the resultant data. They show that the multi-dimensional formulation helps in reducing the communication costs and also improves the single node performance by effectively utilizing the memory system of the node. They implementedmore » this kernel on the IBM SP1 and observed a performance of 1.25 GFLOPS on a 64-node machine.« less

  8. Family, Community and Clinic Collaboration to Treat Overweight and Obese Children: Stanford GOALS -- a Randomized Controlled Trial of a Three-Year, Multi-Component, Multi-Level, Multi-Setting Intervention

    PubMed Central

    Robinson, Thomas N.; Matheson, Donna; Desai, Manisha; Wilson, Darrell M.; Weintraub, Dana L.; Haskell, William L.; McClain, Arianna; McClure, Samuel; Banda, Jorge; Sanders, Lee M.; Haydel, K. Farish; Killen, Joel D.

    2013-01-01

    Objective To test the effects of a three-year, community-based, multi-component, multi-level, multi-setting (MMM) approach for treating overweight and obese children. Design Two-arm, parallel group, randomized controlled trial with measures at baseline, 12, 24, and 36 months after randomization. Participants Seven through eleven year old, overweight and obese children (BMI ≥ 85th percentile) and their parents/caregivers recruited from community locations in low-income, primarily Latino neighborhoods in Northern California. Interventions Families are randomized to the MMM intervention versus a community health education active-placebo comparison intervention. Interventions last for three years for each participant. The MMM intervention includes a community-based after school team sports program designed specifically for overweight and obese children, a home-based family intervention to reduce screen time, alter the home food/eating environment, and promote self-regulatory skills for eating and activity behavior change, and a primary care behavioral counseling intervention linked to the community and home interventions. The active-placebo comparison intervention includes semi-annual health education home visits, monthly health education newsletters for children and for parents/guardians, and a series of community-based health education events for families. Main Outcome Measure Body mass index trajectory over the three-year study. Secondary outcome measures include waist circumference, triceps skinfold thickness, accelerometer-measured physical activity, 24-hour dietary recalls, screen time and other sedentary behaviors, blood pressure, fasting lipids, glucose, insulin, hemoglobin A1c, C-reactive protein, alanine aminotransferase, and psychosocial measures. Conclusions The Stanford GOALS trial is testing the efficacy of a novel community-based multi-component, multi-level, multi-setting treatment for childhood overweight and obesity in low-income, Latino families. PMID:24028942

  9. Family, community and clinic collaboration to treat overweight and obese children: Stanford GOALS-A randomized controlled trial of a three-year, multi-component, multi-level, multi-setting intervention.

    PubMed

    Robinson, Thomas N; Matheson, Donna; Desai, Manisha; Wilson, Darrell M; Weintraub, Dana L; Haskell, William L; McClain, Arianna; McClure, Samuel; Banda, Jorge A; Sanders, Lee M; Haydel, K Farish; Killen, Joel D

    2013-11-01

    To test the effects of a three-year, community-based, multi-component, multi-level, multi-setting (MMM) approach for treating overweight and obese children. Two-arm, parallel group, randomized controlled trial with measures at baseline, 12, 24, and 36 months after randomization. Seven through eleven year old, overweight and obese children (BMI ≥ 85th percentile) and their parents/caregivers recruited from community locations in low-income, primarily Latino neighborhoods in Northern California. Families are randomized to the MMM intervention versus a community health education active-placebo comparison intervention. Interventions last for three years for each participant. The MMM intervention includes a community-based after school team sports program designed specifically for overweight and obese children, a home-based family intervention to reduce screen time, alter the home food/eating environment, and promote self-regulatory skills for eating and activity behavior change, and a primary care behavioral counseling intervention linked to the community and home interventions. The active-placebo comparison intervention includes semi-annual health education home visits, monthly health education newsletters for children and for parents/guardians, and a series of community-based health education events for families. Body mass index trajectory over the three-year study. Secondary outcome measures include waist circumference, triceps skinfold thickness, accelerometer-measured physical activity, 24-hour dietary recalls, screen time and other sedentary behaviors, blood pressure, fasting lipids, glucose, insulin, hemoglobin A1c, C-reactive protein, alanine aminotransferase, and psychosocial measures. The Stanford GOALS trial is testing the efficacy of a novel community-based multi-component, multi-level, multi-setting treatment for childhood overweight and obesity in low-income, Latino families. © 2013 Elsevier Inc. All rights reserved.

  10. A WENO-Limited, ADER-DT, Finite-Volume Scheme for Efficient, Robust, and Communication-Avoiding Multi-Dimensional Transport

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Norman, Matthew R

    2014-01-01

    The novel ADER-DT time discretization is applied to two-dimensional transport in a quadrature-free, WENO- and FCT-limited, Finite-Volume context. Emphasis is placed on (1) the serial and parallel computational properties of ADER-DT and this framework and (2) the flexibility of ADER-DT and this framework in efficiently balancing accuracy with other constraints important to transport applications. This study demonstrates a range of choices for the user when approaching their specific application while maintaining good parallel properties. In this method, genuine multi-dimensionality, single-step and single-stage time stepping, strict positivity, and a flexible range of limiting are all achieved with only one parallel synchronizationmore » and data exchange per time step. In terms of parallel data transfers per simulated time interval, this improves upon multi-stage time stepping and post-hoc filtering techniques such as hyperdiffusion. This method is evaluated with standard transport test cases over a range of limiting options to demonstrate quantitatively and qualitatively what a user should expect when employing this method in their application.« less

  11. Reconceptualizing Efficacy in Substance Use Prevention Research: Refusal Response Efficacy and Drug Resistance Self-Efficacy in Adolescent Substance Use

    PubMed Central

    Choi, Hye Jeong; Krieger, Janice L.; Hecht, Michael L.

    2014-01-01

    The purpose of this study is to utilize the Extended Parallel Process Model (EPPM) to expand the construct of efficacy in the adolescent substance use context. Using survey data collected from 2,129 seventh-grade students in 39 rural schools, we examined the construct of drug refusal efficacy and demonstrated relationships among response efficacy (RE), self-efficacy (SE), and adolescent drug use. Consistent with the hypotheses, confirmatory factor analyses of a 12-item scale yielded a three-factor solution: refusal RE, alcohol-resistance self-efficacy (ASE), and marijuana-resistance self-efficacy (MSE). Refusal RE and ASE/MSE were negatively related to alcohol use and marijuana use, whereas MSE was positively associated with alcohol use. These data demonstrate that efficacy is a broader construct than typically considered in drug prevention. Prevention programs should reinforce both refusal RE and substance-specific resistance SE. PMID:23330857

  12. Effects of Prior Aging at 288 deg C in Argon Environment on Creep Response of Carbon Fiber Reinforced PMR-15 Composite with + or - 45 deg Fiber Orientation at 288 deg C

    DTIC Science & Technology

    2009-06-01

    typically consists of a thermoset or thermoplastic polymer matrix reinforced with fibers that are much stronger and stiffer than the matrix. The PMCs are...high thermal or electrical conductivity, stealth characteristics , the ability to self-heal, communication, and sensor capabilities. The multi...have factual evidence of limitations and characteristics so as to utilize the material in a manner consistent with its strengths and weaknesses

  13. Contactless Growth of ZnSe Single Crystals by Physical Vapor Transport

    NASA Technical Reports Server (NTRS)

    Su, Ching-Hua; George, M. A.; Feth, S.; Lehoczky, S. L.

    1998-01-01

    ZnSe crystals were grown by self-seeded physical vapor transport (PVT) technique in the horizontal configuration. The source materials were heat treated by H2 reduction to remove the oxide followed by baking under dynamic vacuum to adjust the source composition toward that of congruent sublimation. Contactless growth of ZnSe single crystals have been performed consistently using three different source materials. The crystals grew away from the wall during the later stage of the growth with large (110) facets tend to align parallel to the gravity direction. The Scanning Electron Micrography (SEM) micrographs and the Atomic Force Microscope (AFM) images showed that large (110) terraces and steps dominate the as-grown facets. The measured residual gas pressures in the processed ampoules agree well among various source materials and the major components were CO and H2. No preferred growth direction was found. The one-dimensional diffusion model on the mass flux of a multi-species PVT system was employed to analyze the conditions for contactless growth. The calculated thermal profile for supersaturation is very close to the thermal profile measured inside the empty furnace bore in the region of contactless growth. The effects of convective flows in the vapor phase inside the ampoule on the growth processes are discussed.

  14. Estimation of sheath potentials in front of ASDEX upgrade ICRF antenna with SSWICH asymptotic code

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Křivská, A., E-mail: alena.krivska@rma.ac.be; Bobkov, V.; Jacquot, J.

    Multi-megawatt Ion Cyclotron Range of Frequencies (ICRF) heating became problematic in ASDEX Upgrade (AUG) tokamak after coating of ICRF antenna limiters and other plasma facing components by tungsten. Strong impurity influx was indeed produced at levels of injected power markedly lower than in the previous experiments. It is assumed that the impurity production is mainly driven by parallel component of Radio-Frequency (RF) antenna electric near-field E// that is rectified in sheaths. In this contribution we estimate poloidal distribution of sheath Direct Current (DC) potential in front of the ICRF antenna and simulate its relative variations over the parametric scans performedmore » during experiments, trying to reproduce some of the experimental observations. In addition, relative comparison between two types of AUG ICRF antenna configurations, used for experiments in 2014, has been performed. For this purpose we use the Torino Polytechnic Ion Cyclotron Antenna (TOPICA) code and asymptotic version of the Self-consistent Sheaths and Waves for Ion Cyclotron Heating (SSWICH) code. Further, we investigate correlation between amplitudes of the calculated oscillating sheath voltages and the E// fields computed at the lateral side of the antenna box, in relation with a heuristic antenna design strategy at IPP Garching to mitigate RF sheaths.« less

  15. Parallel Work of CO2 Ejectors Installed in a Multi-Ejector Module of Refrigeration System

    NASA Astrophysics Data System (ADS)

    Bodys, Jakub; Palacz, Michal; Haida, Michal; Smolka, Jacek; Nowak, Andrzej J.; Banasiak, Krzysztof; Hafner, Armin

    2016-09-01

    A performance analysis on of fixed ejectors installed in a multi-ejector module in a CO2 refrigeration system is presented in this study. The serial and the parallel work of four fixed-geometry units that compose the multi-ejector pack was carried out. The executed numerical simulations were performed with the use of validated Homogeneous Equilibrium Model (HEM). The computational tool ejectorPL for typical transcritical parameters at the motive nozzle were used in all the tests. A wide range of the operating conditions for supermarket applications in three different European climate zones were taken into consideration. The obtained results present the high and stable performance of all the ejectors in the multi-ejector pack.

  16. Development and study of a parallel algorithm of iteratively forming latent functionally-determined structures for classification and analysis of meteorological data

    NASA Astrophysics Data System (ADS)

    Sorokin, V. A.; Volkov, Yu V.; Sherstneva, A. I.; Botygin, I. A.

    2016-11-01

    This paper overviews a method of generating climate regions based on an analytic signal theory. When applied to atmospheric surface layer temperature data sets, the method allows forming climatic structures with the corresponding changes in the temperature to make conclusions on the uniformity of climate in an area and to trace the climate changes in time by analyzing the type group shifts. The algorithm is based on the fact that the frequency spectrum of the thermal oscillation process is narrow-banded and has only one mode for most weather stations. This allows using the analytic signal theory, causality conditions and introducing an oscillation phase. The annual component of the phase, being a linear function, was removed by the least squares method. The remaining phase fluctuations allow consistent studying of their coordinated behavior and timing, using the Pearson correlation coefficient for dependence evaluation. This study includes program experiments to evaluate the calculation efficiency in the phase grouping task. The paper also overviews some single-threaded and multi-threaded computing models. It is shown that the phase grouping algorithm for meteorological data can be parallelized and that a multi-threaded implementation leads to a 25-30% increase in the performance.

  17. Space Technology 5 Multi-Point Observations of Temporal Variability of Field-Aligned Currents

    NASA Technical Reports Server (NTRS)

    Le, Guan; Wang, Yongli; Slavin, James A.; Strangeway, Robert J.

    2008-01-01

    Space Technology 5 (ST5) is a three micro-satellite constellation deployed into a 300 x 4500 km, dawn-dusk, sun-synchronous polar orbit from March 22 to June 21, 2006, for technology validations. In this paper, we present a study of the temporal variability of field-aligned currents using multi-point magnetic field measurements from ST5. The data demonstrate that meso-scale current structures are commonly embedded within large-scale field-aligned current sheets. The meso-scale current structures are very dynamic with highly variable current density and/or polarity in time scales of approximately 10 min. They exhibit large temporal variations during both quiet and disturbed times in such time scales. On the other hand, the data also shown that the time scales for the currents to be relatively stable are approximately 1 min for meso-scale currents and approximately 10 min for large scale current sheets. These temporal features are obviously associated with dynamic variations of their particle carriers (mainly electrons) as they respond to the variations of the parallel electric field in auroral acceleration region. The characteristic time scales for the temporal variability of meso-scale field-aligned currents are found to be consistent with those of auroral parallel electric field.

  18. Space Technology 5 Multi-point Observations of Field-aligned Currents: Temporal Variability of Meso-Scale Structures

    NASA Technical Reports Server (NTRS)

    Le, Guan; Wang, Yongli; Slavin, James A.; Strangeway, Robert J.

    2007-01-01

    Space Technology 5 (ST5) is a three micro-satellite constellation deployed into a 300 x 4500 km, dawn-dusk, sun-synchronous polar orbit from March 22 to June 21, 2006, for technology validations. In this paper, we present a study of the temporal variability of field-aligned currents using multi-point magnetic field measurements from ST5. The data demonstrate that meso-scale current structures are commonly embedded within large-scale field-aligned current sheets. The meso-scale current structures are very dynamic with highly variable current density and/or polarity in time scales of - 10 min. They exhibit large temporal variations during both quiet and disturbed times in such time scales. On the other hand, the data also shown that the time scales for the currents to be relatively stable are approx. 1 min for meso-scale currents and approx. 10 min for large scale current sheets. These temporal features are obviously associated with dynamic variations of their particle carriers (mainly electrons) as they respond to the variations of the parallel electric field in auroral acceleration region. The characteristic time scales for the temporal variability of meso-scale field-aligned currents are found to be consistent with those of auroral parallel electric field.

  19. Device for sectioning prostatectomy specimens to facilitate comparison between histology and in vivo MRI

    PubMed Central

    Drew, Bryn; Jones, Edward C.; Reinsberg, Stefan; Yung, Andrew C.; Goldenberg, S. Larry; Kozlowski, Piotr

    2012-01-01

    Purpose To develop a device for sectioning prostatectomy specimens that would facilitate comparison between histology and in vivo MRI. Materials and methods A multi-bladed cutting device was developed, which consists of an adjustable box capable of accommodating a prostatectomy specimen up to 85 mm in size in the lateral direction, a “plunger” tool to press on the excised gland from the top to prevent it from rolling or sliding during sectioning, and a multi-bladed knife assembly capable of holding up to 21 blades at 4 mm intervals. The device was tested on a formalin fixed piece of meat and subsequently used to section a prostatectomy specimen. Histology sections were compared with T2-weighted MR images acquired in vivo prior to the prostatectomy procedure. Results The prostatectomy specimen slices were very uniform in thickness with each face parallel to the other with no visible sawing marks on the sections by the blades after the cut. MRI and histology comparison showed good correspondence between the two images. Conclusion The developed device allows sectioning of prostatectomy specimens into parallel cuts at a specific orientation and fixed intervals. Such a device is useful in facilitating accurate correlation between histology and MRI data. PMID:20882632

  20. Parallel and Preemptable Dynamically Dimensioned Search Algorithms for Single and Multi-objective Optimization in Water Resources

    NASA Astrophysics Data System (ADS)

    Tolson, B.; Matott, L. S.; Gaffoor, T. A.; Asadzadeh, M.; Shafii, M.; Pomorski, P.; Xu, X.; Jahanpour, M.; Razavi, S.; Haghnegahdar, A.; Craig, J. R.

    2015-12-01

    We introduce asynchronous parallel implementations of the Dynamically Dimensioned Search (DDS) family of algorithms including DDS, discrete DDS, PA-DDS and DDS-AU. These parallel algorithms are unique from most existing parallel optimization algorithms in the water resources field in that parallel DDS is asynchronous and does not require an entire population (set of candidate solutions) to be evaluated before generating and then sending a new candidate solution for evaluation. One key advance in this study is developing the first parallel PA-DDS multi-objective optimization algorithm. The other key advance is enhancing the computational efficiency of solving optimization problems (such as model calibration) by combining a parallel optimization algorithm with the deterministic model pre-emption concept. These two efficiency techniques can only be combined because of the asynchronous nature of parallel DDS. Model pre-emption functions to terminate simulation model runs early, prior to completely simulating the model calibration period for example, when intermediate results indicate the candidate solution is so poor that it will definitely have no influence on the generation of further candidate solutions. The computational savings of deterministic model preemption available in serial implementations of population-based algorithms (e.g., PSO) disappear in synchronous parallel implementations as these algorithms. In addition to the key advances above, we implement the algorithms across a range of computation platforms (Windows and Unix-based operating systems from multi-core desktops to a supercomputer system) and package these for future modellers within a model-independent calibration software package called Ostrich as well as MATLAB versions. Results across multiple platforms and multiple case studies (from 4 to 64 processors) demonstrate the vast improvement over serial DDS-based algorithms and highlight the important role model pre-emption plays in the performance of parallel, pre-emptable DDS algorithms. Case studies include single- and multiple-objective optimization problems in water resources model calibration and in many cases linear or near linear speedups are observed.

  1. Experiments with a Parallel Multi-Objective Evolutionary Algorithm for Scheduling

    NASA Technical Reports Server (NTRS)

    Brown, Matthew; Johnston, Mark D.

    2013-01-01

    Evolutionary multi-objective algorithms have great potential for scheduling in those situations where tradeoffs among competing objectives represent a key requirement. One challenge, however, is runtime performance, as a consequence of evolving not just a single schedule, but an entire population, while attempting to sample the Pareto frontier as accurately and uniformly as possible. The growing availability of multi-core processors in end user workstations, and even laptops, has raised the question of the extent to which such hardware can be used to speed up evolutionary algorithms. In this paper we report on early experiments in parallelizing a Generalized Differential Evolution (GDE) algorithm for scheduling long-range activities on NASA's Deep Space Network. Initial results show that significant speedups can be achieved, but that performance does not necessarily improve as more cores are utilized. We describe our preliminary results and some initial suggestions from parallelizing the GDE algorithm. Directions for future work are outlined.

  2. Performance of multi-hop parallel free-space optical communication over gamma-gamma fading channel with pointing errors.

    PubMed

    Gao, Zhengguang; Liu, Hongzhan; Ma, Xiaoping; Lu, Wei

    2016-11-10

    Multi-hop parallel relaying is considered in a free-space optical (FSO) communication system deploying binary phase-shift keying (BPSK) modulation under the combined effects of a gamma-gamma (GG) distribution and misalignment fading. Based on the best path selection criterion, the cumulative distribution function (CDF) of this cooperative random variable is derived. Then the performance of this optical mesh network is analyzed in detail. A Monte Carlo simulation is also conducted to demonstrate the effectiveness of the results for the average bit error rate (ABER) and outage probability. The numerical result proves that it needs a smaller average transmitted optical power to achieve the same ABER and outage probability when using the multi-hop parallel network in FSO links. Furthermore, the system use of more number of hops and cooperative paths can improve the quality of the communication.

  3. Multi-channel temperature measurement system for automotive battery stack

    NASA Astrophysics Data System (ADS)

    Lewczuk, Radoslaw; Wojtkowski, Wojciech

    2017-08-01

    A multi-channel temperature measurement system for monitoring of automotive battery stack is presented in the paper. The presented system is a complete battery temperature measuring system for hybrid / electric vehicles that incorporates multi-channel temperature measurements with digital temperature sensors communicating through 1-Wire buses, individual 1-Wire bus for each sensor for parallel computing (parallel measurements instead of sequential), FPGA device which collects data from sensors and translates it for CAN bus frames. CAN bus is incorporated for communication with car Battery Management System and uses additional CAN bus controller which communicates with FPGA device through SPI bus. The described system can parallel measure up to 12 temperatures but can be easily extended in the future in case of additional needs. The structure of the system as well as particular devices are described in the paper. Selected results of experimental investigations which show proper operation of the system are presented as well.

  4. A Parallel, Multi-Scale Watershed-Hydrologic-Inundation Model with Adaptively Switching Mesh for Capturing Flooding and Lake Dynamics

    NASA Astrophysics Data System (ADS)

    Ji, X.; Shen, C.

    2017-12-01

    Flood inundation presents substantial societal hazards and also changes biogeochemistry for systems like the Amazon. It is often expensive to simulate high-resolution flood inundation and propagation in a long-term watershed-scale model. Due to the Courant-Friedrichs-Lewy (CFL) restriction, high resolution and large local flow velocity both demand prohibitively small time steps even for parallel codes. Here we develop a parallel surface-subsurface process-based model enhanced by multi-resolution meshes that are adaptively switched on or off. The high-resolution overland flow meshes are enabled only when the flood wave invades to floodplains. This model applies semi-implicit, semi-Lagrangian (SISL) scheme in solving dynamic wave equations, and with the assistant of the multi-mesh method, it also adaptively chooses the dynamic wave equation only in the area of deep inundation. Therefore, the model achieves a balance between accuracy and computational cost.

  5. Fast data reconstructed method of Fourier transform imaging spectrometer based on multi-core CPU

    NASA Astrophysics Data System (ADS)

    Yu, Chunchao; Du, Debiao; Xia, Zongze; Song, Li; Zheng, Weijian; Yan, Min; Lei, Zhenggang

    2017-10-01

    Imaging spectrometer can gain two-dimensional space image and one-dimensional spectrum at the same time, which shows high utility in color and spectral measurements, the true color image synthesis, military reconnaissance and so on. In order to realize the fast reconstructed processing of the Fourier transform imaging spectrometer data, the paper designed the optimization reconstructed algorithm with OpenMP parallel calculating technology, which was further used for the optimization process for the HyperSpectral Imager of `HJ-1' Chinese satellite. The results show that the method based on multi-core parallel computing technology can control the multi-core CPU hardware resources competently and significantly enhance the calculation of the spectrum reconstruction processing efficiency. If the technology is applied to more cores workstation in parallel computing, it will be possible to complete Fourier transform imaging spectrometer real-time data processing with a single computer.

  6. Parallelized computation for computer simulation of electrocardiograms using personal computers with multi-core CPU and general-purpose GPU.

    PubMed

    Shen, Wenfeng; Wei, Daming; Xu, Weimin; Zhu, Xin; Yuan, Shizhong

    2010-10-01

    Biological computations like electrocardiological modelling and simulation usually require high-performance computing environments. This paper introduces an implementation of parallel computation for computer simulation of electrocardiograms (ECGs) in a personal computer environment with an Intel CPU of Core (TM) 2 Quad Q6600 and a GPU of Geforce 8800GT, with software support by OpenMP and CUDA. It was tested in three parallelization device setups: (a) a four-core CPU without a general-purpose GPU, (b) a general-purpose GPU plus 1 core of CPU, and (c) a four-core CPU plus a general-purpose GPU. To effectively take advantage of a multi-core CPU and a general-purpose GPU, an algorithm based on load-prediction dynamic scheduling was developed and applied to setting (c). In the simulation with 1600 time steps, the speedup of the parallel computation as compared to the serial computation was 3.9 in setting (a), 16.8 in setting (b), and 20.0 in setting (c). This study demonstrates that a current PC with a multi-core CPU and a general-purpose GPU provides a good environment for parallel computations in biological modelling and simulation studies. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.

  7. Self assembled multi-layer nanocomposite of graphene and metal oxide materials

    DOEpatents

    Liu, Jun; Aksay, Ilhan A; Choi, Daiwon; Kou, Rong; Nie, Zimin; Wang, Donghai; Yang, Zhenguo

    2013-10-22

    Nanocomposite materials having at least two layers, each layer consisting of one metal oxide bonded to at least one graphene layer were developed. The nanocomposite materials will typically have many alternating layers of metal oxides and graphene layers, bonded in a sandwich type construction and will be incorporated into an electrochemical or energy storage device.

  8. Self assembled multi-layer nanocomposite of graphene and metal oxide materials

    DOEpatents

    Liu, Jun; Aksay, Ilhan A; Choi, Daiwon; Kou, Rong; Nie, Zimin; Wang, Donghai; Yang, Zhenguo

    2015-04-28

    Nanocomposite materials having at least two layers, each layer consisting of one metal oxide bonded to at least one graphene layer were developed. The nanocomposite materials will typically have many alternating layers of metal oxides and graphene layers, bonded in a sandwich type construction and will be incorporated into an electrochemical or energy storage device.

  9. Self assembled multi-layer nanocomposite of graphene and metal oxide materials

    DOEpatents

    Liu, Jun; Choi, Daiwon; Kou, Rong; Nie, Zimin; Wang, Donghai; Yang, Zhenguo

    2014-09-16

    Nanocomposite materials having at least two layers, each layer consisting of one metal oxide bonded to at least one graphene layer were developed. The nanocomposite materials will typically have many alternating layers of metal oxides and graphene layers, bonded in a sandwich type construction and will be incorporated into an electrochemical or energy storage device.

  10. A reconfigurable multicarrier demodulator architecture

    NASA Technical Reports Server (NTRS)

    Kwatra, S. C.; Jamali, M. M.

    1991-01-01

    An architecture based on parallel and pipline design approaches has been developed for the Frequency Division Multiple Access/Time Domain Multiplexed (FDMA/TDM) conversion system. The architecture has two main modules namely the transmultiplexer and the demodulator. The transmultiplexer has two pipelined modules. These are the shared multiplexed polyphase filter and the Fast Fourier Transform (FFT). The demodulator consists of carrier, clock, and data recovery modules which are interactive. Progress on the design of the MultiCarrier Demodulator (MCD) using commercially available chips and Application Specific Integrated Circuits (ASIC) and simulation studies using Viewlogic software will be presented at the conference.

  11. An object-oriented approach for parallel self adaptive mesh refinement on block structured grids

    NASA Technical Reports Server (NTRS)

    Lemke, Max; Witsch, Kristian; Quinlan, Daniel

    1993-01-01

    Self-adaptive mesh refinement dynamically matches the computational demands of a solver for partial differential equations to the activity in the application's domain. In this paper we present two C++ class libraries, P++ and AMR++, which significantly simplify the development of sophisticated adaptive mesh refinement codes on (massively) parallel distributed memory architectures. The development is based on our previous research in this area. The C++ class libraries provide abstractions to separate the issues of developing parallel adaptive mesh refinement applications into those of parallelism, abstracted by P++, and adaptive mesh refinement, abstracted by AMR++. P++ is a parallel array class library to permit efficient development of architecture independent codes for structured grid applications, and AMR++ provides support for self-adaptive mesh refinement on block-structured grids of rectangular non-overlapping blocks. Using these libraries, the application programmers' work is greatly simplified to primarily specifying the serial single grid application and obtaining the parallel and self-adaptive mesh refinement code with minimal effort. Initial results for simple singular perturbation problems solved by self-adaptive multilevel techniques (FAC, AFAC), being implemented on the basis of prototypes of the P++/AMR++ environment, are presented. Singular perturbation problems frequently arise in large applications, e.g. in the area of computational fluid dynamics. They usually have solutions with layers which require adaptive mesh refinement and fast basic solvers in order to be resolved efficiently.

  12. Massively parallel algorithm and implementation of RI-MP2 energy calculation for peta-scale many-core supercomputers.

    PubMed

    Katouda, Michio; Naruse, Akira; Hirano, Yukihiko; Nakajima, Takahito

    2016-11-15

    A new parallel algorithm and its implementation for the RI-MP2 energy calculation utilizing peta-flop-class many-core supercomputers are presented. Some improvements from the previous algorithm (J. Chem. Theory Comput. 2013, 9, 5373) have been performed: (1) a dual-level hierarchical parallelization scheme that enables the use of more than 10,000 Message Passing Interface (MPI) processes and (2) a new data communication scheme that reduces network communication overhead. A multi-node and multi-GPU implementation of the present algorithm is presented for calculations on a central processing unit (CPU)/graphics processing unit (GPU) hybrid supercomputer. Benchmark results of the new algorithm and its implementation using the K computer (CPU clustering system) and TSUBAME 2.5 (CPU/GPU hybrid system) demonstrate high efficiency. The peak performance of 3.1 PFLOPS is attained using 80,199 nodes of the K computer. The peak performance of the multi-node and multi-GPU implementation is 514 TFLOPS using 1349 nodes and 4047 GPUs of TSUBAME 2.5. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.

  13. Process-to-Panel Modeling and Multiprobe Characterization of Silicon Heterojunction Solar Cell Technology

    NASA Astrophysics Data System (ADS)

    Chavali, Raghu Vamsi Krishna

    The large-scale deployment of PV technology is very sensitive to the material and process costs. There are several potential candidates among p-n heterojunction (HJ) solar cells competing for higher efficiencies at lower material and process costs. These systems are, however, generally complex, involve diverse materials, and are not well understood. The direct translation of classical p-n homojunction theory to p-n HJ cells may not always be self-consistent and can lead, therefore, to misinterpretation of experimental results. Ultimately, this translation may not be useful for modeling and characterization of these solar cells. Hence, there is a strong need to redefine/reinterpret the modeling/characterization methodologies for HJ solar cells to produce a self-consistent framework for optimizing HJ solar cell designs. Towards this goal, we explore the physics and interpret characterization experiments of p-n HJs using Silicon HJ (HIT) solar cells. We will: (1) identify the key HJ properties that affect the cell efficiency; (2) analyze the dependence of key HJ properties on the carrier transport under light and dark conditions; (3) provide a selfconsistent multi-probe approach to extract the HJ parameters using several characterization techniques including dark I-V, light I-V, C-V, impedance spectroscopy, and Suns-Voc; (4) propose design guidelines to address the HJ bottlenecks of HIT cells; and (5) develop a process-to-module modeling framework to establish the module performance limits. The guidelines resulting from this multi-scale and self-consistent framework can be used to improve performance of HIT cells as well as other HJ based solar cells.

  14. Coupling of free space sub-terahertz waves into dielectric slabs using PC waveguides.

    PubMed

    Ghattan, Z; Hasek, T; Shahabadi, M; Koch, M

    2008-04-28

    The paper presents theoretical and experimental results on photonic crystal structures which work under the self-collimation condition to couple free space waves into dielectric slabs in the sub-terahertz range. Using a standard machining process, two-dimensional photonic crystal structures consisting of a square array of air holes in the dielectric medium are fabricated. One of the structures has two adjacent parallel line-defects that improve the coupling efficiency. This leads to a combination of self-collimation and directional emission of electromagnetic waves. The experimental results are in good agreement with those of the Finite- Element-Method calculations. Experimentally we achieve a coupling efficiency of 63%.

  15. The Children's Report of Sleep Patterns: Validity and Reliability of the Sleep Hygiene Index and Sleep Disturbances Scale in Adolescents

    PubMed Central

    Meltzer, Lisa J.; Brimeyer, Chasity; Russell, Kathryn; Avis, Kristin T.; Biggs, Sarah; Reynolds, Amy C.; Crabtree, Valerie McLaughlin

    2014-01-01

    Objective Sleep is critical for adolescent health and well-being. However, there are a limited number of validated self-report measures of sleep for adolescents, and no well-validated measures of sleep that can be used across middle childhood and adolescence. The Children's Report of Sleep Patterns (CRSP) has already been validated in children ages 8-12 years. The purpose of this study was to examine the psychometric properties of the CRSP as a multidimensional, self-report sleep measure for adolescents. Methods Participants included 570 adolescents aged 13 – 18 years, 60% female, recruited from pediatricians’ offices, sleep clinics, children's hospitals, schools, and the general population. A multi-method, multi-reporter approach was used to validate the CRSP. Along with the CRSP, a subset of the sample completed the Adolescent Sleep Hygiene Scale (ASHS), with a different subset of adolescents undergoing polysomnography. Results The CRSP demonstrated good reliability and validity. Group differences on the CRSP were found for adolescents presenting to a sleep or medical clinic (vs. community sample), for older adolescents (vs. younger adolescents), those who regularly napped (vs. infrequently napped), and those with poor sleep quality (vs. good sleep quality). Self-reported sleep quality in adolescents was also associated with higher apnea-hypopnea index scores from polysomnography. Finally, the CRSP Sleep Hygiene Indices were significantly correlated with indices of the ASHS. Conclusions The CRSP is a valid and reliable measure of adolescent sleep hygiene and sleep disturbances. With a parallel version for middle childhood (8-12 years), the CRSP likely provides clinicians and researchers the ability to measure self-reported sleep across development. PMID:25441749

  16. Particle simulation of plasmas on the massively parallel processor

    NASA Technical Reports Server (NTRS)

    Gledhill, I. M. A.; Storey, L. R. O.

    1987-01-01

    Particle simulations, in which collective phenomena in plasmas are studied by following the self consistent motions of many discrete particles, involve several highly repetitive sets of calculations that are readily adaptable to SIMD parallel processing. A fully electromagnetic, relativistic plasma simulation for the massively parallel processor is described. The particle motions are followed in 2 1/2 dimensions on a 128 x 128 grid, with periodic boundary conditions. The two dimensional simulation space is mapped directly onto the processor network; a Fast Fourier Transform is used to solve the field equations. Particle data are stored according to an Eulerian scheme, i.e., the information associated with each particle is moved from one local memory to another as the particle moves across the spatial grid. The method is applied to the study of the nonlinear development of the whistler instability in a magnetospheric plasma model, with an anisotropic electron temperature. The wave distribution function is included as a new diagnostic to allow simulation results to be compared with satellite observations.

  17. Quantum supercharger library: hyper-parallelism of the Hartree-Fock method.

    PubMed

    Fernandes, Kyle D; Renison, C Alicia; Naidoo, Kevin J

    2015-07-05

    We present here a set of algorithms that completely rewrites the Hartree-Fock (HF) computations common to many legacy electronic structure packages (such as GAMESS-US, GAMESS-UK, and NWChem) into a massively parallel compute scheme that takes advantage of hardware accelerators such as Graphical Processing Units (GPUs). The HF compute algorithm is core to a library of routines that we name the Quantum Supercharger Library (QSL). We briefly evaluate the QSL's performance and report that it accelerates a HF 6-31G Self-Consistent Field (SCF) computation by up to 20 times for medium sized molecules (such as a buckyball) when compared with mature Central Processing Unit algorithms available in the legacy codes in regular use by researchers. It achieves this acceleration by massive parallelization of the one- and two-electron integrals and optimization of the SCF and Direct Inversion in the Iterative Subspace routines through the use of GPU linear algebra libraries. © 2015 Wiley Periodicals, Inc. © 2015 Wiley Periodicals, Inc.

  18. Depressive Symptoms and Conversational Self-Focus in Adolescents’ Friendships

    PubMed Central

    Schwartz-Mette, Rebecca A.; Rose, Amanda J.

    2015-01-01

    This multi-method, longitudinal study considered the interplay among depressive symptoms, aversive interpersonal behavior, and interpersonal rejection in early and middle adolescents’ friendships. In particular, the study examined a newly identified interpersonal process, conversational self-focus (i.e., the tendency to redirect conversations about problems to focus on the self). Traditional interpersonal theories of depression suggest that individuals with depressive symptoms engage in aversive behaviors (such as conversational self-focus) and are rejected by others. However, in the current study, not all adolescents with depressive symptoms engaged in conversational self-focus and were rejected by friends. Instead, conversational self-focus moderated prospective relations of depressive symptoms and later friendship problems such that only adolescents with depressive symptoms who engaged in conversational self-focus were rejected by friends. These findings are consistent with current conceptualizations of the development of psychopathology that highlight heterogeneity among youth who share similar symptoms and the possibility of multifinality of outcomes. PMID:25640911

  19. Multi-Stage Hybrid Rocket Conceptual Design for Micro-Satellites Launch using Genetic Algorithm

    NASA Astrophysics Data System (ADS)

    Kitagawa, Yosuke; Kitagawa, Koki; Nakamiya, Masaki; Kanazaki, Masahiro; Shimada, Toru

    The multi-objective genetic algorithm (MOGA) is applied to the multi-disciplinary conceptual design problem for a three-stage launch vehicle (LV) with a hybrid rocket engine (HRE). MOGA is an optimization tool used for multi-objective problems. The parallel coordinate plot (PCP), which is a data mining method, is employed in the post-process in MOGA for design knowledge discovery. A rocket that can deliver observing micro-satellites to the sun-synchronous orbit (SSO) is designed. It consists of an oxidizer tank containing liquid oxidizer, a combustion chamber containing solid fuel, a pressurizing tank and a nozzle. The objective functions considered in this study are to minimize the total mass of the rocket and to maximize the ratio of the payload mass to the total mass. To calculate the thrust and the engine size, the regression rate is estimated based on an empirical model for a paraffin (FT-0070) propellant. Several non-dominated solutions are obtained using MOGA, and design knowledge is discovered for the present hybrid rocket design problem using a PCP analysis. As a result, substantial knowledge on the design of an LV with an HRE is obtained for use in space transportation.

  20. Parallel Program Systems for the Analysis of Wave Processes in Elastic-Plastic, Granular, Porous and Multi-Blocky Media

    NASA Astrophysics Data System (ADS)

    Sadovskaya, Oxana; Sadovskii, Vladimir

    2017-04-01

    Under modeling the wave propagation processes in geomaterials (granular and porous media, soils and rocks) it is necessary to take into account the structural inhomogeneity of these materials. Parallel program systems for numerical solution of 2D and 3D problems of the dynamics of deformable media with constitutive relationships of rather general form on the basis of universal mathematical model describing small strains of elastic, elastic-plastic, granular and porous materials are worked out. In the case of an elastic material, the model is reduced to the system of equations, hyperbolic by Friedrichs, written in terms of velocities and stresses in a symmetric form. In the case of an elastic-plastic material, the model is a special formulation of the Prandtl-Reuss theory in the form of variational inequality with one-sided constraints on the stress tensor. Generalization of the model to describe granularity and the collapse of pores is obtained by means of the rheological approach, taking into account different resistance of materials to tension and compression. Rotational motion of particles in the material microstructure is considered within the framework of a mathematical model of the Cosserat continuum. Computational domain may have a blocky structure, composed of an arbitrary number of layers, strips in a layer and blocks in a strip from different materials with self-consistent curvilinear interfaces. At the external boundaries of computational domain the main types of dissipative boundary conditions in terms of velocities, stresses or mixed boundary conditions can be given. Shock-capturing algorithm is proposed for implementation of the model on supercomputers with cluster architecture. It is based on the two-cyclic splitting method with respect to spatial variables and the special procedures of the stresses correction to take into account plasticity, granularity or porosity of a material. An explicit monotone ENO-scheme is applied for solving one-dimensional systems of equations at the stages of splitting method. The parallelizing of computations is carried out using the MPI library and the SPMD technology. The data exchange between processors occurs at step "predictor" of the finite-difference scheme. Program systems allow simulate the propagation of waves produced by external mechanical effects in a medium, aggregated of arbitrary number of heterogeneous blocks. Some computations of dynamic problems with and without taking into account the moment properties of a material were performed on clusters of ICM SB RAS (Krasnoyarsk) and JSCC RAS (Moscow). Parallel program systems 2Dyn_Granular, 3Dyn_Granular, 2Dyn_Cosserat, 3Dyn_Cosserat and 2Dyn_Blocks_MPI for numerical solution of 2D and 3D elastic-plastic problems of the dynamics of granular media and problems of the Cosserat elasticity theory, as well as for modeling of the dynamic processes in multi-blocky media with pliant viscoelastic, porous and fluid-saturated interlayers on cluster systems were registered by Rospatent.

  1. Hybrid Parallelization of Adaptive MHD-Kinetic Module in Multi-Scale Fluid-Kinetic Simulation Suite

    DOE PAGES

    Borovikov, Sergey; Heerikhuisen, Jacob; Pogorelov, Nikolai

    2013-04-01

    The Multi-Scale Fluid-Kinetic Simulation Suite has a computational tool set for solving partially ionized flows. In this paper we focus on recent developments of the kinetic module which solves the Boltzmann equation using the Monte-Carlo method. The module has been recently redesigned to utilize intra-node hybrid parallelization. We describe in detail the redesign process, implementation issues, and modifications made to the code. Finally, we conduct a performance analysis.

  2. Self-Control and Emotional and Verbal Aggression in Dating Relationships: A Dyadic Understanding.

    PubMed

    Baker, Elizabeth A; Klipfel, Katherine M; van Dulmen, Manfred H M

    2016-08-01

    Guided by the dynamic developmental systems perspective, this study extends past research by examining the association between self-control and emotional and verbal aggression (EVA) using a dyadic multi-method design. Guided by empirical research and the dynamic developmental systems perspective, we hypothesized that (a) there would be a negative association between one's own self-control and one's own perpetration of EVA and (b) there would also be a negative association between one's partner's self-control and one's own perpetration of EVA. One hundred twenty heterosexual dating couples (ages 18-25 years) provided data on self-control (Grasmick et al.'s Low Self-Control Scale; reverse scored for ease of interpretation), self-reported perpetration of EVA (Emotional and Verbal Abuse subscale of the Conflict in Adolescent Dating Relationships Inventory), and observationally assessed perpetration of EVA. Data were analyzed using path analyses within the Actor-Partner Interdependence Model (APIM) framework. Consistent with previous findings, we found that self-control was negatively associated with the perpetration of EVA. Furthermore, we found partner effects, such that female-but not male-self-control predicted partner-observed perpetration of EVA. These findings highlight the importance of examining risk factors for EVA of both partners. Our findings also suggest that the association between self-control and EVA is partially a function of whether EVA is assessed through self-report or observational methodology. This highlights the need to conduct multi-method assessments in future research. As discussed in the article, our findings have implications for theories on intimate partner violence, study designs, and couple interventions.

  3. Work organization in hospital wards and nurses' emotional exhaustion: A multi-method study of observation-based assessment and nurses' self-reports.

    PubMed

    Stab, Nicole; Hacker, Winfried; Weigl, Matthias

    2016-09-01

    Ward organization is a major determinant for nurses' well-being on the job. The majority of previous research on this relationship is based on single source methods, which have been criticized as skewed estimations mainly due to subjectivity of the ratings and due to common source bias. To investigate the association of ward organization characteristics and nurses' exhaustion by combining observation-based assessments with nurses' self-reports. Cross-sectional study on 25 wards of four hospitals and 245 nurses. Our multi-method approach to evaluate hospital ward organization consisted of on-site observations with a standardized assessment tool and of questionnaires to evaluate nurses' self-reports and exhaustion. After establishing the reliability of our measures, we applied multi-level regression analyses to determine associations between determinant and outcome variables. We found substantial convergence in ward organization between the observation-based assessments and nurses' self-reports, which supports the validity of our external assessments. Furthermore, two observation-based characteristics, namely participation and patient-focused care, were significantly associated with lower emotional exhaustion among the nurses. Our results suggest that observation-based assessments are a valid and feasible way to assess ward organization in hospitals. Nurses' self-reported as well as observation-based ratings on ward organization were associated with nurses' emotional exhaustion. This is of interest mainly for identifying alternative measures in evaluating nurses' work environments, to inform health promotion activities and to evaluate job redesign intervention. Copyright © 2016 Elsevier Ltd. All rights reserved.

  4. Self-consistent Non-LTE Model of Infrared Molecular Emissions and Oxygen Dayglows in the Mesosphere and Lower Thermosphere

    NASA Technical Reports Server (NTRS)

    Feofilov, Artem G.; Yankovsky, Valentine A.; Pesnell, William D.; Kutepov, Alexander A.; Goldberg, Richard A.; Mauilova, Rada O.

    2007-01-01

    We present the new version of the ALI-ARMS (for Accelerated Lambda Iterations for Atmospheric Radiation and Molecular Spectra) model. The model allows simultaneous self-consistent calculating the non-LTE populations of the electronic-vibrational levels of the O3 and O2 photolysis products and vibrational level populations of CO2, N2,O2, O3, H2O, CO and other molecules with detailed accounting for the variety of the electronic-vibrational, vibrational-vibrational and vibrational-translational energy exchange processes. The model was used as the reference one for modeling the O2 dayglows and infrared molecular emissions for self-consistent diagnostics of the multi-channel space observations of MLT in the SABER experiment It also allows reevaluating the thermalization efficiency of the absorbed solar ultraviolet energy and infrared radiative cooling/heating of MLT by detailed accounting of the electronic-vibrational relaxation of excited photolysis products via the complex chain of collisional energy conversion processes down to the vibrational energy of optically active trace gas molecules.

  5. Array-based Hierarchical Mesh Generation in Parallel

    DOE PAGES

    Ray, Navamita; Grindeanu, Iulian; Zhao, Xinglin; ...

    2015-11-03

    In this paper, we describe an array-based hierarchical mesh generation capability through uniform refinement of unstructured meshes for efficient solution of PDE's using finite element methods and multigrid solvers. A multi-degree, multi-dimensional and multi-level framework is designed to generate the nested hierarchies from an initial mesh that can be used for a number of purposes such as multi-level methods to generating large meshes. The capability is developed under the parallel mesh framework “Mesh Oriented dAtaBase” a.k.a MOAB. We describe the underlying data structures and algorithms to generate such hierarchies and present numerical results for computational efficiency and mesh quality. Inmore » conclusion, we also present results to demonstrate the applicability of the developed capability to a multigrid finite-element solver.« less

  6. Multi-petascale highly efficient parallel supercomputer

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Asaad, Sameh; Bellofatto, Ralph E.; Blocksome, Michael A.

    A Multi-Petascale Highly Efficient Parallel Supercomputer of 100 petaflop-scale includes node architectures based upon System-On-a-Chip technology, where each processing node comprises a single Application Specific Integrated Circuit (ASIC). The ASIC nodes are interconnected by a five dimensional torus network that optimally maximize the throughput of packet communications between nodes and minimize latency. The network implements collective network and a global asynchronous network that provides global barrier and notification functions. Integrated in the node design include a list-based prefetcher. The memory system implements transaction memory, thread level speculation, and multiversioning cache that improves soft error rate at the same time andmore » supports DMA functionality allowing for parallel processing message-passing.« less

  7. A fast ultrasonic simulation tool based on massively parallel implementations

    NASA Astrophysics Data System (ADS)

    Lambert, Jason; Rougeron, Gilles; Lacassagne, Lionel; Chatillon, Sylvain

    2014-02-01

    This paper presents a CIVA optimized ultrasonic inspection simulation tool, which takes benefit of the power of massively parallel architectures: graphical processing units (GPU) and multi-core general purpose processors (GPP). This tool is based on the classical approach used in CIVA: the interaction model is based on Kirchoff, and the ultrasonic field around the defect is computed by the pencil method. The model has been adapted and parallelized for both architectures. At this stage, the configurations addressed by the tool are : multi and mono-element probes, planar specimens made of simple isotropic materials, planar rectangular defects or side drilled holes of small diameter. Validations on the model accuracy and performances measurements are presented.

  8. Multi-Resolution Climate Ensemble Parameter Analysis with Nested Parallel Coordinates Plots.

    PubMed

    Wang, Junpeng; Liu, Xiaotong; Shen, Han-Wei; Lin, Guang

    2017-01-01

    Due to the uncertain nature of weather prediction, climate simulations are usually performed multiple times with different spatial resolutions. The outputs of simulations are multi-resolution spatial temporal ensembles. Each simulation run uses a unique set of values for multiple convective parameters. Distinct parameter settings from different simulation runs in different resolutions constitute a multi-resolution high-dimensional parameter space. Understanding the correlation between the different convective parameters, and establishing a connection between the parameter settings and the ensemble outputs are crucial to domain scientists. The multi-resolution high-dimensional parameter space, however, presents a unique challenge to the existing correlation visualization techniques. We present Nested Parallel Coordinates Plot (NPCP), a new type of parallel coordinates plots that enables visualization of intra-resolution and inter-resolution parameter correlations. With flexible user control, NPCP integrates superimposition, juxtaposition and explicit encodings in a single view for comparative data visualization and analysis. We develop an integrated visual analytics system to help domain scientists understand the connection between multi-resolution convective parameters and the large spatial temporal ensembles. Our system presents intricate climate ensembles with a comprehensive overview and on-demand geographic details. We demonstrate NPCP, along with the climate ensemble visualization system, based on real-world use-cases from our collaborators in computational and predictive science.

  9. Self-consistent analysis of radiation and relativistic electron beam dynamics in a helical wiggler using Lienard-Wiechert fields

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tecimer, M.; Elias, L.R.

    1995-12-31

    Lienard-Wiechert (LW) fields, which are exact solutions of the Wave Equation for a point charge in free space, are employed to formulate a self-consistent treatment of the electron beam dynamics and the evolution of the generated radiation in long undulators. In a relativistic electron beam the internal forces leading to the interaction of the electrons with each other can be computed by means of retarded LW fields. The resulting electron beam dynamics enables us to obtain three dimensional radiation fields starting from an initial incoherent spontaneous emission, without introducing a seed wave at start-up. Based on the formalism employed here,more » both the evolution of the multi-bucket electron phase space dynamics in the beam body as well as edges and the relative slippage of the radiation with respect to the electrons in the considered short bunch are naturally embedded into the simulation model. In this paper, we present electromagnetic radiation studies, including multi-bucket electron phase dynamics and angular distribution of radiation in the time and frequency domain produced by a relativistic short electron beam bunch interacting with a circularly polarized magnetic undulator.« less

  10. Shared Memory Parallelism for 3D Cartesian Discrete Ordinates Solver

    NASA Astrophysics Data System (ADS)

    Moustafa, Salli; Dutka-Malen, Ivan; Plagne, Laurent; Ponçot, Angélique; Ramet, Pierre

    2014-06-01

    This paper describes the design and the performance of DOMINO, a 3D Cartesian SN solver that implements two nested levels of parallelism (multicore+SIMD) on shared memory computation nodes. DOMINO is written in C++, a multi-paradigm programming language that enables the use of powerful and generic parallel programming tools such as Intel TBB and Eigen. These two libraries allow us to combine multi-thread parallelism with vector operations in an efficient and yet portable way. As a result, DOMINO can exploit the full power of modern multi-core processors and is able to tackle very large simulations, that usually require large HPC clusters, using a single computing node. For example, DOMINO solves a 3D full core PWR eigenvalue problem involving 26 energy groups, 288 angular directions (S16), 46 × 106 spatial cells and 1 × 1012 DoFs within 11 hours on a single 32-core SMP node. This represents a sustained performance of 235 GFlops and 40:74% of the SMP node peak performance for the DOMINO sweep implementation. The very high Flops/Watt ratio of DOMINO makes it a very interesting building block for a future many-nodes nuclear simulation tool.

  11. Fast parallel algorithm for slicing STL based on pipeline

    NASA Astrophysics Data System (ADS)

    Ma, Xulong; Lin, Feng; Yao, Bo

    2016-05-01

    In Additive Manufacturing field, the current researches of data processing mainly focus on a slicing process of large STL files or complicated CAD models. To improve the efficiency and reduce the slicing time, a parallel algorithm has great advantages. However, traditional algorithms can't make full use of multi-core CPU hardware resources. In the paper, a fast parallel algorithm is presented to speed up data processing. A pipeline mode is adopted to design the parallel algorithm. And the complexity of the pipeline algorithm is analyzed theoretically. To evaluate the performance of the new algorithm, effects of threads number and layers number are investigated by a serial of experiments. The experimental results show that the threads number and layers number are two remarkable factors to the speedup ratio. The tendency of speedup versus threads number reveals a positive relationship which greatly agrees with the Amdahl's law, and the tendency of speedup versus layers number also keeps a positive relationship agreeing with Gustafson's law. The new algorithm uses topological information to compute contours with a parallel method of speedup. Another parallel algorithm based on data parallel is used in experiments to show that pipeline parallel mode is more efficient. A case study at last shows a suspending performance of the new parallel algorithm. Compared with the serial slicing algorithm, the new pipeline parallel algorithm can make full use of the multi-core CPU hardware, accelerate the slicing process, and compared with the data parallel slicing algorithm, the new slicing algorithm in this paper adopts a pipeline parallel model, and a much higher speedup ratio and efficiency is achieved.

  12. Implementing the PM Programming Language using MPI and OpenMP - a New Tool for Programming Geophysical Models on Parallel Systems

    NASA Astrophysics Data System (ADS)

    Bellerby, Tim

    2015-04-01

    PM (Parallel Models) is a new parallel programming language specifically designed for writing environmental and geophysical models. The language is intended to enable implementers to concentrate on the science behind the model rather than the details of running on parallel hardware. At the same time PM leaves the programmer in control - all parallelisation is explicit and the parallel structure of any given program may be deduced directly from the code. This paper describes a PM implementation based on the Message Passing Interface (MPI) and Open Multi-Processing (OpenMP) standards, looking at issues involved with translating the PM parallelisation model to MPI/OpenMP protocols and considering performance in terms of the competing factors of finer-grained parallelisation and increased communication overhead. In order to maximise portability, the implementation stays within the MPI 1.3 standard as much as possible, with MPI-2 MPI-IO file handling the only significant exception. Moreover, it does not assume a thread-safe implementation of MPI. PM adopts a two-tier abstract representation of parallel hardware. A PM processor is a conceptual unit capable of efficiently executing a set of language tasks, with a complete parallel system consisting of an abstract N-dimensional array of such processors. PM processors may map to single cores executing tasks using cooperative multi-tasking, to multiple cores or even to separate processing nodes, efficiently sharing tasks using algorithms such as work stealing. While tasks may move between hardware elements within a PM processor, they may not move between processors without specific programmer intervention. Tasks are assigned to processors using a nested parallelism approach, building on ideas from Reyes et al. (2009). The main program owns all available processors. When the program enters a parallel statement then either processors are divided out among the newly generated tasks (number of new tasks < number of processors) or tasks are divided out among the available processors (number of tasks > number of processors). Nested parallel statements may further subdivide the processor set owned by a given task. Tasks or processors are distributed evenly by default, but uneven distributions are possible under programmer control. It is also possible to explicitly enable child tasks to migrate within the processor set owned by their parent task, reducing load unbalancing at the potential cost of increased inter-processor message traffic. PM incorporates some programming structures from the earlier MIST language presented at a previous EGU General Assembly, while adopting a significantly different underlying parallelisation model and type system. PM code is available at www.pm-lang.org under an unrestrictive MIT license. Reference Ruymán Reyes, Antonio J. Dorta, Francisco Almeida, Francisco de Sande, 2009. Automatic Hybrid MPI+OpenMP Code Generation with llc, Recent Advances in Parallel Virtual Machine and Message Passing Interface, Lecture Notes in Computer Science Volume 5759, 185-195

  13. Self-consistent gyrokinetic modeling of neoclassical and turbulent impurity transport

    NASA Astrophysics Data System (ADS)

    Estève, D.; Sarazin, Y.; Garbet, X.; Grandgirard, V.; Breton, S.; Donnel, P.; Asahi, Y.; Bourdelle, C.; Dif-Pradalier, G.; Ehrlacher, C.; Emeriau, C.; Ghendrih, Ph.; Gillot, C.; Latu, G.; Passeron, C.

    2018-03-01

    Trace impurity transport is studied with the flux-driven gyrokinetic GYSELA code (Grandgirard et al 2016 Comput. Phys. Commun. 207 35). A reduced and linearized multi-species collision operator has been recently implemented, so that both neoclassical and turbulent transport channels can be treated self-consistently on an equal footing. In the Pfirsch-Schlüter regime that is probably relevant for tungsten, the standard expression for the neoclassical impurity flux is shown to be recovered from gyrokinetics with the employed collision operator. Purely neoclassical simulations of deuterium plasma with trace impurities of helium, carbon and tungsten lead to impurity diffusion coefficients, inward pinch velocities due to density peaking, and thermo-diffusion terms which quantitatively agree with neoclassical predictions and NEO simulations (Belli et al 2012 Plasma Phys. Control. Fusion 54 015015). The thermal screening factor appears to be less than predicted analytically in the Pfirsch-Schlüter regime, which can be detrimental to fusion performance. Finally, self-consistent nonlinear simulations have revealed that the tungsten impurity flux is not the sum of turbulent and neoclassical fluxes computed separately, as is usually assumed. The synergy partly results from the turbulence-driven in-out poloidal asymmetry of tungsten density. This result suggests the need for self-consistent simulations of impurity transport, i.e. including both turbulence and neoclassical physics, in view of quantitative predictions for ITER.

  14. Parallel consistent labeling algorithms

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Samal, A.; Henderson, T.

    Mackworth and Freuder have analyzed the time complexity of several constraint satisfaction algorithms. Mohr and Henderson have given new algorithms, AC-4 and PC-3, for arc and path consistency, respectively, and have shown that the arc consistency algorithm is optimal in time complexity and of the same order space complexity as the earlier algorithms. In this paper, they give parallel algorithms for solving node and arc consistency. They show that any parallel algorithm for enforcing arc consistency in the worst case must have O(na) sequential steps, where n is number of nodes, and a is the number of labels per node.more » They give several parallel algorithms to do arc consistency. It is also shown that they all have optimal time complexity. The results of running the parallel algorithms on a BBN Butterfly multiprocessor are also presented.« less

  15. A 32-bit Ultrafast Parallel Correlator using Resonant Tunneling Devices

    NASA Technical Reports Server (NTRS)

    Kulkarni, Shriram; Mazumder, Pinaki; Haddad, George I.

    1995-01-01

    An ultrafast 32-bit pipeline correlator has been implemented using resonant tunneling diodes (RTD) and hetero-junction bipolar transistors (HBT). The negative differential resistance (NDR) characteristics of RTD's is the basis of logic gates with the self-latching property that eliminates pipeline area and delay overheads which limit throughput in conventional technologies. The circuit topology also allows threshold logic functions such as minority/majority to be implemented in a compact manner resulting in reduction of the overall complexity and delay of arbitrary logic circuits. The parallel correlator is an essential component in code division multi-access (CDMA) transceivers used for the continuous calculation of correlation between an incoming data stream and a PN sequence. Simulation results show that a nano-pipelined correlator can provide and effective throughput of one 32-bit correlation every 100 picoseconds, using minimal hardware, with a power dissipation of 1.5 watts. RTD plus HBT based logic gates have been fabricated and the RTD plus HBT based correlator is compared with state of the art complementary metal oxide semiconductor (CMOS) implementations.

  16. The influence of the self-consistent mode structure on the Coriolis pinch effect

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Peeters, A. G.; Camenen, Y.; Casson, F. J.

    This paper discusses the effect of the mode structure on the Coriolis pinch effect [A. G. Peeters, C. Angioni, and D. Strintzi, Phys. Rev. Lett. 98, 265003 (2007)]. It is shown that the Coriolis drift effect can be compensated for by a finite parallel wave vector, resulting in a reduced momentum pinch velocity. Gyrokinetic simulations in full toroidal geometry reveal that parallel dynamics effectively removes the Coriolis pinch for the case of adiabatic electrons, while the compensation due to the parallel dynamics is incomplete for the case of kinetic electrons, resulting in a finite pinch velocity. The finite flux inmore » the case of kinetic electrons is interpreted to be related to the electron trapping, which prevents a strong asymmetry in the electrostatic potential with respect to the low field side position. The physics picture developed here leads to the discovery and explanation of two unexpected effects: First the pinch velocity scales with the trapped particle fraction (root of the inverse aspect ratio), and second there is no strong collisionality dependence. The latter is related to the role of the trapped electrons, which retain some symmetry in the eigenmode, but play no role in the perturbed parallel velocity.« less

  17. Electrophysiological evidence for parallel and serial processing during visual search.

    PubMed

    Luck, S J; Hillyard, S A

    1990-12-01

    Event-related potentials were recorded from young adults during a visual search task in order to evaluate parallel and serial models of visual processing in the context of Treisman's feature integration theory. Parallel and serial search strategies were produced by the use of feature-present and feature-absent targets, respectively. In the feature-absent condition, the slopes of the functions relating reaction time and latency of the P3 component to set size were essentially identical, indicating that the longer reaction times observed for larger set sizes can be accounted for solely by changes in stimulus identification and classification time, rather than changes in post-perceptual processing stages. In addition, the amplitude of the P3 wave on target-present trials in this condition increased with set size and was greater when the preceding trial contained a target, whereas P3 activity was minimal on target-absent trials. These effects are consistent with the serial self-terminating search model and appear to contradict parallel processing accounts of attention-demanding visual search performance, at least for a subset of search paradigms. Differences in ERP scalp distributions further suggested that different physiological processes are utilized for the detection of feature presence and absence.

  18. A Tutorial on Parallel and Concurrent Programming in Haskell

    NASA Astrophysics Data System (ADS)

    Peyton Jones, Simon; Singh, Satnam

    This practical tutorial introduces the features available in Haskell for writing parallel and concurrent programs. We first describe how to write semi-explicit parallel programs by using annotations to express opportunities for parallelism and to help control the granularity of parallelism for effective execution on modern operating systems and processors. We then describe the mechanisms provided by Haskell for writing explicitly parallel programs with a focus on the use of software transactional memory to help share information between threads. Finally, we show how nested data parallelism can be used to write deterministically parallel programs which allows programmers to use rich data types in data parallel programs which are automatically transformed into flat data parallel versions for efficient execution on multi-core processors.

  19. Enhancing membrane protein subcellular localization prediction by parallel fusion of multi-view features.

    PubMed

    Yu, Dongjun; Wu, Xiaowei; Shen, Hongbin; Yang, Jian; Tang, Zhenmin; Qi, Yong; Yang, Jingyu

    2012-12-01

    Membrane proteins are encoded by ~ 30% in the genome and function importantly in the living organisms. Previous studies have revealed that membrane proteins' structures and functions show obvious cell organelle-specific properties. Hence, it is highly desired to predict membrane protein's subcellular location from the primary sequence considering the extreme difficulties of membrane protein wet-lab studies. Although many models have been developed for predicting protein subcellular locations, only a few are specific to membrane proteins. Existing prediction approaches were constructed based on statistical machine learning algorithms with serial combination of multi-view features, i.e., different feature vectors are simply serially combined to form a super feature vector. However, such simple combination of features will simultaneously increase the information redundancy that could, in turn, deteriorate the final prediction accuracy. That's why it was often found that prediction success rates in the serial super space were even lower than those in a single-view space. The purpose of this paper is investigation of a proper method for fusing multiple multi-view protein sequential features for subcellular location predictions. Instead of serial strategy, we propose a novel parallel framework for fusing multiple membrane protein multi-view attributes that will represent protein samples in complex spaces. We also proposed generalized principle component analysis (GPCA) for feature reduction purpose in the complex geometry. All the experimental results through different machine learning algorithms on benchmark membrane protein subcellular localization datasets demonstrate that the newly proposed parallel strategy outperforms the traditional serial approach. We also demonstrate the efficacy of the parallel strategy on a soluble protein subcellular localization dataset indicating the parallel technique is flexible to suite for other computational biology problems. The software and datasets are available at: http://www.csbio.sjtu.edu.cn/bioinf/mpsp.

  20. Self-corrected chip-based dual-comb spectrometer.

    PubMed

    Hébert, Nicolas Bourbeau; Genest, Jérôme; Deschênes, Jean-Daniel; Bergeron, Hugo; Chen, George Y; Khurmi, Champak; Lancaster, David G

    2017-04-03

    We present a dual-comb spectrometer based on two passively mode-locked waveguide lasers integrated in a single Er-doped ZBLAN chip. This original design yields two free-running frequency combs having a high level of mutual stability. We developed in parallel a self-correction algorithm that compensates residual relative fluctuations and yields mode-resolved spectra without the help of any reference laser or control system. Fluctuations are extracted directly from the interferograms using the concept of ambiguity function, which leads to a significant simplification of the instrument that will greatly ease its widespread adoption and commercial deployment. Comparison with a correction algorithm relying on a single-frequency laser indicates discrepancies of only 50 attoseconds on optical timings. The capacities of this instrument are finally demonstrated with the acquisition of a high-resolution molecular spectrum covering 20 nm. This new chip-based multi-laser platform is ideal for the development of high-repetition-rate, compact and fieldable comb spectrometers in the near- and mid-infrared.

  1. When the lowest energy does not induce native structures: parallel minimization of multi-energy values by hybridizing searching intelligences.

    PubMed

    Lü, Qiang; Xia, Xiao-Yan; Chen, Rong; Miao, Da-Jun; Chen, Sha-Sha; Quan, Li-Jun; Li, Hai-Ou

    2012-01-01

    Protein structure prediction (PSP), which is usually modeled as a computational optimization problem, remains one of the biggest challenges in computational biology. PSP encounters two difficult obstacles: the inaccurate energy function problem and the searching problem. Even if the lowest energy has been luckily found by the searching procedure, the correct protein structures are not guaranteed to obtain. A general parallel metaheuristic approach is presented to tackle the above two problems. Multi-energy functions are employed to simultaneously guide the parallel searching threads. Searching trajectories are in fact controlled by the parameters of heuristic algorithms. The parallel approach allows the parameters to be perturbed during the searching threads are running in parallel, while each thread is searching the lowest energy value determined by an individual energy function. By hybridizing the intelligences of parallel ant colonies and Monte Carlo Metropolis search, this paper demonstrates an implementation of our parallel approach for PSP. 16 classical instances were tested to show that the parallel approach is competitive for solving PSP problem. This parallel approach combines various sources of both searching intelligences and energy functions, and thus predicts protein conformations with good quality jointly determined by all the parallel searching threads and energy functions. It provides a framework to combine different searching intelligence embedded in heuristic algorithms. It also constructs a container to hybridize different not-so-accurate objective functions which are usually derived from the domain expertise.

  2. When the Lowest Energy Does Not Induce Native Structures: Parallel Minimization of Multi-Energy Values by Hybridizing Searching Intelligences

    PubMed Central

    Lü, Qiang; Xia, Xiao-Yan; Chen, Rong; Miao, Da-Jun; Chen, Sha-Sha; Quan, Li-Jun; Li, Hai-Ou

    2012-01-01

    Background Protein structure prediction (PSP), which is usually modeled as a computational optimization problem, remains one of the biggest challenges in computational biology. PSP encounters two difficult obstacles: the inaccurate energy function problem and the searching problem. Even if the lowest energy has been luckily found by the searching procedure, the correct protein structures are not guaranteed to obtain. Results A general parallel metaheuristic approach is presented to tackle the above two problems. Multi-energy functions are employed to simultaneously guide the parallel searching threads. Searching trajectories are in fact controlled by the parameters of heuristic algorithms. The parallel approach allows the parameters to be perturbed during the searching threads are running in parallel, while each thread is searching the lowest energy value determined by an individual energy function. By hybridizing the intelligences of parallel ant colonies and Monte Carlo Metropolis search, this paper demonstrates an implementation of our parallel approach for PSP. 16 classical instances were tested to show that the parallel approach is competitive for solving PSP problem. Conclusions This parallel approach combines various sources of both searching intelligences and energy functions, and thus predicts protein conformations with good quality jointly determined by all the parallel searching threads and energy functions. It provides a framework to combine different searching intelligence embedded in heuristic algorithms. It also constructs a container to hybridize different not-so-accurate objective functions which are usually derived from the domain expertise. PMID:23028708

  3. Bond Strength of a Novel One Bottle Multi-mode Adhesive to Human Dentin After Six Months of Storage

    PubMed Central

    Manfroi, Fernanda Borguetti; Marcondes, Maurem Leitão; Somacal, Deise Caren; Borges, Gilberto Antonio; Júnior, Luiz Henrique Burnett; Spohr, Ana Maria

    2016-01-01

    Objective: The aim of the study was to evaluate the microtensile bond strength (µTBS) of Scotchbond Universal to dentin using the etch-and-rinse or the self-etch technique after 24 h and 6 months of storage. Materials and Methods: Flat dentin surfaces were obtained in 24 third molars. The teeth were divided into four groups: G1 – Scotchbond Universal applied in the etch-and-rinse mode; G2 – Scotchbond Universal applied in the self-etch mode; G3 – Scotchbond Multi-Purpose; G4 – Clearfil SE Bond. A block of composite was built on the adhesive area. The tooth/resin sets were cut parallel to the long axis to obtain 40 beams (~0.8 mm2) for each group. Twenty specimens were immediately submitted to the µTBS test, and the remaining 20 were stored in water for 6 months. Failures and the adhesive interface were analyzed by SEM. Results: According to two-way ANOVA, the interaction between adhesive and storage time was significant (p=0.015).The µTBS (MPa) means were the following: 24 h – G1 (39.37±10.82), G2 (31.02±13.76), G3 (35.09±14.03) and G4 (35.84±11.06); 6 months – G1 (36.99±8.78), G2 (40.58±8.07), G3 (32.44±6.07) and G4 (41.75±8.25). Most failures were mixed. Evidence of hybrid layer and numerous resin tags were noted for Scotchbond Universal applied with the etch-and-rinse mode and Scotchbond Multi-Purpose. A thinner hybrid layer and fewer resin tags were noted for Scotchbond Universal applied in the self-etch mode and Clearfil SE Bond. Conclusion: The results indicate that the µTBS for Scotchbond Universal is comparable to the gold-standard adhesives. Scotchbond Universal applied in the self-etch mode and Clearfil SE Bond revealed higher bond stability compared to the etch-and-rinse mode. PMID:27347230

  4. Modeling Plasma Turbulence and Flows in LAPD using BOUT++

    NASA Astrophysics Data System (ADS)

    Friedman, B.; Carter, T. A.; Schaffner, D.; Popovich, P.; Umansky, M. V.; Dudson, B.

    2010-11-01

    A Braginskii fluid model of plasma turbulence in the BOUT code has recently been applied to LAPD at UCLA [1]. While these initial simulations with a reduced model and periodic axial boundary conditions have shown good agreement with measurements (e.g. power spectrum, correlation lengths), these simulations have lacked physics essential for modeling self-consistent, quantitatively correct flows. In particular, the model did not contain parallel plasma flow induced by sheath boundary conditions, and the axisymmetric radial electric field was not consistent with experiment. This work addresses these issues by extending the simulation model in the BOUT++ code [2], a more advanced version of BOUT. Specifically, end-plate sheath boundary conditions are added, as well as equations to evolve electron temperature and parallel ion velocity. Finally, various techniques are used to attempt to match the experimental electric potential profile, including fixing an equilibrium profile, fixing the radial boundaries, and adding an angular momentum source. [4pt] [1] Popovich et al., http://arxiv.org/abs/1005.2418 (2010).[0pt] [2] Dudson et al., Computer Physics Communications 180 (2009).

  5. A self-consistent first-principle based approach to model carrier mobility in organic materials

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Meded, Velimir; Friederich, Pascal; Symalla, Franz

    2015-12-31

    Transport through thin organic amorphous films, utilized in OLEDs and OPVs, has been a challenge to model by using ab-initio methods. Charge carrier mobility depends strongly on the disorder strength and reorganization energy, both of which are significantly affected by the details in environment of each molecule. Here we present a multi-scale approach to describe carrier mobility in which the materials morphology is generated using DEPOSIT, a Monte Carlo based atomistic simulation approach, or, alternatively by molecular dynamics calculations performed with GROMACS. From this morphology we extract the material specific hopping rates, as well as the on-site energies using amore » fully self-consistent embedding approach to compute the electronic structure parameters, which are then used in an analytic expression for the carrier mobility. We apply this strategy to compute the carrier mobility for a set of widely studied molecules and obtain good agreement between experiment and theory varying over several orders of magnitude in the mobility without any freely adjustable parameters. The work focuses on the quantum mechanical step of the multi-scale workflow, explains the concept along with the recently published workflow optimization, which combines density functional with semi-empirical tight binding approaches. This is followed by discussion on the analytic formula and its agreement with established percolation fits as well as kinetic Monte Carlo numerical approaches. Finally, we skatch an unified multi-disciplinary approach that integrates materials science simulation and high performance computing, developed within EU project MMM@HPC.« less

  6. Jali - Unstructured Mesh Infrastructure for Multi-Physics Applications

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Garimella, Rao V; Berndt, Markus; Coon, Ethan

    2017-04-13

    Jali is a parallel unstructured mesh infrastructure library designed for use by multi-physics simulations. It supports 2D and 3D arbitrary polyhedral meshes distributed over hundreds to thousands of nodes. Jali can read write Exodus II meshes along with fields and sets on the mesh and support for other formats is partially implemented or is (https://github.com/MeshToolkit/MSTK), an open source general purpose unstructured mesh infrastructure library from Los Alamos National Laboratory. While it has been made to work with other mesh frameworks such as MOAB and STKmesh in the past, support for maintaining the interface to these frameworks has been suspended formore » now. Jali supports distributed as well as on-node parallelism. Support of on-node parallelism is through direct use of the the mesh in multi-threaded constructs or through the use of "tiles" which are submeshes or sub-partitions of a partition destined for a compute node.« less

  7. Topical perspective on massive threading and parallelism.

    PubMed

    Farber, Robert M

    2011-09-01

    Unquestionably computer architectures have undergone a recent and noteworthy paradigm shift that now delivers multi- and many-core systems with tens to many thousands of concurrent hardware processing elements per workstation or supercomputer node. GPGPU (General Purpose Graphics Processor Unit) technology in particular has attracted significant attention as new software development capabilities, namely CUDA (Compute Unified Device Architecture) and OpenCL™, have made it possible for students as well as small and large research organizations to achieve excellent speedup for many applications over more conventional computing architectures. The current scientific literature reflects this shift with numerous examples of GPGPU applications that have achieved one, two, and in some special cases, three-orders of magnitude increased computational performance through the use of massive threading to exploit parallelism. Multi-core architectures are also evolving quickly to exploit both massive-threading and massive-parallelism such as the 1.3 million threads Blue Waters supercomputer. The challenge confronting scientists in planning future experimental and theoretical research efforts--be they individual efforts with one computer or collaborative efforts proposing to use the largest supercomputers in the world is how to capitalize on these new massively threaded computational architectures--especially as not all computational problems will scale to massive parallelism. In particular, the costs associated with restructuring software (and potentially redesigning algorithms) to exploit the parallelism of these multi- and many-threaded machines must be considered along with application scalability and lifespan. This perspective is an overview of the current state of threading and parallelize with some insight into the future. Published by Elsevier Inc.

  8. Signal acquisition module design for multi-channel surface magnetic resonance sounding system

    NASA Astrophysics Data System (ADS)

    Lin, Tingting; Chen, Wuqiang; Du, Wenyuan; Zhao, Jing

    2015-11-01

    To obtain a precise 2D/3D image of fissure or karst water, multi-channel magnetic resonance sounding (MRS) systems using edge-to-edge or overlapping receiving coils are needed. Thus, acquiring a nano-volt signal for a small amount of the aquifer and suppressing the mutual coupling between adjacent coils are two important issues for the design of the signal acquisition module in the system. In the present study, we propose to use a passive low pass filter, consisted of a resistance (R) and capacitance (C), to inhibit the mutual coupling effects of the coils. Four low-noise operational amplifiers LT1028, OPA124, AD745, and OP27 were compared with respect to achieving the lowest system noise. As a result, 3 pieces of LT1028 were chosen and connected in parallel to serve as preamplifier, with a sensitivity of 1.4 nV/√Hz at 2 kHz. Experimental results are presented for 2D MRS groundwater investigations conducted in the suburb of Changchun, China. The inversion result is consistent with the result of drilling log, suggesting that the signal acquisition module is well developed.

  9. Effects of a parallel electric field and the geomagnetic field in the topside ionosphere on auroral and photoelectron energy distributions

    NASA Technical Reports Server (NTRS)

    Min, Q.-L.; Lummerzheim, D.; Rees, M. H.; Stamnes, K.

    1993-01-01

    The consequences of electric field acceleration and an inhomogeneous magnetic field on auroral electron energy distributions in the topside ionosphere are investigated. The one-dimensional, steady state electron transport equation includes elastic and inelastic collisions, an inhomogeneous magnetic field, and a field-aligned electric field. The case of a self-consistent polarization electric field is considered first. The self-consistent field is derived by solving the continuity equation for all ions of importance, including diffusion of O(+) and H(+), and the electron and ion energy equations to derive the electron and ion temperatures. The system of coupled electron transport, continuity, and energy equations is solved numerically. Recognizing observations of parallel electric fields of larger magnitude than the baseline case of the polarization field, the effect of two model fields on the electron distribution function is investigated. In one case the field is increased from the polarization field magnitude at 300 km to a maximum at the upper boundary of 800 km, and in another case a uniform field is added to the polarization field. Substantial perturbations of the low energy portion of the electron flux are produced: an upward directed electric field accelerates the downward directed flux of low-energy secondary electrons and decelerates the upward directed component. Above about 400 km the inhomogeneous magnetic field produces anisotropies in the angular distribution of the electron flux. The effects of the perturbed energy distributions on auroral spectral emission features are noted.

  10. Effects of a Parallel Electric Field and the Geomagnetic Field in the Topside Ionosphere on Auroral and Photoelectron Energy Distributions

    NASA Technical Reports Server (NTRS)

    Min, Q.-L.; Lummerzheim, D.; Rees, M. H.; Stamnes, K.

    1993-01-01

    The consequences of electric field acceleration and an inhomogencous magnetic field on auroral electron energy distributions in the topside ionosphere are investigated. The one- dimensional, steady state electron transport equation includes elastic and inelastic collisions, an inhomogencous magnetic field, and a field-aligned electric field. The case of a self-consistent polarization electric field is considered first. The self-consistent field is derived by solving the continuity equation for all ions of importance, including diffusion of 0(+) and H(+), and the electron and ion energy equations to derive the electron and ion temperatures. The system of coupled electron transport, continuity, and energy equations is solved numerically. Recognizing observations of parallel electric fields of larger magnitude than the baseline case of the polarization field, the effect of two model fields on the electron distribution function in investigated. In one case the field is increased from the polarization field magnitude at 300 km to a maximum at the upper boundary of 800 km, and in another case a uniform field is added to the polarization field. Substantial perturbations of the low energy portion of the electron flux are produced: an upward directed electric field accelerates the downward directed flux of low-energy secondary electrons and decelerates the upward directed component. Above about 400 km the inhomogencous magnetic field produces anisotropies in the angular distribution of the electron flux. The effects of the perturbed energy distributions on auroral spectral emission features are noted.

  11. Parallelized multi–graphics processing unit framework for high-speed Gabor-domain optical coherence microscopy

    PubMed Central

    Tankam, Patrice; Santhanam, Anand P.; Lee, Kye-Sung; Won, Jungeun; Canavesi, Cristina; Rolland, Jannick P.

    2014-01-01

    Abstract. Gabor-domain optical coherence microscopy (GD-OCM) is a volumetric high-resolution technique capable of acquiring three-dimensional (3-D) skin images with histological resolution. Real-time image processing is needed to enable GD-OCM imaging in a clinical setting. We present a parallelized and scalable multi-graphics processing unit (GPU) computing framework for real-time GD-OCM image processing. A parallelized control mechanism was developed to individually assign computation tasks to each of the GPUs. For each GPU, the optimal number of amplitude-scans (A-scans) to be processed in parallel was selected to maximize GPU memory usage and core throughput. We investigated five computing architectures for computational speed-up in processing 1000×1000 A-scans. The proposed parallelized multi-GPU computing framework enables processing at a computational speed faster than the GD-OCM image acquisition, thereby facilitating high-speed GD-OCM imaging in a clinical setting. Using two parallelized GPUs, the image processing of a 1×1×0.6  mm3 skin sample was performed in about 13 s, and the performance was benchmarked at 6.5 s with four GPUs. This work thus demonstrates that 3-D GD-OCM data may be displayed in real-time to the examiner using parallelized GPU processing. PMID:24695868

  12. Study of the mapping of Navier-Stokes algorithms onto multiple-instruction/multiple-data-stream computers

    NASA Technical Reports Server (NTRS)

    Eberhardt, D. S.; Baganoff, D.; Stevens, K.

    1984-01-01

    Implicit approximate-factored algorithms have certain properties that are suitable for parallel processing. A particular computational fluid dynamics (CFD) code, using this algorithm, is mapped onto a multiple-instruction/multiple-data-stream (MIMD) computer architecture. An explanation of this mapping procedure is presented, as well as some of the difficulties encountered when trying to run the code concurrently. Timing results are given for runs on the Ames Research Center's MIMD test facility which consists of two VAX 11/780's with a common MA780 multi-ported memory. Speedups exceeding 1.9 for characteristic CFD runs were indicated by the timing results.

  13. Electron Heating at Kinetic Scales in Magnetosheath Turbulence

    NASA Technical Reports Server (NTRS)

    Chasapis, Alexandros; Matthaeus, W. H.; Parashar, T. N.; Lecontel, O.; Retino, A.; Breuillard, H.; Khotyaintsev, Y.; Vaivads, A.; Lavraud, B.; Eriksson, E.; hide

    2017-01-01

    We present a statistical study of coherent structures at kinetic scales, using data from the Magnetospheric Multiscale mission in the Earths magnetosheath. We implemented the multi-spacecraft partial variance of increments (PVI) technique to detect these structures, which are associated with intermittency at kinetic scales. We examine the properties of the electron heating occurring within such structures. We find that, statistically, structures with a high PVI index are regions of significant electron heating. We also focus on one such structure, a current sheet, which shows some signatures consistent with magnetic reconnection. Strong parallel electron heating coincides with whistler emissions at the edges of the current sheet.

  14. Equalizer: a scalable parallel rendering framework.

    PubMed

    Eilemann, Stefan; Makhinya, Maxim; Pajarola, Renato

    2009-01-01

    Continuing improvements in CPU and GPU performances as well as increasing multi-core processor and cluster-based parallelism demand for flexible and scalable parallel rendering solutions that can exploit multipipe hardware accelerated graphics. In fact, to achieve interactive visualization, scalable rendering systems are essential to cope with the rapid growth of data sets. However, parallel rendering systems are non-trivial to develop and often only application specific implementations have been proposed. The task of developing a scalable parallel rendering framework is even more difficult if it should be generic to support various types of data and visualization applications, and at the same time work efficiently on a cluster with distributed graphics cards. In this paper we introduce a novel system called Equalizer, a toolkit for scalable parallel rendering based on OpenGL which provides an application programming interface (API) to develop scalable graphics applications for a wide range of systems ranging from large distributed visualization clusters and multi-processor multipipe graphics systems to single-processor single-pipe desktop machines. We describe the system architecture, the basic API, discuss its advantages over previous approaches, present example configurations and usage scenarios as well as scalability results.

  15. Progress with the COGENT Edge Kinetic Code: Collision operator options

    DOE PAGES

    Dorf, M. A.; Cohen, R. H.; Compton, J. C.; ...

    2012-06-27

    In this study, COGENT is a continuum gyrokinetic code for edge plasmas being developed by the Edge Simulation Laboratory collaboration. The code is distinguished by application of the fourth order conservative discretization, and mapped multiblock grid technology to handle the geometric complexity of the tokamak edge. It is written in v∥-μ (parallel velocity – magnetic moment) velocity coordinates, and making use of the gyrokinetic Poisson equation for the calculation of a self-consistent electric potential. In the present manuscript we report on the implementation and initial testing of a succession of increasingly detailed collision operator options, including a simple drag-diffusion operatormore » in the parallel velocity space, Lorentz collisions, and a linearized model Fokker-Planck collision operator conserving momentum and energy (© 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim)« less

  16. A third-generation density-functional-theory-based method for calculating canonical molecular orbitals of large molecules.

    PubMed

    Hirano, Toshiyuki; Sato, Fumitoshi

    2014-07-28

    We used grid-free modified Cholesky decomposition (CD) to develop a density-functional-theory (DFT)-based method for calculating the canonical molecular orbitals (CMOs) of large molecules. Our method can be used to calculate standard CMOs, analytically compute exchange-correlation terms, and maximise the capacity of next-generation supercomputers. Cholesky vectors were first analytically downscaled using low-rank pivoted CD and CD with adaptive metric (CDAM). The obtained Cholesky vectors were distributed and stored on each computer node in a parallel computer, and the Coulomb, Fock exchange, and pure exchange-correlation terms were calculated by multiplying the Cholesky vectors without evaluating molecular integrals in self-consistent field iterations. Our method enables DFT and massively distributed memory parallel computers to be used in order to very efficiently calculate the CMOs of large molecules.

  17. Daily Associations Among Self-control, Heavy Episodic Drinking, and Relationship Functioning: An Examination of Actor and Partner Effects

    PubMed Central

    Crane, Cory A.; Testa, Maria; Derrick, Jaye L.; Leonard, Kenneth E.

    2014-01-01

    An emerging literature suggests that temporary deficits in the ability to inhibit impulsive urges may be proximally associated with intimate partner aggression. The current study examined the experience of alcohol use and the depletion of self-control in the prediction of relationship functioning. Daily diary data collected from 118 heterosexual couples were analyzed using parallel multi-level Actor Partner Interdependence Models to assess the effects of heavy episodic drinking and depletion of self-control across partners on outcomes of participant-reported daily arguing with and anger toward an intimate partner. Heavy episodic drinking among actors predicted greater arguing but failed to interact with either actor or partner depletion. We also found that greater arguing was reported on days of high congruent actor and partner depletion. Both actor and partner depletion, as well as their interaction, predicted greater partner-specific anger. Greater partner-specific anger was generally reported on days of congruent actor and partner depletion, particularly on days of high partner depletion. The current results highlight the importance of independently assessing partner effects (i.e., depletion of self-control), which interact dynamically with disinhibiting actor effects, in the prediction of daily adverse relationship functioning. Results offer further support for the development of prospective individualized and couples-based interventions for partner conflict. PMID:24700558

  18. A DSMC Study of Low Pressure Argon Discharge

    NASA Astrophysics Data System (ADS)

    Hash, David; Meyyappan, M.

    1997-10-01

    Work toward a self-consistent plasma simulation using the DSMC method for examination of the flowfields of low-pressure high density plasma reactors is presented. Presently, DSMC simulations for these applications involve either treating the electrons as a fluid or imposing experimentally determined values for the electron number density profile. In either approach, the electrons themselves are not physically simulated. Self-consistent plasma DSMC simulations have been conducted for aerospace applications but at a severe computational cost due in part to the scalar architectures on which the codes were employed. The present work attempts to conduct such simulations at a more reasonable cost using a plasma version of the object-oriented parallel Cornell DSMC code, MONACO, on an IBM SP-2. Due the availability of experimental data, the GEC reference cell is chosen to conduct preliminary investigations. An argon discharge is examined thus affording a simple chemistry set with eight gas-phase reactions and five species: Ar, Ar^+, Ar^*, Ar_2, and e where Ar^* is a metastable.

  19. HOW SHOULD WE BEHAVE? A REVIEW OF REASONS AND PERSONS BY DEREK PARFIT

    PubMed Central

    Rachlin, Howard

    2010-01-01

    In this book Parfit attempts to develop a rational, non-religious ethics. Instead of asking, “What does God tell us to do?” he asks, “What does reason tell us to do?” Given a set of simple assumptions, he considers whether it is possible to be consistently selfish or consistently good. Analyses of personal dilemmas (problems of self-control) and moral dilemmas (problems of social cooperation) show that neither consistent selfishness nor consistent goodness is logically possible. Instead, a fine balance must be maintained between, on the one hand, our immediate versus long-term good and, on the other, our own good versus that of other people. Ultimately Parfit fails to develop a formula by which such a balance may be struck. Parfit's analysis is consistent with behavioral analysis in its reductionistic view of the self and the parallel it draws between relations with other people and relations with oneself at other times. Parfit's analysis is inconsistent with behaviorism in its view of the mind as internal, available to introspection, and able to cause behavior. His nonfunctional mentalism leads Parfit to inconsistencies and blocks the path to a consistent ethics. Teleological behaviorism's view of the mind in terms of patterns of overt behavior is not hampered by these inconsistencies and may lead to a functional rather than purely rational ethics.

  20. A three-wavelength multi-channel brain functional imager based on digital lock-in photon-counting technique

    NASA Astrophysics Data System (ADS)

    Ding, Xuemei; Wang, Bingyuan; Liu, Dongyuan; Zhang, Yao; He, Jie; Zhao, Huijuan; Gao, Feng

    2018-02-01

    During the past two decades there has been a dramatic rise in the use of functional near-infrared spectroscopy (fNIRS) as a neuroimaging technique in cognitive neuroscience research. Diffuse optical tomography (DOT) and optical topography (OT) can be employed as the optical imaging techniques for brain activity investigation. However, most current imagers with analogue detection are limited by sensitivity and dynamic range. Although photon-counting detection can significantly improve detection sensitivity, the intrinsic nature of sequential excitations reduces temporal resolution. To improve temporal resolution, sensitivity and dynamic range, we develop a multi-channel continuous-wave (CW) system for brain functional imaging based on a novel lock-in photon-counting technique. The system consists of 60 Light-emitting device (LED) sources at three wavelengths of 660nm, 780nm and 830nm, which are modulated by current-stabilized square-wave signals at different frequencies, and 12 photomultiplier tubes (PMT) based on lock-in photon-counting technique. This design combines the ultra-high sensitivity of the photon-counting technique with the parallelism of the digital lock-in technique. We can therefore acquire the diffused light intensity for all the source-detector pairs (SD-pairs) in parallel. The performance assessments of the system are conducted using phantom experiments, and demonstrate its excellent measurement linearity, negligible inter-channel crosstalk, strong noise robustness and high temporal resolution.

  1. Self-assembled microstructures of confined rod-coil diblock copolymers by self-consistent field theory.

    PubMed

    Yang, Guang; Tang, Ping; Yang, Yuliang; Wang, Qiang

    2010-11-25

    We employ the self-consistent field theory (SCFT) incorporating Maier-Saupe orientational interactions between rods to investigate the self-assembly of rod-coil diblock copolymers (RC DBC) in bulk and especially confined into two flat surfaces in 2D space. A unit vector defined on a spherical surface for describing the orientation of rigid blocks in 3D Euclidean space is discretized with an icosahedron triangular mesh to numerically integrate over rod orientation, which is confirmed to have numerical accuracy and stability higher than that of the normal Gaussian quadrature. For the hockey puck-shaped phases in bulk, geometrical confinement, i.e., the film thickness, plays an important role in the self-assembled structures' transitions for the neutral walls. However, for the lamellar phase (monolayer smectic-C) in bulk, the perpendicular lamellae are always stable, less dependent on the film thicknesses because they can relax to the bulk spacing with less-paid coil-stretching in thin films. In particular, a very thin rod layer near the surfaces is formed even in a very thin film. When the walls prefer rods, parallel lamellae are obtained, strongly dependent on the competition between the degree of the surface fields and film geometrical confinement, and the effect of surface field on lamellar structure as a function of film thickness is investigated. Our simulation results provide a guide to understanding the self-assembly of the rod-coil films with desirable application prospects in the fabrication of organic light emitting devices.

  2. West Virginia US Department of Energy experimental program to stimulate competitive research. Section 2: Human resource development; Section 3: Carbon-based structural materials research cluster; Section 3: Data parallel algorithms for scientific computing

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1994-02-02

    This report consists of three separate but related reports. They are (1) Human Resource Development, (2) Carbon-based Structural Materials Research Cluster, and (3) Data Parallel Algorithms for Scientific Computing. To meet the objectives of the Human Resource Development plan, the plan includes K--12 enrichment activities, undergraduate research opportunities for students at the state`s two Historically Black Colleges and Universities, graduate research through cluster assistantships and through a traineeship program targeted specifically to minorities, women and the disabled, and faculty development through participation in research clusters. One research cluster is the chemistry and physics of carbon-based materials. The objective of thismore » cluster is to develop a self-sustaining group of researchers in carbon-based materials research within the institutions of higher education in the state of West Virginia. The projects will involve analysis of cokes, graphites and other carbons in order to understand the properties that provide desirable structural characteristics including resistance to oxidation, levels of anisotropy and structural characteristics of the carbons themselves. In the proposed cluster on parallel algorithms, research by four WVU faculty and three state liberal arts college faculty are: (1) modeling of self-organized critical systems by cellular automata; (2) multiprefix algorithms and fat-free embeddings; (3) offline and online partitioning of data computation; and (4) manipulating and rendering three dimensional objects. This cluster furthers the state Experimental Program to Stimulate Competitive Research plan by building on existing strengths at WVU in parallel algorithms.« less

  3. Hazards Due to Overdischarge in Lithium-ion Cylindrical Cells in Multi-cell Configurations

    NASA Technical Reports Server (NTRS)

    Jeevarajan, Judith; Strangways, Brad; Nelson, Tim

    2010-01-01

    Lithium-ion cells in the cylindrical Commercial-off-the-shelf 18650 design format were used to study the hazards associated with overdischarge. The cells in series or in parallel configurations were subjected to different conditions of overdischarge. The cells in parallel configurations were all overdischarged to 2.0 V for 75 cycles with one cell removed at 25 cycles to study the health of the cell. The cells in series were designed to be in an unbalanced configuration by discharging one cell in each series configuration before the start of test. The discharge consisted of removing a pre-determined capacity from the cell. This ranged from 50 to 150 mAh removal. The cells were discharged down to a predetermined end-of-discharge voltage cutoff which allowed the cell with lower capacity to go into an overdischarge mode. The cell modules that survived the 75 cycles were subjected to one overvoltage test to 4.4 V/cell.

  4. Microwave conductance properties of aligned multiwall carbon nanotube textile sheets

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brown, Brian L.; Martinez, Patricia; Zakhidov, Anvar A.

    2015-07-06

    Understanding the conductance properties of multi-walled carbon nanotube (MWNT) textile sheets in the microwave regime is essential for their potential use in high-speed and high-frequency applications. To expand current knowledge, complex high-frequency conductance measurements from 0.01 to 50 GHz and across temperatures from 4.2 K to 300 K and magnetic fields up to 2 T were made on textile sheets of highly aligned MWNTs with strand alignment oriented both parallel and perpendicular to the microwave electric field polarization. Sheets were drawn from 329 and 520 μm high MWNT forests that resulted in different DC resistance anisotropy. For all samples, themore » microwave conductance can be modeled approximately by a shunt capacitance in parallel with a frequency-independent conductance, but with no inductive contribution. Finally, this is consistent with diffusive Drude conduction as the primary transport mechanism up to 50 GHz. Further, it is found that the microwave conductance is essentially independent of both temperature and magnetic field.« less

  5. Highly efficient spatial data filtering in parallel using the opensource library CPPPO

    NASA Astrophysics Data System (ADS)

    Municchi, Federico; Goniva, Christoph; Radl, Stefan

    2016-10-01

    CPPPO is a compilation of parallel data processing routines developed with the aim to create a library for "scale bridging" (i.e. connecting different scales by mean of closure models) in a multi-scale approach. CPPPO features a number of parallel filtering algorithms designed for use with structured and unstructured Eulerian meshes, as well as Lagrangian data sets. In addition, data can be processed on the fly, allowing the collection of relevant statistics without saving individual snapshots of the simulation state. Our library is provided with an interface to the widely-used CFD solver OpenFOAM®, and can be easily connected to any other software package via interface modules. Also, we introduce a novel, extremely efficient approach to parallel data filtering, and show that our algorithms scale super-linearly on multi-core clusters. Furthermore, we provide a guideline for choosing the optimal Eulerian cell selection algorithm depending on the number of CPU cores used. Finally, we demonstrate the accuracy and the parallel scalability of CPPPO in a showcase focusing on heat and mass transfer from a dense bed of particles.

  6. Influence of post pattern and resin cement curing mode on the retention of glass fibre posts.

    PubMed

    Poskus, L T; Sgura, R; Paragó, F E M; Silva, E M; Guimarães, J G A

    2010-04-01

    To evaluate the influence of post design and roughness and cement system (dual- or self-cured) on the retention of glass fibre posts. Two tapered and smooth posts (Exacto Cônico No. 2 and White Post No. 1) and two parallel-sided and serrated posts (Fibrekor 1.25 mm and Reforpost No. 2) were adhesively luted with two different resin cements--a dual-cured (Rely-X ARC) and a self-cured (Cement Post)--in 40 single-rooted teeth. The teeth were divided into eight experimental groups (n = 5): PFD--Parallel-serrated-Fibrekor/dual-cured; PRD--Parallel-serrated-Reforpost/dual-cured; TED--Tapered-smooth-Exacto Cônico/dual-cured; TWD--Tapered-smooth-White Post/dual-cured; PFS--Parallel-serrated-Fibrekor/self-cured; PRS--Parallel-serrated-Reforpost/self-cured; TES--Tapered-smooth-Exacto Cônico/self-cured; TWS--Tapered-smooth-White Post/self-cured. The specimens were submitted to a pull-out test at a crosshead speed of 0.5 mm min(-1). Data were analysed using analysis of variance and Bonferroni's multiple comparison test (alpha = 0.05). Pull-out results (MPa) were: PFD = 8.13 (+/-1.71); PRD = 8.30 (+/-0.46); TED = 8.68 (+/-1.71); TWD = 9.35 (+/-1.99); PFS = 8.54 (+/-2.23); PRS = 7.09 (+/-1.96); TES = 8.27 (+/-3.92); TWS = 7.57 (+/-2.35). No statistical significant difference was detected for posts and cement factors and their interaction. The retention of glass fibre posts was not affected by post design or surface roughness nor by resin cement-curing mode. These results imply that the choice for serrated posts and self-cured cements is not related to an improvement in retention.

  7. Development of the hepatitis C self-management program.

    PubMed

    Groessl, Erik J; Weingart, Kimberly R; Gifford, Allen L; Asch, Steven M; Ho, Samuel B

    2011-05-01

    Chronic hepatitis C infection (HCV) is a major health problem that disproportionately affects people with limited resources. Many people with HCV are ineligible or refuse antiviral treatment, but less curative treatment options exist. These options include adhering to follow-up health visits, lifestyle changes, and avoiding hepatotoxins like alcohol. Herein, we describe a recently developed self-management program designed to assist HCV-infected patients with adherence and improve their health-related quality of life (HRQOL). The development of the Hepatitis C Self-Management Program (HCV-SMP) was informed by scientific literature, qualitative interviews with HCV-infected patients, self-management training, and feedback from HCV clinical experts. The Hepatitis C Self-Management Program (HCV-SMP) is a multi-faceted program that employs cognitive-behavioral principles and is designed to provide HCV-infected people with knowledge and skills for improving their HRQOL. The program consists of six 2-h workshop sessions which are held weekly. The sessions consist of a variety of group activities, including disease-specific information dissemination, action planning, and problem-solving. The intervention teaches skills for adhering to challenging treatment recommendations using a validated theoretical model. A randomized trial will test the efficacy of this novel HCV self-management program for improving HRQOL in a difficult to reach population. Published by Elsevier Ireland Ltd.

  8. Multi-resonant electromagnetic shunt in base isolation for vibration damping and energy harvesting

    NASA Astrophysics Data System (ADS)

    Pei, Yalu; Liu, Yilun; Zuo, Lei

    2018-06-01

    This paper investigates multi-resonant electromagnetic shunts applied to base isolation for dual-function vibration damping and energy harvesting. Two multi-mode shunt circuit configurations, namely parallel and series, are proposed and optimized based on the H2 criteria. The root-mean-square (RMS) value of the relative displacement between the base and the primary structure is minimized. Practically, this will improve the safety of base-isolated buildings subjected the broad bandwidth ground acceleration. Case studies of a base-isolated building are conducted in both the frequency and time domains to investigate the effectiveness of multi-resonant electromagnetic shunts under recorded earthquake signals. It shows that both multi-mode shunt circuits outperform traditional single mode shunt circuits by suppressing the first and the second vibration modes simultaneously. Moreover, for the same stiffness ratio, the parallel shunt circuit is more effective at harvesting energy and suppressing vibration, and can more robustly handle parameter mistuning than the series shunt circuit. Furthermore, this paper discusses experimental validation of the effectiveness of multi-resonant electromagnetic shunts for vibration damping and energy harvesting on a scaled-down base isolation system.

  9. Self-Consistent Thermal Accretion Disk Corona Models for Compact Objects. I: Properties of the Corona and the Spectrum of Escaping Radiation

    NASA Technical Reports Server (NTRS)

    Dove, James B.; Wilms, Jorn; Begelman, Mitchell C.

    1997-01-01

    We present the properties of accretion disk corona (ADC) models in which the radiation field, the temperature, and the total opacity of the corona are determined self-consistently. We use a nonlinear Monte Carlo code to perform the calculations. As an example, we discuss models in which the corona is situated above and below a cold accretion disk with a plane-parallel (slab) geometry, similar to the model of Haardt & Maraschi. By Comptonizing the soft radiation emitted by the accretion disk, the corona is responsible for producing the high-energy component of the escaping radiation. Our models include the reprocessing of radiation in the accretion disk. Here the photons either are Compton-reflected or photoabsorbed, giving rise to fluorescent line emission and thermal emission. The self- consistent coronal temperature is determined by balancing heating (due to viscous energy dissipation) with Compton cooling, determined using the fully relativistic, angle-dependent cross sections. The total opacity is found by balancing pair productions with annihilations. We find that, for a disk temperature kT(sub BB) approx. less than 200 eV, these coronae are unable to have a self-consistent temperature higher than approx. 140 keV if the total optical depth is approx. less than 0.2, regardless of the compactness parameter of the corona and the seed opacity. This limitation corresponds to the angle-averaged spectrum of escaping radiation having a photon index approx. greater than 1.8 within the 5-30 keV band. Finally, all models that have reprocessing features also predict a large thermal excess at lower energies. These constraints make explaining the X-ray spectra of persistent black hole candidates with ADC models very problematic.

  10. Halvade-RNA: Parallel variant calling from transcriptomic data using MapReduce.

    PubMed

    Decap, Dries; Reumers, Joke; Herzeel, Charlotte; Costanza, Pascal; Fostier, Jan

    2017-01-01

    Given the current cost-effectiveness of next-generation sequencing, the amount of DNA-seq and RNA-seq data generated is ever increasing. One of the primary objectives of NGS experiments is calling genetic variants. While highly accurate, most variant calling pipelines are not optimized to run efficiently on large data sets. However, as variant calling in genomic data has become common practice, several methods have been proposed to reduce runtime for DNA-seq analysis through the use of parallel computing. Determining the effectively expressed variants from transcriptomics (RNA-seq) data has only recently become possible, and as such does not yet benefit from efficiently parallelized workflows. We introduce Halvade-RNA, a parallel, multi-node RNA-seq variant calling pipeline based on the GATK Best Practices recommendations. Halvade-RNA makes use of the MapReduce programming model to create and manage parallel data streams on which multiple instances of existing tools such as STAR and GATK operate concurrently. Whereas the single-threaded processing of a typical RNA-seq sample requires ∼28h, Halvade-RNA reduces this runtime to ∼2h using a small cluster with two 20-core machines. Even on a single, multi-core workstation, Halvade-RNA can significantly reduce runtime compared to using multi-threading, thus providing for a more cost-effective processing of RNA-seq data. Halvade-RNA is written in Java and uses the Hadoop MapReduce 2.0 API. It supports a wide range of distributions of Hadoop, including Cloudera and Amazon EMR.

  11. ATDM LANL FleCSI: Topology and Execution Framework

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bergen, Benjamin Karl

    FleCSI is a compile-time configurable C++ framework designed to support multi-physics application development. As such, FleCSI attempts to provide a very general set of infrastructure design patterns that can be specialized and extended to suit the needs of a broad variety of solver and data requirements. This means that FleCSI is potentially useful to many different ECP projects. Current support includes multidimensional mesh topology, mesh geometry, and mesh adjacency information, n-dimensional hashed-tree data structures, graph partitioning interfaces, and dependency closures (to identify data dependencies between distributed-memory address spaces). FleCSI introduces a functional programming model with control, execution, and data abstractionsmore » that are consistent with state-of-the-art task-based runtimes such as Legion and Charm++. The model also provides support for fine-grained, data-parallel execution with backend support for runtimes such as OpenMP and C++17. The FleCSI abstraction layer provides the developer with insulation from the underlying runtimes, while allowing support for multiple runtime systems, including conventional models like asynchronous MPI. The intent is to give developers a concrete set of user-friendly programming tools that can be used now, while allowing flexibility in choosing runtime implementations and optimizations that can be applied to architectures and runtimes that arise in the future. This project is essential to the ECP Ristra Next-Generation Code project, part of ASC ATDM, because it provides a hierarchically parallel programming model that is consistent with the design of modern system architectures, but which allows for the straightforward expression of algorithmic parallelism in a portably performant manner.« less

  12. Dynamic Self-Consistent Field Theories for Polymer Blends and Block Copolymers

    NASA Astrophysics Data System (ADS)

    Kawakatsu, Toshihiro

    Understanding the behavior of the phase separated domain structures and rheological properties of multi-component polymeric systems require detailed information on the dynamics of domains and that of conformations of constituent polymer chains. Self-consistent field (SCF) theory is a useful tool to treat such a problem because the conformation entropy of polymer chains in inhomogeneous systems can be evaluated quantitatively using this theory. However, when we turn our attention to the dynamic properties in a non-equilibrium state, the basic assumption of the SCF theory, i.e. the assumption of equilibrium chain conformation, breaks down. In order to avoid such a difficulty, dynamic SCF theories were developed. In this chapter, we give a brief review of the recent developments of dynamic SCF theories, and discuss where the cutting-edge of this theory is.

  13. DGDFT: A massively parallel method for large scale density functional theory calculations.

    PubMed

    Hu, Wei; Lin, Lin; Yang, Chao

    2015-09-28

    We describe a massively parallel implementation of the recently developed discontinuous Galerkin density functional theory (DGDFT) method, for efficient large-scale Kohn-Sham DFT based electronic structure calculations. The DGDFT method uses adaptive local basis (ALB) functions generated on-the-fly during the self-consistent field iteration to represent the solution to the Kohn-Sham equations. The use of the ALB set provides a systematic way to improve the accuracy of the approximation. By using the pole expansion and selected inversion technique to compute electron density, energy, and atomic forces, we can make the computational complexity of DGDFT scale at most quadratically with respect to the number of electrons for both insulating and metallic systems. We show that for the two-dimensional (2D) phosphorene systems studied here, using 37 basis functions per atom allows us to reach an accuracy level of 1.3 × 10(-4) Hartree/atom in terms of the error of energy and 6.2 × 10(-4) Hartree/bohr in terms of the error of atomic force, respectively. DGDFT can achieve 80% parallel efficiency on 128,000 high performance computing cores when it is used to study the electronic structure of 2D phosphorene systems with 3500-14 000 atoms. This high parallel efficiency results from a two-level parallelization scheme that we will describe in detail.

  14. Parallel Agent-Based Simulations on Clusters of GPUs and Multi-Core Processors

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Aaby, Brandon G; Perumalla, Kalyan S; Seal, Sudip K

    2010-01-01

    An effective latency-hiding mechanism is presented in the parallelization of agent-based model simulations (ABMS) with millions of agents. The mechanism is designed to accommodate the hierarchical organization as well as heterogeneity of current state-of-the-art parallel computing platforms. We use it to explore the computation vs. communication trade-off continuum available with the deep computational and memory hierarchies of extant platforms and present a novel analytical model of the tradeoff. We describe our implementation and report preliminary performance results on two distinct parallel platforms suitable for ABMS: CUDA threads on multiple, networked graphical processing units (GPUs), and pthreads on multi-core processors. Messagemore » Passing Interface (MPI) is used for inter-GPU as well as inter-socket communication on a cluster of multiple GPUs and multi-core processors. Results indicate the benefits of our latency-hiding scheme, delivering as much as over 100-fold improvement in runtime for certain benchmark ABMS application scenarios with several million agents. This speed improvement is obtained on our system that is already two to three orders of magnitude faster on one GPU than an equivalent CPU-based execution in a popular simulator in Java. Thus, the overall execution of our current work is over four orders of magnitude faster when executed on multiple GPUs.« less

  15. Adaptive parallel logic networks

    NASA Technical Reports Server (NTRS)

    Martinez, Tony R.; Vidal, Jacques J.

    1988-01-01

    Adaptive, self-organizing concurrent systems (ASOCS) that combine self-organization with massive parallelism for such applications as adaptive logic devices, robotics, process control, and system malfunction management, are presently discussed. In ASOCS, an adaptive network composed of many simple computing elements operating in combinational and asynchronous fashion is used and problems are specified by presenting if-then rules to the system in the form of Boolean conjunctions. During data processing, which is a different operational phase from adaptation, the network acts as a parallel hardware circuit.

  16. Multi-water-bag models of ion temperature gradient instability in cylindrical geometry

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Coulette, David; Besse, Nicolas

    2013-05-15

    Ion temperature gradient instabilities play a major role in the understanding of anomalous transport in core fusion plasmas. In the considered cylindrical geometry, ion dynamics is described using a drift-kinetic multi-water-bag model for the parallel velocity dependency of the ion distribution function. In a first stage, global linear stability analysis is performed. From the obtained normal modes, parametric dependencies of the main spectral characteristics of the instability are then examined. Comparison of the multi-water-bag results with a reference continuous Maxwellian case allows us to evaluate the effects of discrete parallel velocity sampling induced by the Multi-Water-Bag model. Differences between themore » global model and local models considered in previous works are discussed. Using results from linear, quasilinear, and nonlinear numerical simulations, an analysis of the first stage saturation dynamics of the instability is proposed, where the divergence between the three models is examined.« less

  17. Implementation and performance of parallel Prolog interpreter

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wei, S.; Kale, L.V.; Balkrishna, R.

    1988-01-01

    In this paper, the authors discuss the implementation of a parallel Prolog interpreter on different parallel machines. The implementation is based on the REDUCE--OR process model which exploits both AND and OR parallelism in logic programs. It is machine independent as it runs on top of the chare-kernel--a machine-independent parallel programming system. The authors also give the performance of the interpreter running a diverse set of benchmark pargrams on parallel machines including shared memory systems: an Alliant FX/8, Sequent and a MultiMax, and a non-shared memory systems: Intel iPSC/32 hypercube, in addition to its performance on a multiprocessor simulation system.

  18. Final Technical Report for Years 1-4 of the Early Career Research Project "Viscosity and equation of state of hot and dense QCD matter" - ARRA portion

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Molnar, Denes

    2014-04-14

    The Section below summarizes research activities and achievements during the first four years of the PI’s Early Career Research Project (ECRP). Two main areas have been advanced: i) radiative 3 ↔ 2 radiative transport, via development of a new computer code MPC/Grid that solves the Boltzmann transport equation in full 6+1D (3X+3V+time) on both single-CPU and parallel computers; ii) development of a self-consistent framework to convert viscous fluids to particles, and application of this framework to relativistic heavy-ion collisions, in particular, determination of the shear viscosity. Year 5 of the ECRP is under a separate award number, and therefore itmore » has its own report document ’Final Technical Report for Year 5 of the Early Career Research Project “Viscosity and equation of state of hot and dense QCDmatter”’ (award DE-SC0008028). The PI’s group was also part of the DOE JET Topical Collaboration, a multi-institution project that overlapped in time significantly with the ECRP. Purdue achievements as part of the JET Topical Collaboration are in a separate report “Final Technical Report summarizing Purdue research activities as part of the DOE JET Topical Collaboration” (award DE-SC0004077).« less

  19. Insight into a conformation of the PNA-PNA duplex with (2‧R,4‧R)- and (2‧R,4‧S)-prolyl-(1S,2S)-2-aminocyclopentanecarboxylic acid backbones

    NASA Astrophysics Data System (ADS)

    Maitarad, Amphawan; Poomsuk, Nattawee; Vilaivan, Chotima; Vilaivan, Tirayut; Siriwong, Khatcharin

    2018-04-01

    Suitable conformations for peptide nucleic acid (PNA) self-hybrids with (2‧R,4‧R)- and (2‧R,4‧S)-prolyl-(1S,2S)-2-aminocyclopentanecarboxylic acid backbones (namely, acpcPNA and epi-acpcPNA, respectively) were investigated based on molecular dynamics simulations. The results revealed that hybridization of the acpcPNA was observed only in the parallel direction, with a conformation close to the P-type structure. In contrast, self-hybrids of the epi-acpcPNA were formed in the antiparallel and parallel directions; the antiparallel duplex adopted the B-form conformation, and the parallel duplex was between B- and P-forms. The calculated binding energies and the experimental data indicate that the antiparallel epi-acpcPNA self-hybrid was more stable than the parallel duplex.

  20. Growth of self-textured Ga3+-substituted Li7La3Zr2O12 ceramics by solid state reaction and their significant enhancement in ionic conductivity

    NASA Astrophysics Data System (ADS)

    Qin, Shiying; Zhu, Xiaohong; Jiang, Yue; Ling, Ming'en; Hu, Zhiwei; Zhu, Jiliang

    2018-03-01

    A highly self-textured Ga2O3-substituted Li7La3Zr2O12 (LLZO-Ga) solid electrolyte with a nominal composition of Li6.55Ga0.15La3Zr2O12 is obtained by a simple and low-cost solid-state reaction technique, requiring no seed crystals to achieve grain orientation. The as-prepared self-textured LLZO-Ga shows a strong (420) preferred orientation with a high Lotgering factor of 0.91. Coherently, a terrace-shaped microstructure consisting of many parallel layers, indicating a two-dimensional-like growth mode, is clearly observed in the self-textured sample. As a result, the highly self-textured garnet-type lithium-ion conducting solid electrolyte of LLZO-Ga exhibits an extremely high ionic conductivity, reaching a state-of-the-art level of 2.06 × 10-3 S cm-1 at room temperature (25 °C) and thus shedding light on an important strategy for improving the structure and ionic conductivity of solid electrolytes.

  1. Technology-based self-care methods of improving antiretroviral adherence: a systematic review.

    PubMed

    Saberi, Parya; Johnson, Mallory O

    2011-01-01

    As HIV infection has shifted to a chronic condition, self-care practices have emerged as an important topic for HIV-positive individuals in maintaining an optimal level of health. Self-care refers to activities that patients undertake to maintain and improve health, such as strategies to achieve and maintain high levels of antiretroviral adherence. Technology-based methods are increasingly used to enhance antiretroviral adherence; therefore, we systematically reviewed the literature to examine technology-based self-care methods that HIV-positive individuals utilize to improve adherence. Seven electronic databases were searched from 1/1/1980 through 12/31/2010. We included quantitative and qualitative studies. Among quantitative studies, the primary outcomes included ARV adherence, viral load, and CD4+ cell count and secondary outcomes consisted of quality of life, adverse effects, and feasibility/acceptability data. For qualitative/descriptive studies, interview themes, reports of use, and perceptions of use were summarized. Thirty-six publications were included (24 quantitative and 12 qualitative/descriptive). Studies with exclusive utilization of medication reminder devices demonstrated less evidence of enhancing adherence in comparison to multi-component methods. This systematic review offers support for self-care technology-based approaches that may result in improved antiretroviral adherence. There was a clear pattern of results that favored individually-tailored, multi-function technologies, which allowed for periodic communication with health care providers rather than sole reliance on electronic reminder devices.

  2. Method and structure for skewed block-cyclic distribution of lower-dimensional data arrays in higher-dimensional processor grids

    DOEpatents

    Chatterjee, Siddhartha [Yorktown Heights, NY; Gunnels, John A [Brewster, NY

    2011-11-08

    A method and structure of distributing elements of an array of data in a computer memory to a specific processor of a multi-dimensional mesh of parallel processors includes designating a distribution of elements of at least a portion of the array to be executed by specific processors in the multi-dimensional mesh of parallel processors. The pattern of the designating includes a cyclical repetitive pattern of the parallel processor mesh, as modified to have a skew in at least one dimension so that both a row of data in the array and a column of data in the array map to respective contiguous groupings of the processors such that a dimension of the contiguous groupings is greater than one.

  3. A master-slave parallel hybrid multi-objective evolutionary algorithm for groundwater remediation design under general hydrogeological conditions

    NASA Astrophysics Data System (ADS)

    Wu, J.; Yang, Y.; Luo, Q.; Wu, J.

    2012-12-01

    This study presents a new hybrid multi-objective evolutionary algorithm, the niched Pareto tabu search combined with a genetic algorithm (NPTSGA), whereby the global search ability of niched Pareto tabu search (NPTS) is improved by the diversification of candidate solutions arose from the evolving nondominated sorting genetic algorithm II (NSGA-II) population. Also, the NPTSGA coupled with the commonly used groundwater flow and transport codes, MODFLOW and MT3DMS, is developed for multi-objective optimal design of groundwater remediation systems. The proposed methodology is then applied to a large-scale field groundwater remediation system for cleanup of large trichloroethylene (TCE) plume at the Massachusetts Military Reservation (MMR) in Cape Cod, Massachusetts. Furthermore, a master-slave (MS) parallelization scheme based on the Message Passing Interface (MPI) is incorporated into the NPTSGA to implement objective function evaluations in distributed processor environment, which can greatly improve the efficiency of the NPTSGA in finding Pareto-optimal solutions to the real-world application. This study shows that the MS parallel NPTSGA in comparison with the original NPTS and NSGA-II can balance the tradeoff between diversity and optimality of solutions during the search process and is an efficient and effective tool for optimizing the multi-objective design of groundwater remediation systems under complicated hydrogeologic conditions.

  4. A mixed parallel strategy for the solution of coupled multi-scale problems at finite strains

    NASA Astrophysics Data System (ADS)

    Lopes, I. A. Rodrigues; Pires, F. M. Andrade; Reis, F. J. P.

    2018-02-01

    A mixed parallel strategy for the solution of homogenization-based multi-scale constitutive problems undergoing finite strains is proposed. The approach aims to reduce the computational time and memory requirements of non-linear coupled simulations that use finite element discretization at both scales (FE^2). In the first level of the algorithm, a non-conforming domain decomposition technique, based on the FETI method combined with a mortar discretization at the interface of macroscopic subdomains, is employed. A master-slave scheme, which distributes tasks by macroscopic element and adopts dynamic scheduling, is then used for each macroscopic subdomain composing the second level of the algorithm. This strategy allows the parallelization of FE^2 simulations in computers with either shared memory or distributed memory architectures. The proposed strategy preserves the quadratic rates of asymptotic convergence that characterize the Newton-Raphson scheme. Several examples are presented to demonstrate the robustness and efficiency of the proposed parallel strategy.

  5. Optimizing the Performance of Reactive Molecular Dynamics Simulations for Multi-core Architectures

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Aktulga, Hasan Metin; Coffman, Paul; Shan, Tzu-Ray

    2015-12-01

    Hybrid parallelism allows high performance computing applications to better leverage the increasing on-node parallelism of modern supercomputers. In this paper, we present a hybrid parallel implementation of the widely used LAMMPS/ReaxC package, where the construction of bonded and nonbonded lists and evaluation of complex ReaxFF interactions are implemented efficiently using OpenMP parallelism. Additionally, the performance of the QEq charge equilibration scheme is examined and a dual-solver is implemented. We present the performance of the resulting ReaxC-OMP package on a state-of-the-art multi-core architecture Mira, an IBM BlueGene/Q supercomputer. For system sizes ranging from 32 thousand to 16.6 million particles, speedups inmore » the range of 1.5-4.5x are observed using the new ReaxC-OMP software. Sustained performance improvements have been observed for up to 262,144 cores (1,048,576 processes) of Mira with a weak scaling efficiency of 91.5% in larger simulations containing 16.6 million particles.« less

  6. Real-time multi-mode neutron multiplicity counter

    DOEpatents

    Rowland, Mark S; Alvarez, Raymond A

    2013-02-26

    Embodiments are directed to a digital data acquisition method that collects data regarding nuclear fission at high rates and performs real-time preprocessing of large volumes of data into directly useable forms for use in a system that performs non-destructive assaying of nuclear material and assemblies for mass and multiplication of special nuclear material (SNM). Pulses from a multi-detector array are fed in parallel to individual inputs that are tied to individual bits in a digital word. Data is collected by loading a word at the individual bit level in parallel, to reduce the latency associated with current shift-register systems. The word is read at regular intervals, all bits simultaneously, with no manipulation. The word is passed to a number of storage locations for subsequent processing, thereby removing the front-end problem of pulse pileup. The word is used simultaneously in several internal processing schemes that assemble the data in a number of more directly useable forms. The detector includes a multi-mode counter that executes a number of different count algorithms in parallel to determine different attributes of the count data.

  7. Adaptation of a Multi-Block Structured Solver for Effective Use in a Hybrid CPU/GPU Massively Parallel Environment

    NASA Astrophysics Data System (ADS)

    Gutzwiller, David; Gontier, Mathieu; Demeulenaere, Alain

    2014-11-01

    Multi-Block structured solvers hold many advantages over their unstructured counterparts, such as a smaller memory footprint and efficient serial performance. Historically, multi-block structured solvers have not been easily adapted for use in a High Performance Computing (HPC) environment, and the recent trend towards hybrid GPU/CPU architectures has further complicated the situation. This paper will elaborate on developments and innovations applied to the NUMECA FINE/Turbo solver that have allowed near-linear scalability with real-world problems on over 250 hybrid GPU/GPU cluster nodes. Discussion will focus on the implementation of virtual partitioning and load balancing algorithms using a novel meta-block concept. This implementation is transparent to the user, allowing all pre- and post-processing steps to be performed using a simple, unpartitioned grid topology. Additional discussion will elaborate on developments that have improved parallel performance, including fully parallel I/O with the ADIOS API and the GPU porting of the computationally heavy CPUBooster convergence acceleration module. Head of HPC and Release Management, Numeca International.

  8. SNAVA-A real-time multi-FPGA multi-model spiking neural network simulation architecture.

    PubMed

    Sripad, Athul; Sanchez, Giovanny; Zapata, Mireya; Pirrone, Vito; Dorta, Taho; Cambria, Salvatore; Marti, Albert; Krishnamourthy, Karthikeyan; Madrenas, Jordi

    2018-01-01

    Spiking Neural Networks (SNN) for Versatile Applications (SNAVA) simulation platform is a scalable and programmable parallel architecture that supports real-time, large-scale, multi-model SNN computation. This parallel architecture is implemented in modern Field-Programmable Gate Arrays (FPGAs) devices to provide high performance execution and flexibility to support large-scale SNN models. Flexibility is defined in terms of programmability, which allows easy synapse and neuron implementation. This has been achieved by using a special-purpose Processing Elements (PEs) for computing SNNs, and analyzing and customizing the instruction set according to the processing needs to achieve maximum performance with minimum resources. The parallel architecture is interfaced with customized Graphical User Interfaces (GUIs) to configure the SNN's connectivity, to compile the neuron-synapse model and to monitor SNN's activity. Our contribution intends to provide a tool that allows to prototype SNNs faster than on CPU/GPU architectures but significantly cheaper than fabricating a customized neuromorphic chip. This could be potentially valuable to the computational neuroscience and neuromorphic engineering communities. Copyright © 2017 Elsevier Ltd. All rights reserved.

  9. Influence of spatial beam inhomogeneities on the parameters of a petawatt laser system based on multi-stage parametric amplification

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Frolov, S A; Trunov, V I; Pestryakov, Efim V

    2013-05-31

    We have developed a technique for investigating the evolution of spatial inhomogeneities in high-power laser systems based on multi-stage parametric amplification. A linearised model of the inhomogeneity development is first devised for parametric amplification with the small-scale self-focusing taken into account. It is shown that the application of this model gives the results consistent (with high accuracy and in a wide range of inhomogeneity parameters) with the calculation without approximations. Using the linearised model, we have analysed the development of spatial inhomogeneities in a petawatt laser system based on multi-stage parametric amplification, developed at the Institute of Laser Physics, Siberianmore » Branch of the Russian Academy of Sciences (ILP SB RAS). (control of laser radiation parameters)« less

  10. Acoustic simulation in architecture with parallel algorithm

    NASA Astrophysics Data System (ADS)

    Li, Xiaohong; Zhang, Xinrong; Li, Dan

    2004-03-01

    In allusion to complexity of architecture environment and Real-time simulation of architecture acoustics, a parallel radiosity algorithm was developed. The distribution of sound energy in scene is solved with this method. And then the impulse response between sources and receivers at frequency segment, which are calculated with multi-process, are combined into whole frequency response. The numerical experiment shows that parallel arithmetic can improve the acoustic simulating efficiency of complex scene.

  11. Development and validation of the Chinese version of the Diabetes Management Self-efficacy Scale.

    PubMed

    Vivienne Wu, Shu-Fang; Courtney, Mary; Edwards, Helen; McDowell, Jan; Shortridge-Baggett, Lillie M; Chang, Pei-Jen

    2008-04-01

    The purpose of this study was to translate the Diabetes Management Self-Efficacy Scale (DMSES) into Chinese and test the validity and reliability of the instrument within a Taiwanese population. A two-stage design was used for this study. Stage I consisted of a multi-stepped process of forward and backward translation, using focus groups and consensus meetings to translate the 20-item Australia/English version DMSES to Chinese and test content validity. Stage II established the psychometric properties of the Chinese version DMSES (C-DMSES) by examining the criterion, convergent and construct validity, internal consistency and stability testing. The sample for Stage II comprised 230 patients with type 2 diabetes aged 30 years or more from a diabetes outpatient clinic in Taiwan. Three items were modified to better reflect Chinese practice. The C-DMSES obtained a total average CVI score of .86. The convergent validity of the C-DMSES correlated well with the validated measure of the General Self-Efficacy Scale in measuring self-efficacy (r=.55; p<.01). Criterion-related validity showed that the C-DMSES was a significant predictor of the Summary of Diabetes Self-Care Activities scores (Beta=.58; t=10.75, p<.01). Factor analysis supported the C-DMSES being composed of four subscales. Good internal consistency (Cronbach's alpha=.77 to .93) and test-retest reliability (Pearson correlation coefficient r=.86, p<.01) were found. The C-DMSES is a brief and psychometrically sound measure for evaluation of self-efficacy towards management of diabetes by persons with type 2 diabetes in Chinese populations.

  12. Laser-induced extreme magnetic field in nanorod targets

    NASA Astrophysics Data System (ADS)

    Lécz, Zsolt; Andreev, Alexander

    2018-03-01

    The application of nano-structured target surfaces in laser-solid interaction has attracted significant attention in the last few years. Their ability to absorb significantly more laser energy promises a possible route for advancing the currently established laser ion acceleration concepts. However, it is crucial to have a better understanding of field evolution and electron dynamics during laser-matter interactions before the employment of such exotic targets. This paper focuses on the magnetic field generation in nano-forest targets consisting of parallel nanorods grown on plane surfaces. A general scaling law for the self-generated quasi-static magnetic field amplitude is given and it is shown that amplitudes up to 1 MT field are achievable with current technology. Analytical results are supported by three-dimensional particle-in-cell simulations. Non-parallel arrangements of nanorods has also been considered which result in the generation of donut-shaped azimuthal magnetic fields in a larger volume.

  13. A 2D electrostatic PIC code for the Mark III Hypercube

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ferraro, R.D.; Liewer, P.C.; Decyk, V.K.

    We have implemented a 2D electrostastic plasma particle in cell (PIC) simulation code on the Caltech/JPL Mark IIIfp Hypercube. The code simulates plasma effects by evolving in time the trajectories of thousands to millions of charged particles subject to their self-consistent fields. Each particle`s position and velocity is advanced in time using a leap frog method for integrating Newton`s equations of motion in electric and magnetic fields. The electric field due to these moving charged particles is calculated on a spatial grid at each time by solving Poisson`s equation in Fourier space. These two tasks represent the largest part ofmore » the computation. To obtain efficient operation on a distributed memory parallel computer, we are using the General Concurrent PIC (GCPIC) algorithm previously developed for a 1D parallel PIC code.« less

  14. A self-managed single exercise programme versus usual physiotherapy treatment for rotator cuff tendinopathy: a randomised controlled trial (the SELF study).

    PubMed

    Littlewood, Chris; Bateman, Marcus; Brown, Kim; Bury, Julie; Mawson, Sue; May, Stephen; Walters, Stephen J

    2016-07-01

    To evaluate the clinical effectiveness of a self-managed single exercise programme versus usual physiotherapy treatment for rotator cuff tendinopathy. Multi-centre pragmatic unblinded parallel group randomised controlled trial. UK National Health Service. Patients with a clinical diagnosis of rotator cuff tendinopathy. The intervention was a programme of self-managed exercise prescribed by a physiotherapist in relation to the most symptomatic shoulder movement. The control group received usual physiotherapy treatment. The primary outcome measure was the Shoulder Pain & Disability Index (SPADI) at three months. Secondary outcomes included the SPADI at six and twelve months. A total of 86 patients (self-managed loaded exercise n=42; usual physiotherapy n=44) were randomised. Twenty-six patients were excluded from the analysis because of lack of primary outcome data at the 3 months follow-up, leaving 60 (n=27; n=33) patients for intention to treat analysis. For the primary outcome, the mean SPADI score at three months was 32.4 (SD 20.2) for the self-managed group, and 30.7 (SD 19.7) for the usual physiotherapy treatment group; mean difference adjusted for baseline score: 3.2 (95% Confidence interval -6.0 to +12.4 P = 0.49).By six and twelve months there remained no significant difference between the groups. This study does not provide sufficient evidence of superiority of one intervention over the other in the short-, mid- or long-term and hence a self-management programme based around a single exercise appears comparable to usual physiotherapy treatment. © The Author(s) 2015.

  15. Self-referenced processing, neurodevelopment and joint attention in autism.

    PubMed

    Mundy, Peter; Gwaltney, Mary; Henderson, Heather

    2010-09-01

    This article describes a parallel and distributed processing model (PDPM) of joint attention, self-referenced processing and autism. According to this model, autism involves early impairments in the capacity for rapid, integrated processing of self-referenced (proprioceptive and interoceptive) and other-referenced (exteroceptive) information. Measures of joint attention have proven useful in research on autism because they are sensitive to the early development of the 'parallel' and integrated processing of self- and other-referenced stimuli. Moreover, joint attention behaviors are a consequence, but also an organizer of the functional development of a distal distributed cortical system involving anterior networks including the prefrontal and insula cortices, as well as posterior neural networks including the temporal and parietal cortices. Measures of joint attention provide early behavioral indicators of atypical development in this parallel and distributed processing system in autism. In addition it is proposed that an early, chronic disturbance in the capacity for integrating self- and other-referenced information may have cascading effects on the development of self awareness in autism. The assumptions, empirical support and future research implications of this model are discussed.

  16. Interpersonal and intrapersonal factors as parallel independent mediators in the association between internalized HIV stigma and ART adherence

    PubMed Central

    Seghatol-Eslami, Victoria C.; Dark, Heather; Raper, James L.; Mugavero, Michael J.; Turan, Janet M.; Turan, Bulent

    2016-01-01

    Introduction People living with HIV (PLWH) need to adhere to antiretroviral therapy (ART) to achieve optimal health. One reason for ART non-adherence is HIV-related stigma. Objectives We aimed to examine whether HIV treatment self-efficacy (an intrapersonal mechanism) mediates the stigma – adherence association. We also examined whether self-efficacy and the concern about being seen while taking HIV medication (an interpersonal mechanism) are parallel mediators independent of each other. Methods 180 PLWH self-reported internalized HIV stigma, ART adherence, HIV treatment self-efficacy, and concerns about being seen while taking HIV medication. We calculated bias-corrected 95% confidence intervals (CIs) for indirect effects using bootstrapping to conduct mediation analyses. Results Adherence self-efficacy mediated the relationship between internalized stigma and ART adherence. Additionally, self-efficacy and concern about being seen while taking HIV medication uniquely mediated and explained almost all of the stigma – adherence association in independent paths (parallel mediation). Conclusion These results can inform intervention strategies to promote ART adherence. PMID:27926668

  17. Feasibility analysis of marine ecological on-line integrated monitoring system

    NASA Astrophysics Data System (ADS)

    Chu, D. Z.; Cao, X.; Zhang, S. W.; Wu, N.; Ma, R.; Zhang, L.; Cao, L.

    2017-08-01

    The in-situ water quality sensors were susceptible to biological attachment. Moreover, sea water corrosion and wave impact damage, and many sensors scattered distribution would cause maintenance inconvenience. The paper proposed a highly integrated marine ecological on-line integrated monitoring system, which can be used inside monitoring station. All sensors were reasonably classified, the similar in series, the overall in parallel. The system composition and workflow were described. In addition, the paper proposed attention issues of the system design and corresponding solutions. Water quality multi-parameters and 5 nutrient salts as the verification index, in-situ and systematic data comparison experiment were carried out. The results showed that the data consistency of nutrient salt, PH and salinity was better. Temperature and dissolved oxygen data trend was consistent, but the data had deviation. Turbidity fluctuated greatly; the chlorophyll trend was similar with it. Aiming at the above phenomena, three points system optimization direction were proposed.

  18. Spatial data analytics on heterogeneous multi- and many-core parallel architectures using python

    USGS Publications Warehouse

    Laura, Jason R.; Rey, Sergio J.

    2017-01-01

    Parallel vector spatial analysis concerns the application of parallel computational methods to facilitate vector-based spatial analysis. The history of parallel computation in spatial analysis is reviewed, and this work is placed into the broader context of high-performance computing (HPC) and parallelization research. The rise of cyber infrastructure and its manifestation in spatial analysis as CyberGIScience is seen as a main driver of renewed interest in parallel computation in the spatial sciences. Key problems in spatial analysis that have been the focus of parallel computing are covered. Chief among these are spatial optimization problems, computational geometric problems including polygonization and spatial contiguity detection, the use of Monte Carlo Markov chain simulation in spatial statistics, and parallel implementations of spatial econometric methods. Future directions for research on parallelization in computational spatial analysis are outlined.

  19. Intelligent manipulation technique for multi-branch robotic systems

    NASA Technical Reports Server (NTRS)

    Chen, Alexander Y. K.; Chen, Eugene Y. S.

    1990-01-01

    New analytical development in kinematics planning is reported. The INtelligent KInematics Planner (INKIP) consists of the kinematics spline theory and the adaptive logic annealing process. Also, a novel framework of robot learning mechanism is introduced. The FUzzy LOgic Self Organized Neural Networks (FULOSONN) integrates fuzzy logic in commands, control, searching, and reasoning, the embedded expert system for nominal robotics knowledge implementation, and the self organized neural networks for the dynamic knowledge evolutionary process. Progress on the mechanical construction of SRA Advanced Robotic System (SRAARS) and the real time robot vision system is also reported. A decision was made to incorporate the Local Area Network (LAN) technology in the overall communication system.

  20. On the extension of the MCSCF/CI method

    NASA Technical Reports Server (NTRS)

    Bauschlicher, C., Jr.; Nelin, C. J.; Komornicki, A.

    1984-01-01

    Research conducted during this period was focused on two main areas: (1) bonding in transition metal oxides; and (2) adsorption of CO on Al and Ni. In both of these theoretical studies a major interest was to obtain a better understanding of the nature of the bonding in transition metal containing systems. The studies used self consistent field (SCF), multi-configuration self cosistent field (MCSCF) and configuration interaction (CI) methods in the treatment of the transition metal oxides and only the SCF method in the adsorption studies. The reports of three principle investigators who contributed to this work during the tenure of the project are presented along with associated published papers.

  1. Third-order linearization for self-beating filtered microwave photonic systems using a dual parallel Mach-Zehnder modulator.

    PubMed

    Pérez, Daniel; Gasulla, Ivana; Capmany, José; Fandiño, Javier S; Muñoz, Pascual; Alavi, Hossein

    2016-09-05

    We develop, analyze and apply a linearization technique based on dual parallel Mach-Zehnder modulator to self-beating microwave photonics systems. The approach enables broadband low-distortion transmission and reception at expense of a moderate electrical power penalty yielding a small optical power penalty (<1 dB).

  2. Computing effective properties of random heterogeneous materials on heterogeneous parallel processors

    NASA Astrophysics Data System (ADS)

    Leidi, Tiziano; Scocchi, Giulio; Grossi, Loris; Pusterla, Simone; D'Angelo, Claudio; Thiran, Jean-Philippe; Ortona, Alberto

    2012-11-01

    In recent decades, finite element (FE) techniques have been extensively used for predicting effective properties of random heterogeneous materials. In the case of very complex microstructures, the choice of numerical methods for the solution of this problem can offer some advantages over classical analytical approaches, and it allows the use of digital images obtained from real material samples (e.g., using computed tomography). On the other hand, having a large number of elements is often necessary for properly describing complex microstructures, ultimately leading to extremely time-consuming computations and high memory requirements. With the final objective of reducing these limitations, we improved an existing freely available FE code for the computation of effective conductivity (electrical and thermal) of microstructure digital models. To allow execution on hardware combining multi-core CPUs and a GPU, we first translated the original algorithm from Fortran to C, and we subdivided it into software components. Then, we enhanced the C version of the algorithm for parallel processing with heterogeneous processors. With the goal of maximizing the obtained performances and limiting resource consumption, we utilized a software architecture based on stream processing, event-driven scheduling, and dynamic load balancing. The parallel processing version of the algorithm has been validated using a simple microstructure consisting of a single sphere located at the centre of a cubic box, yielding consistent results. Finally, the code was used for the calculation of the effective thermal conductivity of a digital model of a real sample (a ceramic foam obtained using X-ray computed tomography). On a computer equipped with dual hexa-core Intel Xeon X5670 processors and an NVIDIA Tesla C2050, the parallel application version features near to linear speed-up progression when using only the CPU cores. It executes more than 20 times faster when additionally using the GPU.

  3. Vectorization for Molecular Dynamics on Intel Xeon Phi Corpocessors

    NASA Astrophysics Data System (ADS)

    Yi, Hongsuk

    2014-03-01

    Many modern processors are capable of exploiting data-level parallelism through the use of single instruction multiple data (SIMD) execution. The new Intel Xeon Phi coprocessor supports 512 bit vector registers for the high performance computing. In this paper, we have developed a hierarchical parallelization scheme for accelerated molecular dynamics simulations with the Terfoff potentials for covalent bond solid crystals on Intel Xeon Phi coprocessor systems. The scheme exploits multi-level parallelism computing. We combine thread-level parallelism using a tightly coupled thread-level and task-level parallelism with 512-bit vector register. The simulation results show that the parallel performance of SIMD implementations on Xeon Phi is apparently superior to their x86 CPU architecture.

  4. Self-propulsion of Leidenfrost Drops between Non-Parallel Structures.

    PubMed

    Luo, Cheng; Mrinal, Manjarik; Wang, Xiang

    2017-09-20

    In this work, we explored self-propulsion of a Leidenfrost drop between non-parallel structures. A theoretical model was first developed to determine conditions for liquid drops to start moving away from the corner of two non-parallel plates. These conditions were then simplified for the case of a Leidenfrost drop. Furthermore, ejection speeds and travel distances of Leidenfrost drops were derived using a scaling law. Subsequently, the theoretical models were validated by experiments. Finally, three new devices have been developed to manipulate Leidenfrost drops in different ways.

  5. Metal Ion-Induced Self-Assembly of a Multi-Responsive Block Copolypeptide into Well-Defined Nanocapsules.

    PubMed

    van Eldijk, Mark B; Schoonen, Lise; Cornelissen, Jeroen J L M; Nolte, Roeland J M; van Hest, Jan C M

    2016-05-01

    Protein cages are an interesting class of biomaterials with potential applications in bionanotechnology. Therefore, substantial effort is spent on the development of capsule-forming designer polypeptides with a tailor-made assembly profile. The expanded assembly profile of a triblock copolypeptide consisting of a metal ion chelating hexahistidine-tag, a stimulus-responsive elastin-like polypeptide block, and a pH-responsive morphology-controlling viral capsid protein is presented. The self-assembly of this multi-responsive protein-based block copolymer is triggered by the addition of divalent metal ions. This assembly process yields monodisperse nanocapsules with a 20 nm diameter composed of 60 polypeptides. The well-defined nanoparticles are the result of the emergent properties of all the blocks of the polypeptide. These results demonstrate the feasibility of hexahistidine-tags to function as supramolecular cross-linkers. Furthermore, their potential for the metal ion-mediated encapsulation of hexahistidine-tagged proteins is shown. © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  6. GPU-BASED MONTE CARLO DUST RADIATIVE TRANSFER SCHEME APPLIED TO ACTIVE GALACTIC NUCLEI

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Heymann, Frank; Siebenmorgen, Ralf, E-mail: fheymann@pa.uky.edu

    2012-05-20

    A three-dimensional parallel Monte Carlo (MC) dust radiative transfer code is presented. To overcome the huge computing-time requirements of MC treatments, the computational power of vectorized hardware is used, utilizing either multi-core computer power or graphics processing units. The approach is a self-consistent way to solve the radiative transfer equation in arbitrary dust configurations. The code calculates the equilibrium temperatures of two populations of large grains and stochastic heated polycyclic aromatic hydrocarbons. Anisotropic scattering is treated applying the Heney-Greenstein phase function. The spectral energy distribution (SED) of the object is derived at low spatial resolution by a photon counting proceduremore » and at high spatial resolution by a vectorized ray tracer. The latter allows computation of high signal-to-noise images of the objects at any frequencies and arbitrary viewing angles. We test the robustness of our approach against other radiative transfer codes. The SED and dust temperatures of one- and two-dimensional benchmarks are reproduced at high precision. The parallelization capability of various MC algorithms is analyzed and included in our treatment. We utilize the Lucy algorithm for the optical thin case where the Poisson noise is high, the iteration-free Bjorkman and Wood method to reduce the calculation time, and the Fleck and Canfield diffusion approximation for extreme optical thick cells. The code is applied to model the appearance of active galactic nuclei (AGNs) at optical and infrared wavelengths. The AGN torus is clumpy and includes fluffy composite grains of various sizes made up of silicates and carbon. The dependence of the SED on the number of clumps in the torus and the viewing angle is studied. The appearance of the 10 {mu}m silicate features in absorption or emission is discussed. The SED of the radio-loud quasar 3C 249.1 is fit by the AGN model and a cirrus component to account for the far-infrared emission.« less

  7. Automated analysis and classification of melanocytic tumor on skin whole slide images.

    PubMed

    Xu, Hongming; Lu, Cheng; Berendt, Richard; Jha, Naresh; Mandal, Mrinal

    2018-06-01

    This paper presents a computer-aided technique for automated analysis and classification of melanocytic tumor on skin whole slide biopsy images. The proposed technique consists of four main modules. First, skin epidermis and dermis regions are segmented by a multi-resolution framework. Next, epidermis analysis is performed, where a set of epidermis features reflecting nuclear morphologies and spatial distributions is computed. In parallel with epidermis analysis, dermis analysis is also performed, where dermal cell nuclei are segmented and a set of textural and cytological features are computed. Finally, the skin melanocytic image is classified into different categories such as melanoma, nevus or normal tissue by using a multi-class support vector machine (mSVM) with extracted epidermis and dermis features. Experimental results on 66 skin whole slide images indicate that the proposed technique achieves more than 95% classification accuracy, which suggests that the technique has the potential to be used for assisting pathologists on skin biopsy image analysis and classification. Copyright © 2018 Elsevier Ltd. All rights reserved.

  8. A 12 GHz wavelength spacing multi-wavelength laser source for wireless communication systems

    NASA Astrophysics Data System (ADS)

    Peng, P. C.; Shiu, R. K.; Bitew, M. A.; Chang, T. L.; Lai, C. H.; Junior, J. I.

    2017-08-01

    This paper presents a multi-wavelength laser source with 12 GHz wavelength spacing based on a single distributed feedback laser. A light wave generated from the distributed feedback laser is fed into a frequency shifter loop consisting of 50:50 coupler, dual-parallel Mach-Zehnder modulator, optical amplifier, optical filter, and polarization controller. The frequency of the input wavelength is shifted and then re-injected into the frequency shifter loop. By re-injecting the shifted wavelengths multiple times, we have generated 84 optical carriers with 12 GHz wavelength spacing and stable output power. For each channel, two wavelengths are modulated by a wireless data using the phase modulator and transmitted through a 25 km single mode fiber. In contrast to previously developed schemes, the proposed laser source does not incur DC bias drift problem. Moreover, it is a good candidate for radio-over-fiber systems to support multiple users using a single distributed feedback laser.

  9. Kinetic Effects in Parametric Instabilities of Finite Amplitude Alfven Waves in a Drifting Multi-Species Plasma

    NASA Astrophysics Data System (ADS)

    Maneva, Y. G.; Araneda, J. A.; Poedts, S.

    2014-12-01

    We consider parametric instabilities of finite-amplitude large-scale Alfven waves in a low-beta collisionless multi-species plasma, consisting of fluid electrons, kinetic protons and a drifting population of minor ions. Complementary to many theoretical studies, relying on fluid or multi-fluid approach, in this work we present the solutions of the parametric instability dispersion relation, including kinetic effects in the parallel direction, along the ambient magnetic field. This provides us with the opportunity to predict the importance of some wave-particle interactions like Landau damping of the daughter ion-acoustic waves for the given pump wave and plasma conditions. We apply the dispersion relation to plasma parameters, typical for low-beta collisionless solar wind close to the Sun. We compare the analytical solutions to the linear stage of hybrid numerical simulations and discuss the application of the model to the problems of preferential heating and differential acceleration of minor ions in the solar corona and the fast solar wind. The results of this study provide tools for prediction and interpretation of the magnetic field and particles data as expected from the future Solar Orbiter and Solar Probe Plus missions.

  10. Circuit for high resolution decoding of multi-anode microchannel array detectors

    NASA Technical Reports Server (NTRS)

    Kasle, David B. (Inventor)

    1995-01-01

    A circuit for high resolution decoding of multi-anode microchannel array detectors consisting of input registers accepting transient inputs from the anode array; anode encoding logic circuits connected to the input registers; midpoint pipeline registers connected to the anode encoding logic circuits; and pixel decoding logic circuits connected to the midpoint pipeline registers is described. A high resolution algorithm circuit operates in parallel with the pixel decoding logic circuit and computes a high resolution least significant bit to enhance the multianode microchannel array detector's spatial resolution by halving the pixel size and doubling the number of pixels in each axis of the anode array. A multiplexer is connected to the pixel decoding logic circuit and allows a user selectable pixel address output according to the actual multi-anode microchannel array detector anode array size. An output register concatenates the high resolution least significant bit onto the standard ten bit pixel address location to provide an eleven bit pixel address, and also stores the full eleven bit pixel address. A timing and control state machine is connected to the input registers, the anode encoding logic circuits, and the output register for managing the overall operation of the circuit.

  11. Parallel Monte Carlo Search for Hough Transform

    NASA Astrophysics Data System (ADS)

    Lopes, Raul H. C.; Franqueira, Virginia N. L.; Reid, Ivan D.; Hobson, Peter R.

    2017-10-01

    We investigate the problem of line detection in digital image processing and in special how state of the art algorithms behave in the presence of noise and whether CPU efficiency can be improved by the combination of a Monte Carlo Tree Search, hierarchical space decomposition, and parallel computing. The starting point of the investigation is the method introduced in 1962 by Paul Hough for detecting lines in binary images. Extended in the 1970s to the detection of space forms, what came to be known as Hough Transform (HT) has been proposed, for example, in the context of track fitting in the LHC ATLAS and CMS projects. The Hough Transform transfers the problem of line detection, for example, into one of optimization of the peak in a vote counting process for cells which contain the possible points of candidate lines. The detection algorithm can be computationally expensive both in the demands made upon the processor and on memory. Additionally, it can have a reduced effectiveness in detection in the presence of noise. Our first contribution consists in an evaluation of the use of a variation of the Radon Transform as a form of improving theeffectiveness of line detection in the presence of noise. Then, parallel algorithms for variations of the Hough Transform and the Radon Transform for line detection are introduced. An algorithm for Parallel Monte Carlo Search applied to line detection is also introduced. Their algorithmic complexities are discussed. Finally, implementations on multi-GPU and multicore architectures are discussed.

  12. [The effects of a multi agent obesity control program in obese school children].

    PubMed

    Ahn, Hye Young; Im, Sook Bin; Hong, Kyung Ja; Hur, Myung Haeng

    2007-02-01

    The purpose of this study was to identify the effects of a multi agent obesity control program in obese school children. This program was composed of strategies to modify diet and exercise habits and to change cognitive behavior variables(stress, coping, and self-efficacy). The subjects were 40 obese school children who participated in our project voluntarily via homepage, TV, newspaper, public paper and school official documents. The program was implemented daily for 4 sessions per day for ten days from August 16 to 26, 2004. The daily program consisted of exercise therapy, dance therapy, cognitive behavior therapy and aroma therapy. The data was analyzed by paired t-test using the SPSSWIN program. There was a significant decrease in children's waist-hip ratio (p=.04) and in children's stress (p=.00) after the program. There was a significant increase in children's self-confidence after the program(p=.02) and a significant decrease in children's diet habit after the program(p=.02). This study provides evidence that a multiagent obese control program is effective in changing waist-hip ratio, stress, self-confidence, and diet habits in obese school children.

  13. Multi-physics transient simulation of monolithic niobium dioxide-tantalum dioxide memristor-selector structures

    NASA Astrophysics Data System (ADS)

    Sevic, John F.; Kobayashi, Nobuhiko P.

    2017-10-01

    Self-assembled niobium dioxide (NbO2) thin-film selectors self-aligned to tantalum dioxide (TaO2) memristive memory cells are studied by a multi-physics transient solution of the heat equation coupled to the nonlinear current continuity equation. While a compact model can resolve the quasi-static bulk negative differential resistance (NDR), a self-consistent coupled transport formulation provides a non-equilibrium picture of NbO2-TaO2 selector-memristor operation ab initio. By employing the drift-diffusion transport approximation, a finite element method is used to study the dynamic electrothermal behavior of our experimentally obtained selector-memristor devices, showing that existing conditions are suitable for electroformation of NbO2 selector thin-films. Both transient and steady-state simulations support our theory, suggesting that the phase change due to insulator-metal transition is responsible for NbO2 selector NDR in our as-fabricated selector-memristor devices. Simulation results further suggest that TiN nano-via may play a central role in electroforming, as its dimensions and material properties establish the mutual electrothermal interaction between TiN nano-via and the selector-memristor.

  14. Knowledge Management in Role Based Agents

    NASA Astrophysics Data System (ADS)

    Kır, Hüseyin; Ekinci, Erdem Eser; Dikenelli, Oguz

    In multi-agent system literature, the role concept is getting increasingly researched to provide an abstraction to scope beliefs, norms, goals of agents and to shape relationships of the agents in the organization. In this research, we propose a knowledgebase architecture to increase applicability of roles in MAS domain by drawing inspiration from the self concept in the role theory of sociology. The proposed knowledgebase architecture has granulated structure that is dynamically organized according to the agent's identification in a social environment. Thanks to this dynamic structure, agents are enabled to work on consistent knowledge in spite of inevitable conflicts between roles and the agent. The knowledgebase architecture is also implemented and incorporated into the SEAGENT multi-agent system development framework.

  15. Crystal plasticity simulation of Zirconium tube rolling using multi-grain representative volume element

    NASA Astrophysics Data System (ADS)

    Isaenkova, Margarita; Perlovich, Yuriy; Zhuk, Dmitry; Krymskaya, Olga

    2017-10-01

    The rolling of Zirconium tube is studied by means of the crystal plasticity viscoplastic self-consistent (VPSC) constitutive modeling. This modeling performed by a dislocation-based constitutive model and a spectral solver using open-source simulation of DAMASK kit. The multi-grain representative volume elements with periodic boundary conditions are used to predict the texture evolution and distributions of strain and stresses. Two models for randomly textured and partially rolled material are deformed to 30% reduction in tube wall thickness and 7% reduction in tube diameter. The resulting shapes of the models are shown and distributions of strain are plotted. Also, evolution of grain's shape during deformation is shown.

  16. SU(N) affine Toda solitons and breathers from transparent Dirac potentials

    NASA Astrophysics Data System (ADS)

    Thies, Michael

    2017-05-01

    Transparent scalar and pseudoscalar potentials in the one-dimensional Dirac equation play an important role as self-consistent mean fields in 1  +  1 dimensional four-fermion theories (Gross-Neveu, Nambu-Jona Lasinio models) and quasi-one dimensional superconductors (Bogoliubov-de Gennes equation). Here, we show that they also serve as seed to generate a large class of classical multi-soliton and multi-breather solutions of su(N) affine Toda field theories, including the Lax representation and the corresponding vector. This generalizes previous findings about the relationship between real kinks in the Gross-Neveu model and classical solitons of the sinh-Gordon equation to complex twisted kinks.

  17. Electron Heating at Kinetic Scales in Magnetosheath Turbulence

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chasapis, Alexandros; Matthaeus, W. H.; Parashar, T. N.

    2017-02-20

    We present a statistical study of coherent structures at kinetic scales, using data from the Magnetospheric Multiscale mission in the Earth’s magnetosheath. We implemented the multi-spacecraft partial variance of increments (PVI) technique to detect these structures, which are associated with intermittency at kinetic scales. We examine the properties of the electron heating occurring within such structures. We find that, statistically, structures with a high PVI index are regions of significant electron heating. We also focus on one such structure, a current sheet, which shows some signatures consistent with magnetic reconnection. Strong parallel electron heating coincides with whistler emissions at themore » edges of the current sheet.« less

  18. Accelerating the Gillespie Exact Stochastic Simulation Algorithm using hybrid parallel execution on graphics processing units.

    PubMed

    Komarov, Ivan; D'Souza, Roshan M

    2012-01-01

    The Gillespie Stochastic Simulation Algorithm (GSSA) and its variants are cornerstone techniques to simulate reaction kinetics in situations where the concentration of the reactant is too low to allow deterministic techniques such as differential equations. The inherent limitations of the GSSA include the time required for executing a single run and the need for multiple runs for parameter sweep exercises due to the stochastic nature of the simulation. Even very efficient variants of GSSA are prohibitively expensive to compute and perform parameter sweeps. Here we present a novel variant of the exact GSSA that is amenable to acceleration by using graphics processing units (GPUs). We parallelize the execution of a single realization across threads in a warp (fine-grained parallelism). A warp is a collection of threads that are executed synchronously on a single multi-processor. Warps executing in parallel on different multi-processors (coarse-grained parallelism) simultaneously generate multiple trajectories. Novel data-structures and algorithms reduce memory traffic, which is the bottleneck in computing the GSSA. Our benchmarks show an 8×-120× performance gain over various state-of-the-art serial algorithms when simulating different types of models.

  19. PCTO-SIM: Multiple-point geostatistical modeling using parallel conditional texture optimization

    NASA Astrophysics Data System (ADS)

    Pourfard, Mohammadreza; Abdollahifard, Mohammad J.; Faez, Karim; Motamedi, Sayed Ahmad; Hosseinian, Tahmineh

    2017-05-01

    Multiple-point Geostatistics is a well-known general statistical framework by which complex geological phenomena have been modeled efficiently. Pixel-based and patch-based are two major categories of these methods. In this paper, the optimization-based category is used which has a dual concept in texture synthesis as texture optimization. Our extended version of texture optimization uses the energy concept to model geological phenomena. While honoring the hard point, the minimization of our proposed cost function forces simulation grid pixels to be as similar as possible to training images. Our algorithm has a self-enrichment capability and creates a richer training database from a sparser one through mixing the information of all surrounding patches of the simulation nodes. Therefore, it preserves pattern continuity in both continuous and categorical variables very well. It also shows a fuzzy result in its every realization similar to the expected result of multi realizations of other statistical models. While the main core of most previous Multiple-point Geostatistics methods is sequential, the parallel main core of our algorithm enabled it to use GPU efficiently to reduce the CPU time. One new validation method for MPS has also been proposed in this paper.

  20. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Popovich, P.; Carter, T. A.; Friedman, B.

    Numerical simulation of plasma turbulence in the Large Plasma Device (LAPD) [W. Gekelman, H. Pfister, Z. Lucky et al., Rev. Sci. Instrum. 62, 2875 (1991)] is presented. The model, implemented in the BOUndary Turbulence code [M. Umansky, X. Xu, B. Dudson et al., Contrib. Plasma Phys. 180, 887 (2009)], includes three-dimensional (3D) collisional fluid equations for plasma density, electron parallel momentum, and current continuity, and also includes the effects of ion-neutral collisions. In nonlinear simulations using measured LAPD density profiles but assuming constant temperature profile for simplicity, self-consistent evolution of instabilities and nonlinearly generated zonal flows results in a saturatedmore » turbulent state. Comparisons of these simulations with measurements in LAPD plasmas reveal good qualitative and reasonable quantitative agreement, in particular in frequency spectrum, spatial correlation, and amplitude probability distribution function of density fluctuations. For comparison with LAPD measurements, the plasma density profile in simulations is maintained either by direct azimuthal averaging on each time step, or by adding particle source/sink function. The inferred source/sink values are consistent with the estimated ionization source and parallel losses in LAPD. These simulations lay the groundwork for more a comprehensive effort to test fluid turbulence simulation against LAPD data.« less

  1. Multi-stage separations based on dielectrophoresis

    DOEpatents

    Mariella, Jr., Raymond P.

    2004-07-13

    A system utilizing multi-stage traps based on dielectrophoresis. Traps with electrodes arranged transverse to the flow and traps with electrodes arranged parallel to the flow with combinations of direct current and alternating voltage are used to trap, concentrate, separate, and/or purify target particles.

  2. Efficient multi-objective calibration of a computationally intensive hydrologic model with parallel computing software in Python

    USDA-ARS?s Scientific Manuscript database

    With enhanced data availability, distributed watershed models for large areas with high spatial and temporal resolution are increasingly used to understand water budgets and examine effects of human activities and climate change/variability on water resources. Developing parallel computing software...

  3. Array-based, parallel hierarchical mesh refinement algorithms for unstructured meshes

    DOE PAGES

    Ray, Navamita; Grindeanu, Iulian; Zhao, Xinglin; ...

    2016-08-18

    In this paper, we describe an array-based hierarchical mesh refinement capability through uniform refinement of unstructured meshes for efficient solution of PDE's using finite element methods and multigrid solvers. A multi-degree, multi-dimensional and multi-level framework is designed to generate the nested hierarchies from an initial coarse mesh that can be used for a variety of purposes such as in multigrid solvers/preconditioners, to do solution convergence and verification studies and to improve overall parallel efficiency by decreasing I/O bandwidth requirements (by loading smaller meshes and in memory refinement). We also describe a high-order boundary reconstruction capability that can be used tomore » project the new points after refinement using high-order approximations instead of linear projection in order to minimize and provide more control on geometrical errors introduced by curved boundaries.The capability is developed under the parallel unstructured mesh framework "Mesh Oriented dAtaBase" (MOAB Tautges et al. (2004)). We describe the underlying data structures and algorithms to generate such hierarchies in parallel and present numerical results for computational efficiency and effect on mesh quality. Furthermore, we also present results to demonstrate the applicability of the developed capability to study convergence properties of different point projection schemes for various mesh hierarchies and to a multigrid finite-element solver for elliptic problems.« less

  4. pyPaSWAS: Python-based multi-core CPU and GPU sequence alignment.

    PubMed

    Warris, Sven; Timal, N Roshan N; Kempenaar, Marcel; Poortinga, Arne M; van de Geest, Henri; Varbanescu, Ana L; Nap, Jan-Peter

    2018-01-01

    Our previously published CUDA-only application PaSWAS for Smith-Waterman (SW) sequence alignment of any type of sequence on NVIDIA-based GPUs is platform-specific and therefore adopted less than could be. The OpenCL language is supported more widely and allows use on a variety of hardware platforms. Moreover, there is a need to promote the adoption of parallel computing in bioinformatics by making its use and extension more simple through more and better application of high-level languages commonly used in bioinformatics, such as Python. The novel application pyPaSWAS presents the parallel SW sequence alignment code fully packed in Python. It is a generic SW implementation running on several hardware platforms with multi-core systems and/or GPUs that provides accurate sequence alignments that also can be inspected for alignment details. Additionally, pyPaSWAS support the affine gap penalty. Python libraries are used for automated system configuration, I/O and logging. This way, the Python environment will stimulate further extension and use of pyPaSWAS. pyPaSWAS presents an easy Python-based environment for accurate and retrievable parallel SW sequence alignments on GPUs and multi-core systems. The strategy of integrating Python with high-performance parallel compute languages to create a developer- and user-friendly environment should be considered for other computationally intensive bioinformatics algorithms.

  5. Self-regenerating column chromatography

    DOEpatents

    Park, Woo K.

    1995-05-30

    The present invention provides a process for treating both cations and anions by using a self-regenerating, multi-ionic exchange resin column system which requires no separate regeneration steps. The process involves alternating ion-exchange chromatography for cations and anions in a multi-ionic exchange column packed with a mixture of cation and anion exchange resins. The multi-ionic mixed-charge resin column works as a multi-function column, capable of independently processing either cationic or anionic exchange, or simultaneously processing both cationic and anionic exchanges. The major advantage offered by the alternating multi-function ion exchange process is the self-regeneration of the resins.

  6. Crawling and walking infants encounter objects differently in a multi-target environment.

    PubMed

    Dosso, Jill A; Boudreau, J Paul

    2014-10-01

    From birth, infants move their bodies in order to obtain information and stimulation from their environment. Exploratory movements are important for the development of an infant's understanding of the world and are well established as being key to cognitive advances. Newly acquired motor skills increase the potential actions available to the infant. However, the way that infants employ potential actions in environments with multiple potential targets is undescribed. The current work investigated the target object selections of infants across a range of self-produced locomotor experience (11- to 14-month-old crawlers and walkers). Infants repeatedly accessed objects among pairs of objects differing in both distance and preference status, some requiring locomotion. Overall, their object actions were found to be sensitive to object preference status; however, the role of object distance in shaping object encounters was moderated by movement status. Crawlers' actions appeared opportunistic and were biased towards nearby objects while walkers' actions appeared intentional and were independent of object position. Moreover, walkers' movements favoured preferred objects more strongly for children with higher levels of self-produced locomotion experience. The multi-target experimental situation used in this work parallels conditions faced by foraging organisms, and infants' behaviours were discussed with respect to optimal foraging theory. There is a complex interplay between infants' agency, locomotor experience, and environment in shaping their motor actions. Infants' movements, in turn, determine the information and experiences offered to infants by their micro-environment.

  7. The cerebellum: a new key structure in the navigation system

    PubMed Central

    Rochefort, Christelle; Lefort, Julie M.; Rondi-Reig, Laure

    2013-01-01

    Early investigations of cerebellar function focused on motor learning, in particular on eyeblink conditioning and adaptation of the vestibulo-ocular reflex, and led to the general view that cerebellar long-term depression (LTD) at parallel fiber (PF)–Purkinje cell (PC) synapses is the neural correlate of cerebellar motor learning. Thereafter, while the full complexity of cerebellar plasticities was being unraveled, cerebellar involvement in more cognitive tasks—including spatial navigation—was further investigated. However, cerebellar implication in spatial navigation remains a matter of debate because motor deficits frequently associated with cerebellar damage often prevent the dissociation between its role in spatial cognition from its implication in motor function. Here, we review recent findings from behavioral and electrophysiological analyses of cerebellar mutant mouse models, which show that the cerebellum might participate in the construction of hippocampal spatial representation map (i.e., place cells) and thereby in goal-directed navigation. These recent advances in cerebellar research point toward a model in which computation from the cerebellum could be required for spatial representation and would involve the integration of multi-source self-motion information to: (1) transform the reference frame of vestibular signals and (2) distinguish between self- and externally-generated vestibular signals. We eventually present herein anatomical and functional connectivity data supporting a cerebello-hippocampal interaction. Whilst a direct cerebello-hippocampal projection has been suggested, recent investigations rather favor a multi-synaptic pathway involving posterior parietal and retrosplenial cortices, two regions critically involved in spatial navigation. PMID:23493515

  8. Parallel auto-correlative statistics with VTK.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pebay, Philippe Pierre; Bennett, Janine Camille

    2013-08-01

    This report summarizes existing statistical engines in VTK and presents both the serial and parallel auto-correlative statistics engines. It is a sequel to [PT08, BPRT09b, PT09, BPT09, PT10] which studied the parallel descriptive, correlative, multi-correlative, principal component analysis, contingency, k-means, and order statistics engines. The ease of use of the new parallel auto-correlative statistics engine is illustrated by the means of C++ code snippets and algorithm verification is provided. This report justifies the design of the statistics engines with parallel scalability in mind, and provides scalability and speed-up analysis results for the autocorrelative statistics engine.

  9. Influences on day-to-day self-management of type 2 diabetes among African-American women: spirituality, the multi-caregiver role, and other social context factors.

    PubMed

    Samuel-Hodge, C D; Headen, S W; Skelly, A H; Ingram, A F; Keyserling, T C; Jackson, E J; Ammerman, A S; Elasy, T A

    2000-07-01

    Many African-American women are affected by diabetes and its complications, and culturally appropriate lifestyle interventions that lead to improvements in glycemic control are urgently needed. The aim of this qualitative study was to identify culturally relevant psychosocial issues and social context variables influencing lifestyle behaviors--specifically diet and physical activity--of southern African-American women with diabetes. We conducted 10 focus group interviews with 70 southern African-American women with type 2 diabetes. Group interviews were audiotaped and transcripts were coded using qualitative data analysis software. A panel of reviewers analyzed the coded responses for emerging themes and trends. The dominant and most consistent themes that emerged from these focus groups were 1) spirituality as an important factor in general health, disease adjustment, and coping; 2) general life stress and multi-caregiving responsibilities interfering with daily disease management; and 3) the impact of diabetes manifested in feelings of dietary deprivation, physical and emotional "tiredness," "worry," and fear of diabetes complications. Our findings suggest that influences on diabetes self-management behaviors of African-American women may be best understood from a sociocultural and family context. Interventions to improve self-management for this population should recognize the influences of spirituality, general life stress, multi-caregiving responsibilities, and the psychological impact of diabetes. These findings suggest that family-centered and church-based approaches to diabetes care interventions are appropriate.

  10. The motivation for drug abuse treatment: testing cognitive and 12-step theories.

    PubMed

    Bell, D C; Montoya, I D; Richard, A J; Dayton, C A

    1998-11-01

    The purpose of this paper is to evaluate two models of behavior change: cognitive theory and 12-step theory. Research subjects were drawn from three separate, but parallel, samples of adults. The first sample consisted of out-of-treatment chronic drug users, the second consisted of drug users who had applied for treatment at a publicly funded multiple-provider drug treatment facility, and the third consisted of drug users who had applied for treatment at an intensive outpatient program for crack cocaine users. Cognitive theory was supported. Study participants applying for drug abuse treatment reported a higher level of perceived problem severity and a higher level of cognitive functioning than out-of-treatment drug users. Two hypotheses drawn from 12-step theory were not supported. Treatment applicants had more positive emotional functioning than out-of-treatment drug users, and one treatment-seeking sample had higher self-esteem.

  11. Analysis of photoelectron effect on the antenna impedance via Particle-In-Cell simulation

    NASA Astrophysics Data System (ADS)

    Miyake, Y.; Usui, H.

    2008-08-01

    We present photoelectron effects on the impedance of electric field antennas used for plasma wave investigations. To illustrate the photoelectron effects, we applied electromagnetic Particle-In-Cell simulation to the self-consistent antenna impedance analysis. We confirmed the formation of a dense photoelectron region around the sunlit surfaces of the antenna and the spacecraft. The dense photoelectrons enhance the real part, and decrease the absolute value of the imaginary part, of antenna impedance at low frequencies. We also showed that the antenna conductance can be analytically calculated from simulation results of the electron current flowing into or out of the antenna. The antenna impedance in the photoelectron environment is represented by a parallel equivalent circuit consisting of a capacitance and a resistance, which is consistent with empirical knowledge. The results also imply that the impedance varies with the spin of the spacecraft, which causes the variation of the photoelectron density around the antenna.

  12. The ecology of flows and drift wave turbulence in CSDX: A model

    NASA Astrophysics Data System (ADS)

    Hajjar, R. J.; Diamond, P. H.; Tynan, G. R.

    2018-02-01

    This paper describes the ecology of drift wave turbulence and mean flows in the coupled drift-ion acoustic wave plasma of a CSDX linear device. A 1D reduced model that studies the spatiotemporal evolution of plasma mean density n ¯ , and mean flows v¯ y and v¯ z , in addition to fluctuation intensity ε, is presented. Here, ε= is the conserved energy field. The model uses a mixing length lmix inversely proportional to both axial and azimuthal flow shear. This form of lmix closes the loop on total energy. The model self-consistently describes variations in plasma profiles, including mean flows and turbulent stresses. It investigates the energy exchange between the fluctuation intensity and mean profiles via particle flux and Reynolds stresses and . Acoustic coupling breaks parallel symmetry and generates a parallel residual stress Πxzr e s . The model uses a set of equations to explain the acceleration of v¯ y and v¯ z via Πxyr e s∝∇n ¯ and Πxyr e s∝∇n ¯ . Flow dynamics in the parallel direction are related to those in the perpendicular direction through an empirical coupling constant σVT. This constant measures the degree of symmetry breaking in the correlator and determines the efficiency of ∇n ¯ in driving v¯ z . The model also establishes a relation between ∇v¯ y and ∇v¯ z , via the ratio of the stresses Πxyr e s and Πxzr e s . When parallel to perpendicular flow coupling is weak, axial Reynolds power Pxz R e=-∇v¯ z is less than the azimuthal Reynolds power Pxy R e=-∇v¯ y . The model is then reduced to a 2-field predator/prey model where v¯ z is parasitic to the system and fluctuations evolve self-consistently. Finally, turbulent diffusion in CSDX follows the scaling: DCSDX=DBρ⋆0.6 , where DB is the Bohm diffusion coefficient and ρ⋆ is the ion gyroradius normalized to the density gradient |∇n ¯ /n ¯ |-1 .

  13. Spatial filter system as an optical relay line

    DOEpatents

    Hunt, John T.; Renard, Paul A.

    1979-01-01

    A system consisting of a set of spatial filters that are used to optically relay a laser beam from one position to a downstream position with minimal nonlinear phase distortion and beam intensity variation. The use of the device will result in a reduction of deleterious beam self-focusing and produce a significant increase in neutron yield from the implosion of targets caused by their irradiation with multi-beam glass laser systems.

  14. Identity, grief and self-awareness after traumatic brain injury.

    PubMed

    Carroll, Emma; Coetzer, Rudi

    2011-06-01

    The objective of this study was to investigate perceived identity change in adults with traumatic brain injury (TBI) and explore associations between identity change, grief, depression, self-esteem and self-awareness. The participants were 29 adults with TBI who were being followed up by a community brain injury rehabilitation service. Participants were longer post-injury than those more commonly studied. Time since injury ranged from 2.25 to 40 years (mean = 11.17 years, SD = 11.4 years). Participants completed a battery of questionnaires. Significant others and clinicians completed a parallel version of one of these measures. Questionnaires included the Head Injury Semantic Differential Scale (HISDS-III), Brain Injury Grief Inventory (BIGI), Hospital Anxiety and Depression Scale - Depression, Rosenberg Self-Esteem Scale (RSES) and the Awareness Questionnaire (Self/Significant other/Clinician versions). The main findings were that participants reported significant changes in self-concept with current self being viewed negatively in comparison to pre-injury self. Perceived identity change was positively associated with depression and grief and negatively associated with self-esteem and awareness. Awareness was negatively associated with self-esteem and positively associated with depression. These findings were consistent with previous research, revealing changes in identity following TBI. Further research is needed to increase our understanding of the psychological factors involved in emotional adjustment after TBI and to inform brain injury rehabilitation interventions, including psychotherapy approaches.

  15. Implementation of 5-layer thermal diffusion scheme in weather research and forecasting model with Intel Many Integrated Cores

    NASA Astrophysics Data System (ADS)

    Huang, Melin; Huang, Bormin; Huang, Allen H.

    2014-10-01

    For weather forecasting and research, the Weather Research and Forecasting (WRF) model has been developed, consisting of several components such as dynamic solvers and physical simulation modules. WRF includes several Land- Surface Models (LSMs). The LSMs use atmospheric information, the radiative and precipitation forcing from the surface layer scheme, the radiation scheme, and the microphysics/convective scheme all together with the land's state variables and land-surface properties, to provide heat and moisture fluxes over land and sea-ice points. The WRF 5-layer thermal diffusion simulation is an LSM based on the MM5 5-layer soil temperature model with an energy budget that includes radiation, sensible, and latent heat flux. The WRF LSMs are very suitable for massively parallel computation as there are no interactions among horizontal grid points. The features, efficient parallelization and vectorization essentials, of Intel Many Integrated Core (MIC) architecture allow us to optimize this WRF 5-layer thermal diffusion scheme. In this work, we present the results of the computing performance on this scheme with Intel MIC architecture. Our results show that the MIC-based optimization improved the performance of the first version of multi-threaded code on Xeon Phi 5110P by a factor of 2.1x. Accordingly, the same CPU-based optimizations improved the performance on Intel Xeon E5- 2603 by a factor of 1.6x as compared to the first version of multi-threaded code.

  16. SIAM Conference on Parallel Processing for Scientific Computing, 4th, Chicago, IL, Dec. 11-13, 1989, Proceedings

    NASA Technical Reports Server (NTRS)

    Dongarra, Jack (Editor); Messina, Paul (Editor); Sorensen, Danny C. (Editor); Voigt, Robert G. (Editor)

    1990-01-01

    Attention is given to such topics as an evaluation of block algorithm variants in LAPACK and presents a large-grain parallel sparse system solver, a multiprocessor method for the solution of the generalized Eigenvalue problem on an interval, and a parallel QR algorithm for iterative subspace methods on the CM2. A discussion of numerical methods includes the topics of asynchronous numerical solutions of PDEs on parallel computers, parallel homotopy curve tracking on a hypercube, and solving Navier-Stokes equations on the Cedar Multi-Cluster system. A section on differential equations includes a discussion of a six-color procedure for the parallel solution of elliptic systems using the finite quadtree structure, data parallel algorithms for the finite element method, and domain decomposition methods in aerodynamics. Topics dealing with massively parallel computing include hypercube vs. 2-dimensional meshes and massively parallel computation of conservation laws. Performance and tools are also discussed.

  17. Enhancing Image Processing Performance for PCID in a Heterogeneous Network of Multi-core Processors

    DTIC Science & Technology

    2009-09-01

    TFLOPS of Playstation 3 (PS3) nodes with IBM Cell Broadband Engine multi-cores and 15 dual-quad Xeon head nodes. The interconnect fabric includes... 4   3. INFORMATION MANAGEMENT FOR PARALLELIZATION AND...STREAMING............................................................. 7  4 . RESULTS

  18. Reconfigurable microfluidic hanging drop network for multi-tissue interaction and analysis.

    PubMed

    Frey, Olivier; Misun, Patrick M; Fluri, David A; Hengstler, Jan G; Hierlemann, Andreas

    2014-06-30

    Integration of multiple three-dimensional microtissues into microfluidic networks enables new insights in how different organs or tissues of an organism interact. Here, we present a platform that extends the hanging-drop technology, used for multi-cellular spheroid formation, to multifunctional complex microfluidic networks. Engineered as completely open, 'hanging' microfluidic system at the bottom of a substrate, the platform features high flexibility in microtissue arrangements and interconnections, while fabrication is simple and operation robust. Multiple spheroids of different cell types are formed in parallel on the same platform; the different tissues are then connected in physiological order for multi-tissue experiments through reconfiguration of the fluidic network. Liquid flow is precisely controlled through the hanging drops, which enable nutrient supply, substance dosage and inter-organ metabolic communication. The possibility to perform parallelized microtissue formation on the same chip that is subsequently used for complex multi-tissue experiments renders the developed platform a promising technology for 'body-on-a-chip'-related research.

  19. An in situ Comparison of Electron Acceleration at Collisionless Shocks under Differing Upstream Magnetic Field Orientations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Masters, A.; Dougherty, M. K.; Sulaiman, A. H.

    A leading explanation for the origin of Galactic cosmic rays is acceleration at high-Mach number shock waves in the collisionless plasma surrounding young supernova remnants. Evidence for this is provided by multi-wavelength non-thermal emission thought to be associated with ultrarelativistic electrons at these shocks. However, the dependence of the electron acceleration process on the orientation of the upstream magnetic field with respect to the local normal to the shock front (quasi-parallel/quasi-perpendicular) is debated. Cassini spacecraft observations at Saturn’s bow shock have revealed examples of electron acceleration under quasi-perpendicular conditions, and the first in situ evidence of electron acceleration at amore » quasi-parallel shock. Here we use Cassini data to make the first comparison between energy spectra of locally accelerated electrons under these differing upstream magnetic field regimes. We present data taken during a quasi-perpendicular shock crossing on 2008 March 8 and during a quasi-parallel shock crossing on 2007 February 3, highlighting that both were associated with electron acceleration to at least MeV energies. The magnetic signature of the quasi-perpendicular crossing has a relatively sharp upstream–downstream transition, and energetic electrons were detected close to the transition and immediately downstream. The magnetic transition at the quasi-parallel crossing is less clear, energetic electrons were encountered upstream and downstream, and the electron energy spectrum is harder above ∼100 keV. We discuss whether the acceleration is consistent with diffusive shock acceleration theory in each case, and suggest that the quasi-parallel spectral break is due to an energy-dependent interaction between the electrons and short, large-amplitude magnetic structures.« less

  20. Hierarchical fractional-step approximations and parallel kinetic Monte Carlo algorithms

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Arampatzis, Giorgos, E-mail: garab@math.uoc.gr; Katsoulakis, Markos A., E-mail: markos@math.umass.edu; Plechac, Petr, E-mail: plechac@math.udel.edu

    2012-10-01

    We present a mathematical framework for constructing and analyzing parallel algorithms for lattice kinetic Monte Carlo (KMC) simulations. The resulting algorithms have the capacity to simulate a wide range of spatio-temporal scales in spatially distributed, non-equilibrium physiochemical processes with complex chemistry and transport micro-mechanisms. Rather than focusing on constructing exactly the stochastic trajectories, our approach relies on approximating the evolution of observables, such as density, coverage, correlations and so on. More specifically, we develop a spatial domain decomposition of the Markov operator (generator) that describes the evolution of all observables according to the kinetic Monte Carlo algorithm. This domain decompositionmore » corresponds to a decomposition of the Markov generator into a hierarchy of operators and can be tailored to specific hierarchical parallel architectures such as multi-core processors or clusters of Graphical Processing Units (GPUs). Based on this operator decomposition, we formulate parallel Fractional step kinetic Monte Carlo algorithms by employing the Trotter Theorem and its randomized variants; these schemes, (a) are partially asynchronous on each fractional step time-window, and (b) are characterized by their communication schedule between processors. The proposed mathematical framework allows us to rigorously justify the numerical and statistical consistency of the proposed algorithms, showing the convergence of our approximating schemes to the original serial KMC. The approach also provides a systematic evaluation of different processor communicating schedules. We carry out a detailed benchmarking of the parallel KMC schemes using available exact solutions, for example, in Ising-type systems and we demonstrate the capabilities of the method to simulate complex spatially distributed reactions at very large scales on GPUs. Finally, we discuss work load balancing between processors and propose a re-balancing scheme based on probabilistic mass transport methods.« less

  1. Observation of layered antiferromagnetism in self-assembled parallel NiSi nanowire arrays on Si(110) by spin-polarized scanning tunneling spectromicroscopy

    NASA Astrophysics Data System (ADS)

    Hong, Ie-Hong; Hsu, Hsin-Zan

    2018-03-01

    The layered antiferromagnetism of parallel nanowire (NW) arrays self-assembled on Si(110) have been observed at room temperature by direct imaging of both the topographies and magnetic domains using spin-polarized scanning tunneling microscopy/spectroscopy (SP-STM/STS). The topographic STM images reveal that the self-assembled unidirectional and parallel NiSi NWs grow into the Si(110) substrate along the [\\bar{1}10] direction (i.e. the endotaxial growth) and exhibit multiple-layer growth. The spatially-resolved SP-STS maps show that these parallel NiSi NWs of different heights produce two opposite magnetic domains, depending on the heights of either even or odd layers in the layer stack of the NiSi NWs. This layer-wise antiferromagnetic structure can be attributed to an antiferromagnetic interlayer exchange coupling between the adjacent layers in the multiple-layer NiSi NW with a B2 (CsCl-type) crystal structure. Such an endotaxial heterostructure of parallel magnetic NiSi NW arrays with a layered antiferromagnetic ordering in Si(110) provides a new and important perspective for the development of novel Si-based spintronic nanodevices.

  2. Scalable Triadic Analysis of Large-Scale Graphs: Multi-Core vs. Multi-Processor vs. Multi-Threaded Shared Memory Architectures

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chin, George; Marquez, Andres; Choudhury, Sutanay

    2012-09-01

    Triadic analysis encompasses a useful set of graph mining methods that is centered on the concept of a triad, which is a subgraph of three nodes and the configuration of directed edges across the nodes. Such methods are often applied in the social sciences as well as many other diverse fields. Triadic methods commonly operate on a triad census that counts the number of triads of every possible edge configuration in a graph. Like other graph algorithms, triadic census algorithms do not scale well when graphs reach tens of millions to billions of nodes. To enable the triadic analysis ofmore » large-scale graphs, we developed and optimized a triad census algorithm to efficiently execute on shared memory architectures. We will retrace the development and evolution of a parallel triad census algorithm. Over the course of several versions, we continually adapted the code’s data structures and program logic to expose more opportunities to exploit parallelism on shared memory that would translate into improved computational performance. We will recall the critical steps and modifications that occurred during code development and optimization. Furthermore, we will compare the performances of triad census algorithm versions on three specific systems: Cray XMT, HP Superdome, and AMD multi-core NUMA machine. These three systems have shared memory architectures but with markedly different hardware capabilities to manage parallelism.« less

  3. Parallelising a molecular dynamics algorithm on a multi-processor workstation

    NASA Astrophysics Data System (ADS)

    Müller-Plathe, Florian

    1990-12-01

    The Verlet neighbour-list algorithm is parallelised for a multi-processor Hewlett-Packard/Apollo DN10000 workstation. The implementation makes use of memory shared between the processors. It is a genuine master-slave approach by which most of the computational tasks are kept in the master process and the slaves are only called to do part of the nonbonded forces calculation. The implementation features elements of both fine-grain and coarse-grain parallelism. Apart from three calls to library routines, two of which are standard UNIX calls, and two machine-specific language extensions, the whole code is written in standard Fortran 77. Hence, it may be expected that this parallelisation concept can be transfered in parts or as a whole to other multi-processor shared-memory computers. The parallel code is routinely used in production work.

  4. Coordination between Generation and Transmission Maintenance Scheduling by Means of Multi-agent Technique

    NASA Astrophysics Data System (ADS)

    Nagata, Takeshi; Tao, Yasuhiro; Utatani, Masahiro; Sasaki, Hiroshi; Fujita, Hideki

    This paper proposes a multi-agent approach to maintenance scheduling in restructured power systems. The restructuring of electric power industry has resulted in market-based approaches for unbundling a multitude of service provided by self-interested entities such as power generating companies (GENCOs), transmission providers (TRANSCOs) and distribution companies (DISCOs). The Independent System Operator (ISO) is responsible for the security of the system operation. The schedule submitted to ISO by GENCOs and TRANSCOs should satisfy security and reliability constraints. The proposed method consists of several GENCO Agents (GAGs), TARNSCO Agents (TAGs) and a ISO Agent(IAG). The IAG’s role in maintenance scheduling is limited to ensuring that the submitted schedules do not cause transmission congestion or endanger the system reliability. From the simulation results, it can be seen the proposed multi-agent approach could coordinate between generation and transmission maintenance schedules.

  5. Parallel phase-shifting self-interference digital holography with faithful reconstruction using compressive sensing

    NASA Astrophysics Data System (ADS)

    Wan, Yuhong; Man, Tianlong; Wu, Fan; Kim, Myung K.; Wang, Dayong

    2016-11-01

    We present a new self-interference digital holographic approach that allows single-shot capturing three-dimensional intensity distribution of the spatially incoherent objects. The Fresnel incoherent correlation holographic microscopy is combined with parallel phase-shifting technique to instantaneously obtain spatially multiplexed phase-shifting holograms. The compressive-sensing-based reconstruction algorithm is implemented to reconstruct the original object from the under sampled demultiplexed holograms. The scheme is verified with simulations. The validity of the proposed method is experimentally demonstrated in an indirectly way by simulating the use of specific parallel phase-shifting recording device.

  6. Ohm's law for a current sheet

    NASA Technical Reports Server (NTRS)

    Lyons, L. R.; Speiser, T. W.

    1985-01-01

    The paper derives an Ohm's law for single-particle motion in a current sheet, where the magnetic field reverses in direction across the sheet. The result is considerably different from the resistive Ohm's law often used in MHD studies of the geomagnetic tail. Single-particle analysis is extended to obtain a self-consistency relation for a current sheet which agrees with previous results. The results are applicable to the concept of reconnection in that the electric field parallel to the current is obtained for a one-dimensional current sheet with constant normal magnetic field. Dissipated energy goes directly into accelerating particles within the current sheet.

  7. Evolution of large amplitude Alfven waves in solar wind plasmas: Kinetic-fluid models

    NASA Astrophysics Data System (ADS)

    Nariyuki, Y.

    2014-12-01

    Large amplitude Alfven waves are ubiquitously observed in solar wind plasmas. Mjolhus(JPP, 1976) and Mio et al(JPSJ, 1976) found that nonlinear evolution of the uni-directional, parallel propagating Alfven waves can be described by the derivative nonlinear Schrodinger equation (DNLS). Later, the multi-dimensional extension (Mjolhus and Wyller, JPP, 1988; Passot and Sulem, POP, 1993; Gazol et al, POP, 1999) and ion kinetic modification (Mjolhus and Wyller, JPP, 1988; Spangler, POP, 1989; Medvedev and Diamond, POP, 1996; Nariyuki et al, POP, 2013) of DNLS have been reported. Recently, Nariyuki derived multi-dimensional DNLS from an expanding box model of the Hall-MHD system (Nariyuki, submitted). The set of equations including the nonlinear evolution of compressional wave modes (TDNLS) was derived by Hada(GRL, 1993). DNLS can be derived from TDNLS by rescaling of the variables (Mjolhus, Phys. Scr., 2006). Nariyuki and Hada(JPSJ, 2007) derived a kinetically modified TDNLS by using a simple Landau closure (Hammet and Perkins, PRL, 1990; Medvedev and Diamond, POP, 1996). In the present study, we revisit the ion kinetic modification of multi-dimensional TDNLS through more rigorous derivations, which is consistent with the past kinetic modification of DNLS. Although the original TDNLS was derived in the multi-dimensional form, the evolution of waves with finite propagation angles in TDNLS has not been paid much attention. Applicability of the resultant models to solar wind turbulence is discussed.

  8. A Generic Mesh Data Structure with Parallel Applications

    ERIC Educational Resources Information Center

    Cochran, William Kenneth, Jr.

    2009-01-01

    High performance, massively-parallel multi-physics simulations are built on efficient mesh data structures. Most data structures are designed from the bottom up, focusing on the implementation of linear algebra routines. In this thesis, we explore a top-down approach to design, evaluating the various needs of many aspects of simulation, not just…

  9. Hierarchial parallel computer architecture defined by computational multidisciplinary mechanics

    NASA Technical Reports Server (NTRS)

    Padovan, Joe; Gute, Doug; Johnson, Keith

    1989-01-01

    The goal is to develop an architecture for parallel processors enabling optimal handling of multi-disciplinary computation of fluid-solid simulations employing finite element and difference schemes. The goals, philosphical and modeling directions, static and dynamic poly trees, example problems, interpolative reduction, the impact on solvers are shown in viewgraph form.

  10. Parallel computer vision

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Uhr, L.

    1987-01-01

    This book is written by research scientists involved in the development of massively parallel, but hierarchically structured, algorithms, architectures, and programs for image processing, pattern recognition, and computer vision. The book gives an integrated picture of the programs and algorithms that are being developed, and also of the multi-computer hardware architectures for which these systems are designed.

  11. Storing files in a parallel computing system based on user or application specification

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Faibish, Sorin; Bent, John M.; Nick, Jeffrey M.

    2016-03-29

    Techniques are provided for storing files in a parallel computing system based on a user-specification. A plurality of files generated by a distributed application in a parallel computing system are stored by obtaining a specification from the distributed application indicating how the plurality of files should be stored; and storing one or more of the plurality of files in one or more storage nodes of a multi-tier storage system based on the specification. The plurality of files comprise a plurality of complete files and/or a plurality of sub-files. The specification can optionally be processed by a daemon executing on onemore » or more nodes in a multi-tier storage system. The specification indicates how the plurality of files should be stored, for example, identifying one or more storage nodes where the plurality of files should be stored.« less

  12. Development and Validation of the Patient Experience with Treatment and Self-Management (PETS): A Patient-Reported Measure of Treatment Burden

    PubMed Central

    Eton, David T.; Yost, Kathleen J.; Lai, Jin-shei; Ridgeway, Jennifer L.; Egginton, Jason S.; Rosedahl, Jordan K.; Linzer, Mark; Boehm, Deborah H.; Thakur, Azra; Poplau, Sara; Odell, Laura; Montori, Victor M.; May, Carl R.; Anderson, Roger T.

    2017-01-01

    Purpose The purpose of this study was to develop and validate a new comprehensive patient-reported measure of treatment burden – the Patient Experience with Treatment and Self-Management (PETS). Methods A conceptual framework was used to derive the PETS with items reviewed and cognitively tested with patients. A survey battery, including a pilot version of the PETS, was mailed to 838 multi-morbid patients from two healthcare institutions for validation. Results A total of 332 multi-morbid patients returned completed surveys. Diagnostics supported deletion and consolidation of some items and domains. Confirmatory factor analysis supported a domain model for scaling comprised of 9 factors: medical information, medications, medical appointments, monitoring health, interpersonal challenges, medical/healthcare expenses, difficulty with healthcare services, role/social activity limitations, and physical/mental exhaustion. Scales showed good internal consistency (alpha range: 0.79 – 0.95). Higher PETS scores, indicative of greater treatment burden, were correlated with more distress, less satisfaction with medications, lower self-efficacy, worse physical and mental health, and lower convenience of healthcare (Ps<.001). Patients with lower health literacy, less adherence to medications, and more financial difficulties reported higher PETS scores (Ps<.01). Conclusion A comprehensive patient-reported measure of treatment burden can help to better characterize the impact of treatment and self-management burden on patient well-being and guide care toward minimally disruptive medicine. PMID:27566732

  13. Development and validation of the Patient Experience with Treatment and Self-management (PETS): a patient-reported measure of treatment burden.

    PubMed

    Eton, David T; Yost, Kathleen J; Lai, Jin-Shei; Ridgeway, Jennifer L; Egginton, Jason S; Rosedahl, Jordan K; Linzer, Mark; Boehm, Deborah H; Thakur, Azra; Poplau, Sara; Odell, Laura; Montori, Victor M; May, Carl R; Anderson, Roger T

    2017-02-01

    The purpose of this study was to develop and validate a new comprehensive patient-reported measure of treatment burden-the Patient Experience with Treatment and Self-management (PETS). A conceptual framework was used to derive the PETS with items reviewed and cognitively tested with patients. A survey battery, including a pilot version of the PETS, was mailed to 838 multi-morbid patients from two healthcare institutions for validation. A total of 332 multi-morbid patients returned completed surveys. Diagnostics supported deletion and consolidation of some items and domains. Confirmatory factor analysis supported a domain model for scaling comprised of 9 factors: medical information, medications, medical appointments, monitoring health, interpersonal challenges, medical/healthcare expenses, difficulty with healthcare services, role/social activity limitations, and physical/mental exhaustion. Scales showed good internal consistency (α range 0.79-0.95). Higher PETS scores, indicative of greater treatment burden, were correlated with more distress, less satisfaction with medications, lower self-efficacy, worse physical and mental health, and lower convenience of healthcare (Ps < 0.001). Patients with lower health literacy, less adherence to medications, and more financial difficulties reported higher PETS scores (Ps < 0.01). A comprehensive patient-reported measure of treatment burden can help to better characterize the impact of treatment and self-management burden on patient well-being and guide care toward minimally disruptive medicine.

  14. Self-tuning multivariable pole placement control of a multizone crystal growth furnace

    NASA Technical Reports Server (NTRS)

    Batur, C.; Sharpless, R. B.; Duval, W. M. B.; Rosenthal, B. N.

    1992-01-01

    This paper presents the design and implementation of a multivariable self-tuning temperature controller for the control of lead bromide crystal growth. The crystal grows inside a multizone transparent furnace. There are eight interacting heating zones shaping the axial temperature distribution inside the furnace. A multi-input, multi-output furnace model is identified on-line by a recursive least squares estimation algorithm. A multivariable pole placement controller based on this model is derived and implemented. Comparison between single-input, single-output and multi-input, multi-output self-tuning controllers demonstrates that the zone-to-zone interactions can be minimized better by a multi-input, multi-output controller design. This directly affects the quality of crystal grown.

  15. Incorporating Cultural Perspectives into Diabetes Self-Management Programs for East Asian Immigrants: A Mixed-Study Review.

    PubMed

    Park, Chorong; Nam, Soohyun; Whittemore, Robin

    2016-04-01

    It is important to understand East Asian immigrants (EAIs)' unique perspectives in managing diabetes in order to provide culturally-competent care. However, it is not known whether EAIs' perspectives are addressed in diabetes self-management interventions developed for EAIs. Therefore, a mixed-study review was conducted to identify EAIs' perspective from qualitative research (n = 9 studies) and to evaluate the components of EAI diabetes self-management interventions (n = 7). Themes from the qualitative synthesis demonstrated that EAIs have unique cultural values and traditional health beliefs while struggling with multi-contextual barriers due to immigration. The evaluation of EAI diabetes self-management interventions revealed that there was a lack of consensus on cultural strategies for EAIs' across the interventions. Addressing language barriers was the only factor consistently integrated in the cultural components of intervention by employing bilingual interventionists. EAIs' perspectives and experiences need to be incorporated in the future diabetes self-management interventions to better provide culturally-competent care.

  16. Artificial dielectric stepped-refractive-index lens for the terahertz region.

    PubMed

    Hernandez-Serrano, A I; Mendis, Rajind; Reichel, Kimberly S; Zhang, Wei; Castro-Camus, E; Mittleman, Daniel M

    2018-02-05

    In this paper we theoretically and experimentally demonstrate a stepped-refractive-index convergent lens made of a parallel stack of metallic plates for terahertz frequencies based on artificial dielectrics. The lens consist of a non-uniformly spaced stack of metallic plates, forming a mirror-symmetric array of parallel-plate waveguides (PPWGs). The operation of the device is based on the TE 1 mode of the PPWG. The effective refractive index of the TE 1 mode is a function of the frequency of operation and the spacing between the plates of the PPWG. By varying the spacing between the plates, we can modify the local refractive index of the structure in every individual PPWG that constitutes the lens producing a stepped refractive index profile across the multi stack structure. The theoretical and experimental results show that this structure is capable of focusing a 1 cm diameter beam to a line focus of less than 4 mm for the design frequency of 0.18 THz. This structure shows that this artificial-dielectric concept is an important technology for the fabrication of next generation terahertz devices.

  17. Harmony search algorithm: application to the redundancy optimization problem

    NASA Astrophysics Data System (ADS)

    Nahas, Nabil; Thien-My, Dao

    2010-09-01

    The redundancy optimization problem is a well known NP-hard problem which involves the selection of elements and redundancy levels to maximize system performance, given different system-level constraints. This article presents an efficient algorithm based on the harmony search algorithm (HSA) to solve this optimization problem. The HSA is a new nature-inspired algorithm which mimics the improvization process of music players. Two kinds of problems are considered in testing the proposed algorithm, with the first limited to the binary series-parallel system, where the problem consists of a selection of elements and redundancy levels used to maximize the system reliability given various system-level constraints; the second problem for its part concerns the multi-state series-parallel systems with performance levels ranging from perfect operation to complete failure, and in which identical redundant elements are included in order to achieve a desirable level of availability. Numerical results for test problems from previous research are reported and compared. The results of HSA showed that this algorithm could provide very good solutions when compared to those obtained through other approaches.

  18. DSPCP: A Data Scalable Approach for Identifying Relationships in Parallel Coordinates.

    PubMed

    Nguyen, Hoa; Rosen, Paul

    2018-03-01

    Parallel coordinates plots (PCPs) are a well-studied technique for exploring multi-attribute datasets. In many situations, users find them a flexible method to analyze and interact with data. Unfortunately, using PCPs becomes challenging as the number of data items grows large or multiple trends within the data mix in the visualization. The resulting overdraw can obscure important features. A number of modifications to PCPs have been proposed, including using color, opacity, smooth curves, frequency, density, and animation to mitigate this problem. However, these modified PCPs tend to have their own limitations in the kinds of relationships they emphasize. We propose a new data scalable design for representing and exploring data relationships in PCPs. The approach exploits the point/line duality property of PCPs and a local linear assumption of data to extract and represent relationship summarizations. This approach simultaneously shows relationships in the data and the consistency of those relationships. Our approach supports various visualization tasks, including mixed linear and nonlinear pattern identification, noise detection, and outlier detection, all in large data. We demonstrate these tasks on multiple synthetic and real-world datasets.

  19. A multi-platform evaluation of the randomized CX low-rank matrix factorization in Spark

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gittens, Alex; Kottalam, Jey; Yang, Jiyan

    We investigate the performance and scalability of the randomized CX low-rank matrix factorization and demonstrate its applicability through the analysis of a 1TB mass spectrometry imaging (MSI) dataset, using Apache Spark on an Amazon EC2 cluster, a Cray XC40 system, and an experimental Cray cluster. We implemented this factorization both as a parallelized C implementation with hand-tuned optimizations and in Scala using the Apache Spark high-level cluster computing framework. We obtained consistent performance across the three platforms: using Spark we were able to process the 1TB size dataset in under 30 minutes with 960 cores on all systems, with themore » fastest times obtained on the experimental Cray cluster. In comparison, the C implementation was 21X faster on the Amazon EC2 system, due to careful cache optimizations, bandwidth-friendly access of matrices and vector computation using SIMD units. We report these results and their implications on the hardware and software issues arising in supporting data-centric workloads in parallel and distributed environments.« less

  20. Self-Locking Optoelectronic Tweezers for Single-Cell and Microparticle Manipulation across a Large Area in High Conductivity Media

    PubMed Central

    Yang, Yajia; Mao, Yufei; Shin, Kyeong-Sik; Chui, Chi On; Chiou, Pei-Yu

    2016-01-01

    Optoelectronic tweezers (OET) has advanced within the past decade to become a promising tool for cell and microparticle manipulation. Its incompatibility with high conductivity media and limited throughput remain two major technical challenges. Here a novel manipulation concept and corresponding platform called Self-Locking Optoelectronic Tweezers (SLOT) are proposed and demonstrated to tackle these challenges concurrently. The SLOT platform comprises a periodic array of optically tunable phototransistor traps above which randomly dispersed single cells and microparticles are self-aligned to and retained without light illumination. Light beam illumination on a phototransistor turns off the trap and releases the trapped cell, which is then transported downstream via a background flow. The cell trapping and releasing functions in SLOT are decoupled, which is a unique feature that enables SLOT’s stepper-mode function to overcome the small field-of-view issue that all prior OET technologies encountered in manipulation with single-cell resolution across a large area. Massively parallel trapping of more than 100,000 microparticles has been demonstrated in high conductivity media. Even larger scale trapping and manipulation can be achieved by linearly scaling up the number of phototransistors and device area. Cells after manipulation on the SLOT platform maintain high cell viability and normal multi-day divisibility. PMID:26940301

  1. Combined collapse by bridging and self-adhesion in a prototypical polymer model inspired by the bacterial nucleoid

    NASA Astrophysics Data System (ADS)

    Scolari, Vittore F.; Cosentino Lagomarsino, Marco

    Recent experimental results suggest that the E. coli chromosome feels a self-attracting interaction of osmotic origin, and is condensed in foci by bridging interactions. Motivated by these findings, we explore a generic modeling framework combining solely these two ingredients, in order to characterize their joint effects. Specifically, we study a simple polymer physics computational model with weak ubiquitous short-ranged self attraction and stronger sparse bridging interactions. Combining theoretical arguments and simulations, we study the general phenomenology of polymer collapse induced by these dual contributions, in the case of regularly-spaced bridging. Our results distinguish a regime of classical Flory-like coil-globule collapse dictated by the interplay of excluded volume and attractive energy and a switch-like collapse where bridging interaction compete with entropy loss terms from the looped arms of a star-like rosette. Additionally, we show that bridging can induce stable compartmentalized domains. In these configurations, different "cores" of bridging proteins are kept separated by star-like polymer loops in an entropically favorable multi-domain configuration, with a mechanism that parallels micellar polysoaps. Such compartmentalized domains are stable, and do not need any intra-specific interactions driving their segregation. Domains can be stable also in presence of uniform attraction, as long as the uniform collapse is above its theta point.

  2. Data Acquisition System for Multi-Frequency Radar Flight Operations Preparation

    NASA Technical Reports Server (NTRS)

    Leachman, Jonathan

    2010-01-01

    A three-channel data acquisition system was developed for the NASA Multi-Frequency Radar (MFR) system. The system is based on a commercial-off-the-shelf (COTS) industrial PC (personal computer) and two dual-channel 14-bit digital receiver cards. The decimated complex envelope representations of the three radar signals are passed to the host PC via the PCI bus, and then processed in parallel by multiple cores of the PC CPU (central processing unit). The innovation is this parallelization of the radar data processing using multiple cores of a standard COTS multi-core CPU. The data processing portion of the data acquisition software was built using autonomous program modules or threads, which can run simultaneously on different cores. A master program module calculates the optimal number of processing threads, launches them, and continually supplies each with data. The benefit of this new parallel software architecture is that COTS PCs can be used to implement increasingly complex processing algorithms on an increasing number of radar range gates and data rates. As new PCs become available with higher numbers of CPU cores, the software will automatically utilize the additional computational capacity.

  3. Optimization of the coherence function estimation for multi-core central processing unit

    NASA Astrophysics Data System (ADS)

    Cheremnov, A. G.; Faerman, V. A.; Avramchuk, V. S.

    2017-02-01

    The paper considers use of parallel processing on multi-core central processing unit for optimization of the coherence function evaluation arising in digital signal processing. Coherence function along with other methods of spectral analysis is commonly used for vibration diagnosis of rotating machinery and its particular nodes. An algorithm is given for the function evaluation for signals represented with digital samples. The algorithm is analyzed for its software implementation and computational problems. Optimization measures are described, including algorithmic, architecture and compiler optimization, their results are assessed for multi-core processors from different manufacturers. Thus, speeding-up of the parallel execution with respect to sequential execution was studied and results are presented for Intel Core i7-4720HQ и AMD FX-9590 processors. The results show comparatively high efficiency of the optimization measures taken. In particular, acceleration indicators and average CPU utilization have been significantly improved, showing high degree of parallelism of the constructed calculating functions. The developed software underwent state registration and will be used as a part of a software and hardware solution for rotating machinery fault diagnosis and pipeline leak location with acoustic correlation method.

  4. Full pillar extraction at the Kathleen Mine with mobile roof supports

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Grimm, E.S.

    1994-12-31

    The Voest Alpine Breaker Line Supports (ABLS) resemble self-propelled longwall shields. Each individual unit consists of four hydraulic legs extending from the base of the unit, pressing a solid flat canopy against the mine roof. Each support unit is capable of exerting 606 tons of force against the roof. A chain curtain on the sides and rear protects the interior of the support from falling rock. The internal scissoring lemniscate design allows for parallel movement of the canopy as it is raised or lowered. Each ABLS has 750 feet of 4 AWG trailing cable to supply 480 volts AC tomore » a permissible controller and a 40 hp explosion-proof electrical motor. The hydraulic pump and reservoir are self-contained and protected with an automatic fire suppression system.« less

  5. Fabrication of 3D polymer photonic crystals for near-IR applications

    NASA Astrophysics Data System (ADS)

    Yao, Peng; Qiu, Liang; Shi, Shouyuan; Schneider, Garrett J.; Prather, Dennis W.; Sharkawy, Ahmed; Kelmelis, Eric

    2008-02-01

    Photonic crystals[1, 2] have stirred enormous research interest and became a growing enterprise in the last 15 years. Generally, PhCs consist of periodic structures that possess periodicity comparable with the wavelength that the PhCs are designed to modulate. If material and periodic pattern are properly selected, PhCs can be applied to many applications based on their unique properties, including photonic band gaps (PBG)[3], self-collimation[4], super prism[5], etc. Strictly speaking, PhCs need to possess periodicity in three dimensions to maximize their advantageous capabilities. However, many current research is based on scaled two-dimensional PhCs, mainly due to the difficulty of fabrication such three-dimensional PhCs. Many approaches have been explored for the fabrication of 3D photonic crystals, including layer-by-layer surface micromachining[6], glancing angle deposition[7], 3D micro-sculpture method[8], self-assembly[9] and lithographical methods[10-12]. Among them, lithographic methods became increasingly accepted due to low costs and precise control over the photonic crystal structure. There are three mostly developed lithographical methods, namely X-ray lithography[10], holographic lithography[11] and two-photon polymerization[12]. Although significant progress has been made in developing these lithography-based technologies, these approaches still suffer from significant disadvantages. X-ray lithography relies on an expensive radiation source. Holographic lithography lacks the flexibility to create engineered defects, and multi-photon polymerization is not suitable for parallel fabrication. In our previous work, we developed a multi-layer photolithography processes[13, 14] that is based on multiple resist application and enhanced absorption upon exposure. Using a negative lift-off resist (LOR) and 254nm DUV source, we have demonstrated fabrication of 3D arbitrary structures with feature size of several microns. However, severe intermixing problem occurred as we reduced the lattice constant for near-IR applications. In this work, we address this problem by employing SU8. The exposure is vertically confined by using a mismatched 220nm DUV source. Intermixing problem is eliminated due to more densely crosslinked resist molecules. Using this method, we have demonstrated 3D "woodpile" structure with 1.55μm lattice constant and a 2mm-by-2mm pattern area.

  6. Solvers for $$\\mathcal{O} (N)$$ Electronic Structure in the Strong Scaling Limit

    DOE PAGES

    Bock, Nicolas; Challacombe, William M.; Kale, Laxmikant

    2016-01-26

    Here we present a hybrid OpenMP/Charm\\tt++ framework for solving themore » $$\\mathcal{O} (N)$$ self-consistent-field eigenvalue problem with parallelism in the strong scaling regime, $$P\\gg{N}$$, where $P$ is the number of cores, and $N$ is a measure of system size, i.e., the number of matrix rows/columns, basis functions, atoms, molecules, etc. This result is achieved with a nested approach to spectral projection and the sparse approximate matrix multiply [Bock and Challacombe, SIAM J. Sci. Comput., 35 (2013), pp. C72--C98], and involves a recursive, task-parallel algorithm, often employed by generalized $N$-Body solvers, to occlusion and culling of negligible products in the case of matrices with decay. Lastly, employing classic technologies associated with generalized $N$-Body solvers, including overdecomposition, recursive task parallelism, orderings that preserve locality, and persistence-based load balancing, we obtain scaling beyond hundreds of cores per molecule for small water clusters ([H$${}_2$$O]$${}_N$$, $$N \\in \\{ 30, 90, 150 \\}$$, $$P/N \\approx \\{ 819, 273, 164 \\}$$) and find support for an increasingly strong scalability with increasing system size $N$.« less

  7. Edge gyrokinetic theory and continuum simulations

    NASA Astrophysics Data System (ADS)

    Xu, X. Q.; Xiong, Z.; Dorr, M. R.; Hittinger, J. A.; Bodi, K.; Candy, J.; Cohen, B. I.; Cohen, R. H.; Colella, P.; Kerbel, G. D.; Krasheninnikov, S.; Nevins, W. M.; Qin, H.; Rognlien, T. D.; Snyder, P. B.; Umansky, M. V.

    2007-08-01

    The following results are presented from the development and application of TEMPEST, a fully nonlinear (full-f) five-dimensional (3d2v) gyrokinetic continuum edge-plasma code. (1) As a test of the interaction of collisions and parallel streaming, TEMPEST is compared with published analytic and numerical results for endloss of particles confined by combined electrostatic and magnetic wells. Good agreement is found over a wide range of collisionality, confining potential and mirror ratio, and the required velocity space resolution is modest. (2) In a large-aspect-ratio circular geometry, excellent agreement is found for a neoclassical equilibrium with parallel ion flow in the banana regime with zero temperature gradient and radial electric field. (3) The four-dimensional (2d2v) version of the code produces the first self-consistent simulation results of collisionless damping of geodesic acoustic modes and zonal flow (Rosenbluth-Hinton residual) with Boltzmann electrons using a full-f code. The electric field is also found to agree with the standard neoclassical expression for steep density and ion temperature gradients in the plateau regime. In divertor geometry, it is found that the endloss of particles and energy induces parallel flow stronger than the core neoclassical predictions in the SOL.

  8. Density-based parallel skin lesion border detection with webCL

    PubMed Central

    2015-01-01

    Background Dermoscopy is a highly effective and noninvasive imaging technique used in diagnosis of melanoma and other pigmented skin lesions. Many aspects of the lesion under consideration are defined in relation to the lesion border. This makes border detection one of the most important steps in dermoscopic image analysis. In current practice, dermatologists often delineate borders through a hand drawn representation based upon visual inspection. Due to the subjective nature of this technique, intra- and inter-observer variations are common. Because of this, the automated assessment of lesion borders in dermoscopic images has become an important area of study. Methods Fast density based skin lesion border detection method has been implemented in parallel with a new parallel technology called WebCL. WebCL utilizes client side computing capabilities to use available hardware resources such as multi cores and GPUs. Developed WebCL-parallel density based skin lesion border detection method runs efficiently from internet browsers. Results Previous research indicates that one of the highest accuracy rates can be achieved using density based clustering techniques for skin lesion border detection. While these algorithms do have unfavorable time complexities, this effect could be mitigated when implemented in parallel. In this study, density based clustering technique for skin lesion border detection is parallelized and redesigned to run very efficiently on the heterogeneous platforms (e.g. tablets, SmartPhones, multi-core CPUs, GPUs, and fully-integrated Accelerated Processing Units) by transforming the technique into a series of independent concurrent operations. Heterogeneous computing is adopted to support accessibility, portability and multi-device use in the clinical settings. For this, we used WebCL, an emerging technology that enables a HTML5 Web browser to execute code in parallel for heterogeneous platforms. We depicted WebCL and our parallel algorithm design. In addition, we tested parallel code on 100 dermoscopy images and showed the execution speedups with respect to the serial version. Results indicate that parallel (WebCL) version and serial version of density based lesion border detection methods generate the same accuracy rates for 100 dermoscopy images, in which mean of border error is 6.94%, mean of recall is 76.66%, and mean of precision is 99.29% respectively. Moreover, WebCL version's speedup factor for 100 dermoscopy images' lesion border detection averages around ~491.2. Conclusions When large amount of high resolution dermoscopy images considered in a usual clinical setting along with the critical importance of early detection and diagnosis of melanoma before metastasis, the importance of fast processing dermoscopy images become obvious. In this paper, we introduce WebCL and the use of it for biomedical image processing applications. WebCL is a javascript binding of OpenCL, which takes advantage of GPU computing from a web browser. Therefore, WebCL parallel version of density based skin lesion border detection introduced in this study can supplement expert dermatologist, and aid them in early diagnosis of skin lesions. While WebCL is currently an emerging technology, a full adoption of WebCL into the HTML5 standard would allow for this implementation to run on a very large set of hardware and software systems. WebCL takes full advantage of parallel computational resources including multi-cores and GPUs on a local machine, and allows for compiled code to run directly from the Web Browser. PMID:26423836

  9. Density-based parallel skin lesion border detection with webCL.

    PubMed

    Lemon, James; Kockara, Sinan; Halic, Tansel; Mete, Mutlu

    2015-01-01

    Dermoscopy is a highly effective and noninvasive imaging technique used in diagnosis of melanoma and other pigmented skin lesions. Many aspects of the lesion under consideration are defined in relation to the lesion border. This makes border detection one of the most important steps in dermoscopic image analysis. In current practice, dermatologists often delineate borders through a hand drawn representation based upon visual inspection. Due to the subjective nature of this technique, intra- and inter-observer variations are common. Because of this, the automated assessment of lesion borders in dermoscopic images has become an important area of study. Fast density based skin lesion border detection method has been implemented in parallel with a new parallel technology called WebCL. WebCL utilizes client side computing capabilities to use available hardware resources such as multi cores and GPUs. Developed WebCL-parallel density based skin lesion border detection method runs efficiently from internet browsers. Previous research indicates that one of the highest accuracy rates can be achieved using density based clustering techniques for skin lesion border detection. While these algorithms do have unfavorable time complexities, this effect could be mitigated when implemented in parallel. In this study, density based clustering technique for skin lesion border detection is parallelized and redesigned to run very efficiently on the heterogeneous platforms (e.g. tablets, SmartPhones, multi-core CPUs, GPUs, and fully-integrated Accelerated Processing Units) by transforming the technique into a series of independent concurrent operations. Heterogeneous computing is adopted to support accessibility, portability and multi-device use in the clinical settings. For this, we used WebCL, an emerging technology that enables a HTML5 Web browser to execute code in parallel for heterogeneous platforms. We depicted WebCL and our parallel algorithm design. In addition, we tested parallel code on 100 dermoscopy images and showed the execution speedups with respect to the serial version. Results indicate that parallel (WebCL) version and serial version of density based lesion border detection methods generate the same accuracy rates for 100 dermoscopy images, in which mean of border error is 6.94%, mean of recall is 76.66%, and mean of precision is 99.29% respectively. Moreover, WebCL version's speedup factor for 100 dermoscopy images' lesion border detection averages around ~491.2. When large amount of high resolution dermoscopy images considered in a usual clinical setting along with the critical importance of early detection and diagnosis of melanoma before metastasis, the importance of fast processing dermoscopy images become obvious. In this paper, we introduce WebCL and the use of it for biomedical image processing applications. WebCL is a javascript binding of OpenCL, which takes advantage of GPU computing from a web browser. Therefore, WebCL parallel version of density based skin lesion border detection introduced in this study can supplement expert dermatologist, and aid them in early diagnosis of skin lesions. While WebCL is currently an emerging technology, a full adoption of WebCL into the HTML5 standard would allow for this implementation to run on a very large set of hardware and software systems. WebCL takes full advantage of parallel computational resources including multi-cores and GPUs on a local machine, and allows for compiled code to run directly from the Web Browser.

  10. Segment Fixed Priority Scheduling for Self Suspending Real Time Tasks

    DTIC Science & Technology

    2016-08-11

    Segment-Fixed Priority Scheduling for Self-Suspending Real -Time Tasks Junsung Kim, Department of Electrical and Computer Engineering, Carnegie...4 2.1 Application of a Multi-Segment Self-Suspending Real -Time Task Model ............................. 5 3 Fixed Priority Scheduling...1 Figure 2: A multi-segment self-suspending real -time task model

  11. A scalable parallel black oil simulator on distributed memory parallel computers

    NASA Astrophysics Data System (ADS)

    Wang, Kun; Liu, Hui; Chen, Zhangxin

    2015-11-01

    This paper presents our work on developing a parallel black oil simulator for distributed memory computers based on our in-house parallel platform. The parallel simulator is designed to overcome the performance issues of common simulators that are implemented for personal computers and workstations. The finite difference method is applied to discretize the black oil model. In addition, some advanced techniques are employed to strengthen the robustness and parallel scalability of the simulator, including an inexact Newton method, matrix decoupling methods, and algebraic multigrid methods. A new multi-stage preconditioner is proposed to accelerate the solution of linear systems from the Newton methods. Numerical experiments show that our simulator is scalable and efficient, and is capable of simulating extremely large-scale black oil problems with tens of millions of grid blocks using thousands of MPI processes on parallel computers.

  12. Parallel Task Management Library for MARTe

    NASA Astrophysics Data System (ADS)

    Valcarcel, Daniel F.; Alves, Diogo; Neto, Andre; Reux, Cedric; Carvalho, Bernardo B.; Felton, Robert; Lomas, Peter J.; Sousa, Jorge; Zabeo, Luca

    2014-06-01

    The Multithreaded Application Real-Time executor (MARTe) is a real-time framework with increasing popularity and support in the thermonuclear fusion community. It allows modular code to run in a multi-threaded environment leveraging on the current multi-core processor (CPU) technology. One application that relies on the MARTe framework is the Joint European Torus (JET) tokamak WAll Load Limiter System (WALLS). It calculates and monitors the temperature on metal tiles and plasma facing components (PFCs) that can melt or flake if their temperature gets too high when exposed to power loads. One of the main time consuming tasks in WALLS is the calculation of thermal diffusion models in real-time. These models tend to be described by very large state-space models thus making them perfect candidates for parallelisation. MARTe's traditional approach for task parallelisation is to split the problem into several Real-Time Threads, each responsible for a self-contained sequential execution of an input-to-output chain. This is usually possible, but it might not always be practical for algorithmic or technical reasons. Also, it might not be easily scalable with an increase in the number of available CPU cores. The WorkLibrary introduces a “GPU-like approach” of splitting work among the available cores of modern CPUs that is (i) straightforward to use in an application, (ii) scalable with the availability of cores and all of this (iii) without rewriting or recompiling the source code. The first part of this article explains the motivation behind the library, its architecture and implementation. The second part presents a real application for WALLS, a parallel version of a large state-space model describing the 2D thermal diffusion on a JET tile.

  13. Results from the OH-PT model: a Kinetic-MHD Model of the Outer Heliosphere within SWMF

    NASA Astrophysics Data System (ADS)

    Michael, A.; Opher, M.; Tenishev, V.; Borovikov, D.; Toth, G.

    2017-12-01

    We present an update of the OH-PT model, a kinetic-MHD model of the outer heliosphere. The OH-PT model couples the Outer Heliosphere (OH) and Particle Tracker (PT) components within the Space Weather Modeling Framework (SWMF). The OH component utilizes the Block-Adaptive Tree Solarwind Roe-type Upwind Scheme (BATS-R-US) MHD code, a highly parallel, 3D, and block-adaptive solver. As a stand-alone model, the OH component solves the ideal MHD equations for the plasma and a separate set of Euler's equations for the different populations of neutral atoms. The neutrals and plasma in the outer heliosphere are coupled through charge-exchange. While this provides an accurate solution for the plasma, it is an inaccurate description of the neutrals. The charge-exchange mean free path is on the order of the size of the heliosphere; therefore the neutrals cannot be described as a fluid. The PT component is based on the Adaptive Mesh Particle Simulator (AMPS) model, a 3D, direct simulation Monte Carlo model that solves the Boltzmann equation for the motion and interaction of multi-species plasma and is used to model the neutral distribution functions throughout the domain. The charge-exchange process occurs within AMPS, which handles each event on a particle-by-particle basis and calculates the resulting source terms to the MHD equations. The OH-PT model combines the MHD solution for the plasma with the kinetic solution for the neutrals to form a self-consistent model of the heliosphere. In this work, we present verification and validation of the model as well as demonstrate the codes capabilities. Furthermore we provide a comparison of the OH-PT model to our multi-fluid approximation and detail the differences between the models in both the plasma solution and neutral distribution functions.

  14. Continuum kinetic and multi-fluid simulations of classical sheaths

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cagas, P.; Hakim, A.; Juno, J.

    The kinetic study of plasma sheaths is critical, among other things, to understand the deposition of heat on walls, the effect of sputtering, and contamination of the plasma with detrimental impurities. The plasma sheath also provides a boundary condition and can often have a significant global impact on the bulk plasma. In this paper, kinetic studies of classical sheaths are performed with the continuum kinetic code, Gkeyll, which directly solves the Vlasov-Maxwell equations. The code uses a novel version of the finite-element discontinuous Galerkin scheme that conserves energy in the continuous-time limit. The fields are computed using Maxwell equations. Ionizationmore » and scattering collisions are included; however, surface effects are neglected. The aim of this work is to introduce the continuum kinetic method and compare its results with those obtained from an already established finite-volume multi-fluid model also implemented in Gkeyll. Novel boundary conditions on the fluids allow the sheath to form without specifying wall fluxes, so the fluids and fields adjust self-consistently at the wall. Our work demonstrates that the kinetic and fluid results are in agreement for the momentum flux, showing that in certain regimes, a multifluid model can be a useful approximation for simulating the plasma boundary. There are differences in the electrostatic potential between the fluid and kinetic results. Further, the direct solutions of the distribution function presented here highlight the non-Maxwellian distribution of electrons in the sheath, emphasizing the need for a kinetic model. The densities, velocities, and the potential show a good agreement between the kinetic and fluid results. But, kinetic physics is highlighted through higher moments such as parallel and perpendicular temperatures which provide significant differences from the fluid results in which the temperature is assumed to be isotropic. Besides decompression cooling, the heat flux is shown to play a role in the temperature differences that are observed, especially inside the collisionless sheath. Published by AIP Publishing.« less

  15. Continuum kinetic and multi-fluid simulations of classical sheaths

    DOE PAGES

    Cagas, P.; Hakim, A.; Juno, J.; ...

    2017-02-21

    The kinetic study of plasma sheaths is critical, among other things, to understand the deposition of heat on walls, the effect of sputtering, and contamination of the plasma with detrimental impurities. The plasma sheath also provides a boundary condition and can often have a significant global impact on the bulk plasma. In this paper, kinetic studies of classical sheaths are performed with the continuum kinetic code, Gkeyll, which directly solves the Vlasov-Maxwell equations. The code uses a novel version of the finite-element discontinuous Galerkin scheme that conserves energy in the continuous-time limit. The fields are computed using Maxwell equations. Ionizationmore » and scattering collisions are included; however, surface effects are neglected. The aim of this work is to introduce the continuum kinetic method and compare its results with those obtained from an already established finite-volume multi-fluid model also implemented in Gkeyll. Novel boundary conditions on the fluids allow the sheath to form without specifying wall fluxes, so the fluids and fields adjust self-consistently at the wall. Our work demonstrates that the kinetic and fluid results are in agreement for the momentum flux, showing that in certain regimes, a multifluid model can be a useful approximation for simulating the plasma boundary. There are differences in the electrostatic potential between the fluid and kinetic results. Further, the direct solutions of the distribution function presented here highlight the non-Maxwellian distribution of electrons in the sheath, emphasizing the need for a kinetic model. The densities, velocities, and the potential show a good agreement between the kinetic and fluid results. But, kinetic physics is highlighted through higher moments such as parallel and perpendicular temperatures which provide significant differences from the fluid results in which the temperature is assumed to be isotropic. Besides decompression cooling, the heat flux is shown to play a role in the temperature differences that are observed, especially inside the collisionless sheath. Published by AIP Publishing.« less

  16. Formation of Electrostatic Potential Drops in the Auroral Zone

    NASA Technical Reports Server (NTRS)

    Schriver, D.; Ashour-Abdalla, M.; Richard, R. L.

    2001-01-01

    In order to examine the self-consistent formation of large-scale quasi-static parallel electric fields in the auroral zone on a micro/meso scale, a particle in cell simulation has been developed. The code resolves electron Debye length scales so that electron micro-processes are included and a variable grid scheme is used such that the overall length scale of the simulation is of the order of an Earth radii along the magnetic field. The simulation is electrostatic and includes the magnetic mirror force, as well as two types of plasmas, a cold dense ionospheric plasma and a warm tenuous magnetospheric plasma. In order to study the formation of parallel electric fields in the auroral zone, different magnetospheric ion and electron inflow boundary conditions are used to drive the system. It has been found that for conditions in the primary (upward) current region an upward directed quasi-static electric field can form across the system due to magnetic mirroring of the magnetospheric ions and electrons at different altitudes. For conditions in the return (downward) current region it is shown that a quasi-static parallel electric field in the opposite sense of that in the primary current region is formed, i.e., the parallel electric field is directed earthward. The conditions for how these different electric fields can be formed are discussed using satellite observations and numerical simulations.

  17. Ion and Electron Energization in Guide Field Reconnection Outflows with Kinetic Riemann Simulations and Parallel Shock Simulations

    NASA Astrophysics Data System (ADS)

    Zhang, Q.; Drake, J. F.; Swisdak, M.

    2017-12-01

    How ions and electrons are energized in magnetic reconnection outflows is an essential topic throughout the heliosphere. Here we carry out guide field PIC Riemann simulations to explore the ion and electron energization mechanisms far downstream of the x-line. Riemann simulations, with their simple magnetic geometry, facilitate the study of the reconnection outflow far downstream of the x-line in much more detail than is possible with conventional reconnection simulations. We find that the ions get accelerated at rotational discontinuities, counter stream, and give rise to two slow shocks. We demonstrate that the energization mechanism at the slow shocks is essentially the same as that of parallel electrostatic shocks. Also, the electron confining electric potential at the slow shocks is driven by the counterstreaming beams, which tend to break the quasi-neutrality. Based on this picture, we build a kinetic model to self consistently predict the downstream ion and electron temperatures. Additional explorations using parallel shock simulations also imply that in a very low beta(0.001 0.01 for a modest guide field) regime, electron energization will be insignificant compared to the ion energization. Our model and the parallel shock simulations might be used as simple tools to understand and estimate the energization of ions and electrons and the energy partition far downstream of the x-line.

  18. Parallelization and checkpointing of GPU applications through program transformation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Solano-Quinde, Lizandro Damian

    2012-01-01

    GPUs have emerged as a powerful tool for accelerating general-purpose applications. The availability of programming languages that makes writing general-purpose applications for running on GPUs tractable have consolidated GPUs as an alternative for accelerating general purpose applications. Among the areas that have benefited from GPU acceleration are: signal and image processing, computational fluid dynamics, quantum chemistry, and, in general, the High Performance Computing (HPC) Industry. In order to continue to exploit higher levels of parallelism with GPUs, multi-GPU systems are gaining popularity. In this context, single-GPU applications are parallelized for running in multi-GPU systems. Furthermore, multi-GPU systems help to solvemore » the GPU memory limitation for applications with large application memory footprint. Parallelizing single-GPU applications has been approached by libraries that distribute the workload at runtime, however, they impose execution overhead and are not portable. On the other hand, on traditional CPU systems, parallelization has been approached through application transformation at pre-compile time, which enhances the application to distribute the workload at application level and does not have the issues of library-based approaches. Hence, a parallelization scheme for GPU systems based on application transformation is needed. Like any computing engine of today, reliability is also a concern in GPUs. GPUs are vulnerable to transient and permanent failures. Current checkpoint/restart techniques are not suitable for systems with GPUs. Checkpointing for GPU systems present new and interesting challenges, primarily due to the natural differences imposed by the hardware design, the memory subsystem architecture, the massive number of threads, and the limited amount of synchronization among threads. Therefore, a checkpoint/restart technique suitable for GPU systems is needed. The goal of this work is to exploit higher levels of parallelism and to develop support for application-level fault tolerance in applications using multiple GPUs. Our techniques reduce the burden of enhancing single-GPU applications to support these features. To achieve our goal, this work designs and implements a framework for enhancing a single-GPU OpenCL application through application transformation.« less

  19. A privacy-preserving parallel and homomorphic encryption scheme

    NASA Astrophysics Data System (ADS)

    Min, Zhaoe; Yang, Geng; Shi, Jingqi

    2017-04-01

    In order to protect data privacy whilst allowing efficient access to data in multi-nodes cloud environments, a parallel homomorphic encryption (PHE) scheme is proposed based on the additive homomorphism of the Paillier encryption algorithm. In this paper we propose a PHE algorithm, in which plaintext is divided into several blocks and blocks are encrypted with a parallel mode. Experiment results demonstrate that the encryption algorithm can reach a speed-up ratio at about 7.1 in the MapReduce environment with 16 cores and 4 nodes.

  20. An OpenACC-Based Unified Programming Model for Multi-accelerator Systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kim, Jungwon; Lee, Seyong; Vetter, Jeffrey S

    2015-01-01

    This paper proposes a novel SPMD programming model of OpenACC. Our model integrates the different granularities of parallelism from vector-level parallelism to node-level parallelism into a single, unified model based on OpenACC. It allows programmers to write programs for multiple accelerators using a uniform programming model whether they are in shared or distributed memory systems. We implement a prototype of our model and evaluate its performance with a GPU-based supercomputer using three benchmark applications.

  1. Operation of high power converters in parallel

    NASA Technical Reports Server (NTRS)

    Decker, D. K.; Inouye, L. Y.

    1993-01-01

    High power converters that are used in space power subsystems are limited in power handling capability due to component and thermal limitations. For applications, such as Space Station Freedom, where multi-kilowatts of power must be delivered to user loads, parallel operation of converters becomes an attractive option when considering overall power subsystem topologies. TRW developed three different unequal power sharing approaches for parallel operation of converters. These approaches, known as droop, master-slave, and proportional adjustment, are discussed and test results are presented.

  2. Image matrix processor for fast multi-dimensional computations

    DOEpatents

    Roberson, George P.; Skeate, Michael F.

    1996-01-01

    An apparatus for multi-dimensional computation which comprises a computation engine, including a plurality of processing modules. The processing modules are configured in parallel and compute respective contributions to a computed multi-dimensional image of respective two dimensional data sets. A high-speed, parallel access storage system is provided which stores the multi-dimensional data sets, and a switching circuit routes the data among the processing modules in the computation engine and the storage system. A data acquisition port receives the two dimensional data sets representing projections through an image, for reconstruction algorithms such as encountered in computerized tomography. The processing modules include a programmable local host, by which they may be configured to execute a plurality of different types of multi-dimensional algorithms. The processing modules thus include an image manipulation processor, which includes a source cache, a target cache, a coefficient table, and control software for executing image transformation routines using data in the source cache and the coefficient table and loading resulting data in the target cache. The local host processor operates to load the source cache with a two dimensional data set, loads the coefficient table, and transfers resulting data out of the target cache to the storage system, or to another destination.

  3. Advanced Numerical Techniques of Performance Evaluation. Volume 1

    DTIC Science & Technology

    1990-06-01

    system scheduling3thread. The scheduling thread then runs any other ready thread that can be found. A thread can only sleep or switch out on itself...Polychronopoulos and D.J. Kuck. Guided Self- Scheduling : A Practical Scheduling Scheme for Parallel Supercomputers. IEEE Transactions on Computers C...Kuck 1987] C.D. Polychronopoulos and D.J. Kuck. Guided Self- Scheduling : A Practical Scheduling Scheme for Parallel Supercomputers. IEEE Trans. on Comp

  4. System, methods and apparatus for program optimization for multi-threaded processor architectures

    DOEpatents

    Bastoul, Cedric; Lethin, Richard A; Leung, Allen K; Meister, Benoit J; Szilagyi, Peter; Vasilache, Nicolas T; Wohlford, David E

    2015-01-06

    Methods, apparatus and computer software product for source code optimization are provided. In an exemplary embodiment, a first custom computing apparatus is used to optimize the execution of source code on a second computing apparatus. In this embodiment, the first custom computing apparatus contains a memory, a storage medium and at least one processor with at least one multi-stage execution unit. The second computing apparatus contains at least two multi-stage execution units that allow for parallel execution of tasks. The first custom computing apparatus optimizes the code for parallelism, locality of operations and contiguity of memory accesses on the second computing apparatus. This Abstract is provided for the sole purpose of complying with the Abstract requirement rules. This Abstract is submitted with the explicit understanding that it will not be used to interpret or to limit the scope or the meaning of the claims.

  5. Rubus: A compiler for seamless and extensible parallelism.

    PubMed

    Adnan, Muhammad; Aslam, Faisal; Nawaz, Zubair; Sarwar, Syed Mansoor

    2017-01-01

    Nowadays, a typical processor may have multiple processing cores on a single chip. Furthermore, a special purpose processing unit called Graphic Processing Unit (GPU), originally designed for 2D/3D games, is now available for general purpose use in computers and mobile devices. However, the traditional programming languages which were designed to work with machines having single core CPUs, cannot utilize the parallelism available on multi-core processors efficiently. Therefore, to exploit the extraordinary processing power of multi-core processors, researchers are working on new tools and techniques to facilitate parallel programming. To this end, languages like CUDA and OpenCL have been introduced, which can be used to write code with parallelism. The main shortcoming of these languages is that programmer needs to specify all the complex details manually in order to parallelize the code across multiple cores. Therefore, the code written in these languages is difficult to understand, debug and maintain. Furthermore, to parallelize legacy code can require rewriting a significant portion of code in CUDA or OpenCL, which can consume significant time and resources. Thus, the amount of parallelism achieved is proportional to the skills of the programmer and the time spent in code optimizations. This paper proposes a new open source compiler, Rubus, to achieve seamless parallelism. The Rubus compiler relieves the programmer from manually specifying the low-level details. It analyses and transforms a sequential program into a parallel program automatically, without any user intervention. This achieves massive speedup and better utilization of the underlying hardware without a programmer's expertise in parallel programming. For five different benchmarks, on average a speedup of 34.54 times has been achieved by Rubus as compared to Java on a basic GPU having only 96 cores. Whereas, for a matrix multiplication benchmark the average execution speedup of 84 times has been achieved by Rubus on the same GPU. Moreover, Rubus achieves this performance without drastically increasing the memory footprint of a program.

  6. Rubus: A compiler for seamless and extensible parallelism

    PubMed Central

    Adnan, Muhammad; Aslam, Faisal; Sarwar, Syed Mansoor

    2017-01-01

    Nowadays, a typical processor may have multiple processing cores on a single chip. Furthermore, a special purpose processing unit called Graphic Processing Unit (GPU), originally designed for 2D/3D games, is now available for general purpose use in computers and mobile devices. However, the traditional programming languages which were designed to work with machines having single core CPUs, cannot utilize the parallelism available on multi-core processors efficiently. Therefore, to exploit the extraordinary processing power of multi-core processors, researchers are working on new tools and techniques to facilitate parallel programming. To this end, languages like CUDA and OpenCL have been introduced, which can be used to write code with parallelism. The main shortcoming of these languages is that programmer needs to specify all the complex details manually in order to parallelize the code across multiple cores. Therefore, the code written in these languages is difficult to understand, debug and maintain. Furthermore, to parallelize legacy code can require rewriting a significant portion of code in CUDA or OpenCL, which can consume significant time and resources. Thus, the amount of parallelism achieved is proportional to the skills of the programmer and the time spent in code optimizations. This paper proposes a new open source compiler, Rubus, to achieve seamless parallelism. The Rubus compiler relieves the programmer from manually specifying the low-level details. It analyses and transforms a sequential program into a parallel program automatically, without any user intervention. This achieves massive speedup and better utilization of the underlying hardware without a programmer’s expertise in parallel programming. For five different benchmarks, on average a speedup of 34.54 times has been achieved by Rubus as compared to Java on a basic GPU having only 96 cores. Whereas, for a matrix multiplication benchmark the average execution speedup of 84 times has been achieved by Rubus on the same GPU. Moreover, Rubus achieves this performance without drastically increasing the memory footprint of a program. PMID:29211758

  7. Variable Swing Optimal Parallel Links - Minimal Power, Maximal Density for Parallel Links

    DTIC Science & Technology

    2009-01-01

    implemented; it allows controlling the transmitter current by a simple design of a differential pair with a 100 ohms termination resistor. Figure 3.4...optimization. Zuber, P., et al. 2005. 0-7695-2288-2. 21. A 36Gb/s ACCI Multi-Channel Bus using a Fully Differential Pulse Receiver. Wilson, Lei Luo

  8. Distributed run of a one-dimensional model in a regional application using SOAP-based web services

    NASA Astrophysics Data System (ADS)

    Smiatek, Gerhard

    This article describes the setup of a distributed computing system in Perl. It facilitates the parallel run of a one-dimensional environmental model on a number of simple network PC hosts. The system uses Simple Object Access Protocol (SOAP) driven web services offering the model run on remote hosts and a multi-thread environment distributing the work and accessing the web services. Its application is demonstrated in a regional run of a process-oriented biogenic emission model for the area of Germany. Within a network consisting of up to seven web services implemented on Linux and MS-Windows hosts, a performance increase of approximately 400% has been reached compared to a model run on the fastest single host.

  9. GPU COMPUTING FOR PARTICLE TRACKING

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nishimura, Hiroshi; Song, Kai; Muriki, Krishna

    2011-03-25

    This is a feasibility study of using a modern Graphics Processing Unit (GPU) to parallelize the accelerator particle tracking code. To demonstrate the massive parallelization features provided by GPU computing, a simplified TracyGPU program is developed for dynamic aperture calculation. Performances, issues, and challenges from introducing GPU are also discussed. General purpose Computation on Graphics Processing Units (GPGPU) bring massive parallel computing capabilities to numerical calculation. However, the unique architecture of GPU requires a comprehensive understanding of the hardware and programming model to be able to well optimize existing applications. In the field of accelerator physics, the dynamic aperture calculationmore » of a storage ring, which is often the most time consuming part of the accelerator modeling and simulation, can benefit from GPU due to its embarrassingly parallel feature, which fits well with the GPU programming model. In this paper, we use the Tesla C2050 GPU which consists of 14 multi-processois (MP) with 32 cores on each MP, therefore a total of 448 cores, to host thousands ot threads dynamically. Thread is a logical execution unit of the program on GPU. In the GPU programming model, threads are grouped into a collection of blocks Within each block, multiple threads share the same code, and up to 48 KB of shared memory. Multiple thread blocks form a grid, which is executed as a GPU kernel. A simplified code that is a subset of Tracy++ [2] is developed to demonstrate the possibility of using GPU to speed up the dynamic aperture calculation by having each thread track a particle.« less

  10. A multi-process model of self-regulation: influences of mindfulness, integrative self-knowledge and self-control in Iran.

    PubMed

    Ghorbani, Nima; Watson, P J; Farhadi, Mehran; Chen, Zhuo

    2014-04-01

    Self-regulation presumably rests upon multiple processes that include an awareness of ongoing self-experience, enduring self-knowledge and self-control. The present investigation tested this multi-process model using the Five-Facet Mindfulness Questionnaire (FFMQ) and the Integrative Self-Knowledge and Brief Self-Control Scales. Using a sample of 1162 Iranian university students, we confirmed the five-factor structure of the FFMQ in Iran and documented its factorial invariance across males and females. Self-regulatory variables correlated negatively with Perceived Stress, Depression, and Anxiety and positively with Self-Esteem and Satisfaction with Life. Partial mediation effects confirmed that self-regulatory measures ameliorated the disturbing effects of Perceived Stress. Integrative Self-Knowledge and Self-Control interacted to partially mediate the association of Perceived Stress with lower levels of Satisfaction with Life. Integrative Self-Knowledge, alone or in interaction with Self-Control, was the only self-regulation variable to display the expected mediation of Perceived Stress associations with all other measures. Self-Control failed to be implicated in self-regulation only in the mediation of Anxiety. These data confirmed the need to further examine this multi-process model of self-regulation. © 2014 International Union of Psychological Science.

  11. Multi-electrolyte-step anodic aluminum oxide method for the fabrication of self-organized nanochannel arrays

    PubMed Central

    2012-01-01

    Nanochannel arrays were fabricated by the self-organized multi-electrolyte-step anodic aluminum oxide [AAO] method in this study. The anodization conditions used in the multi-electrolyte-step AAO method included a phosphoric acid solution as the electrolyte and an applied high voltage. There was a change in the phosphoric acid by the oxalic acid solution as the electrolyte and the applied low voltage. This method was used to produce self-organized nanochannel arrays with good regularity and circularity, meaning less power loss and processing time than with the multi-step AAO method. PMID:22333268

  12. Does cardiovascular reactivity during speech reflect self-construction processes?

    PubMed

    Lyons, A C; Spicer, J; Tuffin, K; Chamberlain, K

    2000-11-01

    Abstract Substantial empirical research has been undertaken on cardiovascular reactivity (CVR). however interpretation of this research is hampered by a lack of theoretical frameworks. This paper develops a framework initially stimulated by evidence demonstrating that the cardiovascular system increases in activity during communication, and that the extent of this activation depends upon numerous and diverse psychosocial factors. We attempt to account for this phenomenon using merit post-structuralist ideas concerning the constructive nature of language and its centrality to an individual's sense of self. Our theoretical framework proposes that the CVR exhibited during language use is explicable in terms of self-construction - From this analysis we hypothesised that CVR would differ across conversations about private self. public self and non-self topics, and that these differences would depend upon people's speaking histories. We found that the blood pressure and heart rate of 102 women was most reactive when they talked in a laboratory with a stranger about aspects of their private self, and least reactive during non-self talk, whilst their heart rate was most reactive during talk about their public self. Overall the results highlight the inextricable link between our inherent socialness and our cardiovascular systems. SUMMARY The explanatory scheme outlined here is an attempt to provide a social reconceptualisation of a phenomenon that is typically interpreted in individualistic psychophysiological terms, and which is consistent with the notion that repeated exposure to situations which provoke large haemodynamic changes may lead to CHD disease progression. The explanation draws heavily on post-structuralist ideas regarding language, and the social constructionist notion that engaging in language use is central to constructing and maintaining a sense of self. This sense of self is a central theoretical entity in our everyday lives, produced and maintained in our interactions with others. We argue that it is this centrality of self-construction that helps to explain the extraordinary consistency of elevated CVR in conversation. Further, we have noted the striking parallels between those features of conversations that make the self salient, and those that have been associated with elevated CVR. To examine it more explicitly, it needs to be tested empirically with new data, using explicitly derived operationalisations and hypotheses.

  13. PARALLEL PERTURBATION MODEL FOR CYCLE TO CYCLE VARIABILITY PPM4CCV

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ameen, Muhsin Mohammed; Som, Sibendu

    This code consists of a Fortran 90 implementation of the parallel perturbation model to compute cyclic variability in spark ignition (SI) engines. Cycle-to-cycle variability (CCV) is known to be detrimental to SI engine operation resulting in partial burn and knock, and result in an overall reduction in the reliability of the engine. Numerical prediction of cycle-to-cycle variability (CCV) in SI engines is extremely challenging for two key reasons: (i) high-fidelity methods such as large eddy simulation (LES) are required to accurately capture the in-cylinder turbulent flow field, and (ii) CCV is experienced over long timescales and hence the simulations needmore » to be performed for hundreds of consecutive cycles. In the new technique, the strategy is to perform multiple parallel simulations, each of which encompasses 2-3 cycles, by effectively perturbing the simulation parameters such as the initial and boundary conditions. The PPM4CCV code is a pre-processing code and can be coupled with any engine CFD code. PPM4CCV was coupled with Converge CFD code and a 10-time speedup was demonstrated over the conventional multi-cycle LES in predicting the CCV for a motored engine. Recently, the model is also being applied to fired engines including port fuel injected (PFI) and direct injection spark ignition engines and the preliminary results are very encouraging.« less

  14. Kinetic Approaches to Shear-Driven Magnetic Reconnection for Multi-Scale Modeling of CME Initiation

    NASA Astrophysics Data System (ADS)

    Black, C.; Antiochos, S. K.; DeVore, C.; Germaschewski, K.; Karpen, J. T.

    2013-12-01

    In the standard model for coronal mass ejections (CME) and/or solar flares, the free energy for the event resides in the strongly sheared magnetic field of a filament channel. The pre-eruption force balance, consisting of an upward force due to the magnetic pressure of the sheared field balanced by a downward tension due to overlying un-sheared field, is widely believed to be disrupted by magnetic reconnection. Therefore, understanding initiation of solar explosive phenomena requires a true multi-scale model of reconnection onset driven by the buildup of magnetic shear. While the application of magnetic-field shear is a trivial matter in MHD simulations, it is a significant challenge in a PIC code. The driver must be implemented in a self-consistent manner and with boundary conditions that avoid the generation of waves that destroy the applied shear. In this work, we describe drivers for 2.5D, aperiodic, PIC systems and discuss the implementation of driver-consistent boundary conditions that allow a net electric current to flow through the walls. Preliminary tests of these boundaries with a MHD equilibrium are shown. This work was supported, in part, by the NASA Living With a Star TR&T Program.

  15. Parallel software support for computational structural mechanics

    NASA Technical Reports Server (NTRS)

    Jordan, Harry F.

    1987-01-01

    The application of the parallel programming methodology known as the Force was conducted. Two application issues were addressed. The first involves the efficiency of the implementation and its completeness in terms of satisfying the needs of other researchers implementing parallel algorithms. Support for, and interaction with, other Computational Structural Mechanics (CSM) researchers using the Force was the main issue, but some independent investigation of the Barrier construct, which is extremely important to overall performance, was also undertaken. Another efficiency issue which was addressed was that of relaxing the strong synchronization condition imposed on the self-scheduled parallel DO loop. The Force was extended by the addition of logical conditions to the cases of a parallel case construct and by the inclusion of a self-scheduled version of this construct. The second issue involved applying the Force to the parallelization of finite element codes such as those found in the NICE/SPAR testbed system. One of the more difficult problems encountered is the determination of what information in COMMON blocks is actually used outside of a subroutine and when a subroutine uses a COMMON block merely as scratch storage for internal temporary results.

  16. Parallel processing data network of master and slave transputers controlled by a serial control network

    DOEpatents

    Crosetto, D.B.

    1996-12-31

    The present device provides for a dynamically configurable communication network having a multi-processor parallel processing system having a serial communication network and a high speed parallel communication network. The serial communication network is used to disseminate commands from a master processor to a plurality of slave processors to effect communication protocol, to control transmission of high density data among nodes and to monitor each slave processor`s status. The high speed parallel processing network is used to effect the transmission of high density data among nodes in the parallel processing system. Each node comprises a transputer, a digital signal processor, a parallel transfer controller, and two three-port memory devices. A communication switch within each node connects it to a fast parallel hardware channel through which all high density data arrives or leaves the node. 6 figs.

  17. Parallel processing data network of master and slave transputers controlled by a serial control network

    DOEpatents

    Crosetto, Dario B.

    1996-01-01

    The present device provides for a dynamically configurable communication network having a multi-processor parallel processing system having a serial communication network and a high speed parallel communication network. The serial communication network is used to disseminate commands from a master processor (100) to a plurality of slave processors (200) to effect communication protocol, to control transmission of high density data among nodes and to monitor each slave processor's status. The high speed parallel processing network is used to effect the transmission of high density data among nodes in the parallel processing system. Each node comprises a transputer (104), a digital signal processor (114), a parallel transfer controller (106), and two three-port memory devices. A communication switch (108) within each node (100) connects it to a fast parallel hardware channel (70) through which all high density data arrives or leaves the node.

  18. Parallel Computation of the Jacobian Matrix for Nonlinear Equation Solvers Using MATLAB

    NASA Technical Reports Server (NTRS)

    Rose, Geoffrey K.; Nguyen, Duc T.; Newman, Brett A.

    2017-01-01

    Demonstrating speedup for parallel code on a multicore shared memory PC can be challenging in MATLAB due to underlying parallel operations that are often opaque to the user. This can limit potential for improvement of serial code even for the so-called embarrassingly parallel applications. One such application is the computation of the Jacobian matrix inherent to most nonlinear equation solvers. Computation of this matrix represents the primary bottleneck in nonlinear solver speed such that commercial finite element (FE) and multi-body-dynamic (MBD) codes attempt to minimize computations. A timing study using MATLAB's Parallel Computing Toolbox was performed for numerical computation of the Jacobian. Several approaches for implementing parallel code were investigated while only the single program multiple data (spmd) method using composite objects provided positive results. Parallel code speedup is demonstrated but the goal of linear speedup through the addition of processors was not achieved due to PC architecture.

  19. Matrix-Inversion-Free Compressed Sensing With Variable Orthogonal Multi-Matching Pursuit Based on Prior Information for ECG Signals.

    PubMed

    Cheng, Yih-Chun; Tsai, Pei-Yun; Huang, Ming-Hao

    2016-05-19

    Low-complexity compressed sensing (CS) techniques for monitoring electrocardiogram (ECG) signals in wireless body sensor network (WBSN) are presented. The prior probability of ECG sparsity in the wavelet domain is first exploited. Then, variable orthogonal multi-matching pursuit (vOMMP) algorithm that consists of two phases is proposed. In the first phase, orthogonal matching pursuit (OMP) algorithm is adopted to effectively augment the support set with reliable indices and in the second phase, the orthogonal multi-matching pursuit (OMMP) is employed to rescue the missing indices. The reconstruction performance is thus enhanced with the prior information and the vOMMP algorithm. Furthermore, the computation-intensive pseudo-inverse operation is simplified by the matrix-inversion-free (MIF) technique based on QR decomposition. The vOMMP-MIF CS decoder is then implemented in 90 nm CMOS technology. The QR decomposition is accomplished by two systolic arrays working in parallel. The implementation supports three settings for obtaining 40, 44, and 48 coefficients in the sparse vector. From the measurement result, the power consumption is 11.7 mW at 0.9 V and 12 MHz. Compared to prior chip implementations, our design shows good hardware efficiency and is suitable for low-energy applications.

  20. A novel measure of compulsive food restriction in anorexia nervosa: validation of the Self-Starvation Scale (SS).

    PubMed

    Godier, Lauren R; Park, Rebecca J

    2015-04-01

    The characteristic relentless self-starvation behaviour seen in Anorexia Nervosa (AN) has been described as evidence of compulsivity, with increasing suggestion of transdiagnostic parallels with addictive behaviour. There is a paucity of standardised self-report measures of compulsive behaviour in eating disorders (EDs). Measures that index the concept of compulsive self-starvation in AN are needed to explore the suggested parallels with addictions. With this aim a novel measure of self-starvation was developed (the Self-Starvation Scale, SS). 126 healthy participants, and 78 individuals with experience of AN, completed the new measure along with existing measures of eating disorder symptoms, anxiety and depression. Initial validation in the healthy sample indicated good reliability and construct validity, and incremental validity in predicting eating disorder symptoms. The psychometric properties of the SS scale were replicated in the AN sample. The ability of this scale to predict ED symptoms was particularly strong in individuals currently suffering from AN. These results suggest the SS may be a useful index of compulsive food restriction in AN. The concept of 'starvation dependence' in those with eating disorders, as a parallel with addiction, may be of clinical and theoretical importance. Copyright © 2014 Elsevier Ltd. All rights reserved.

  1. The potential of multi-port optical memories in digital computing

    NASA Technical Reports Server (NTRS)

    Alford, C. O.; Gaylord, T. K.

    1975-01-01

    A high-capacity memory with a relatively high data transfer rate and multi-port simultaneous access capability may serve as the basis for new computer architectures. The implementation of a multi-port optical memory is discussed. Several computer structures are presented that might profitably use such a memory. These structures include (1) a simultaneous record access system, (2) a simultaneously shared memory computer system, and (3) a parallel digital processing structure.

  2. Nebo: An efficient, parallel, and portable domain-specific language for numerically solving partial differential equations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Earl, Christopher; Might, Matthew; Bagusetty, Abhishek

    This study presents Nebo, a declarative domain-specific language embedded in C++ for discretizing partial differential equations for transport phenomena on multiple architectures. Application programmers use Nebo to write code that appears sequential but can be run in parallel, without editing the code. Currently Nebo supports single-thread execution, multi-thread execution, and many-core (GPU-based) execution. With single-thread execution, Nebo performs on par with code written by domain experts. With multi-thread execution, Nebo can linearly scale (with roughly 90% efficiency) up to 12 cores, compared to its single-thread execution. Moreover, Nebo’s many-core execution can be over 140x faster than its single-thread execution.

  3. Nebo: An efficient, parallel, and portable domain-specific language for numerically solving partial differential equations

    DOE PAGES

    Earl, Christopher; Might, Matthew; Bagusetty, Abhishek; ...

    2016-01-26

    This study presents Nebo, a declarative domain-specific language embedded in C++ for discretizing partial differential equations for transport phenomena on multiple architectures. Application programmers use Nebo to write code that appears sequential but can be run in parallel, without editing the code. Currently Nebo supports single-thread execution, multi-thread execution, and many-core (GPU-based) execution. With single-thread execution, Nebo performs on par with code written by domain experts. With multi-thread execution, Nebo can linearly scale (with roughly 90% efficiency) up to 12 cores, compared to its single-thread execution. Moreover, Nebo’s many-core execution can be over 140x faster than its single-thread execution.

  4. Real time display Fourier-domain OCT using multi-thread parallel computing with data vectorization

    NASA Astrophysics Data System (ADS)

    Eom, Tae Joong; Kim, Hoon Seop; Kim, Chul Min; Lee, Yeung Lak; Choi, Eun-Seo

    2011-03-01

    We demonstrate a real-time display of processed OCT images using multi-thread parallel computing with a quad-core CPU of a personal computer. The data of each A-line are treated as one vector to maximize the data translation rate between the cores of the CPU and RAM stored image data. A display rate of 29.9 frames/sec for processed OCT data (4096 FFT-size x 500 A-scans) is achieved in our system using a wavelength swept source with 52-kHz swept frequency. The data processing times of the OCT image and a Doppler OCT image with a 4-time average are 23.8 msec and 91.4 msec.

  5. Implementing Shared Memory Parallelism in MCBEND

    NASA Astrophysics Data System (ADS)

    Bird, Adam; Long, David; Dobson, Geoff

    2017-09-01

    MCBEND is a general purpose radiation transport Monte Carlo code from AMEC Foster Wheelers's ANSWERS® Software Service. MCBEND is well established in the UK shielding community for radiation shielding and dosimetry assessments. The existing MCBEND parallel capability effectively involves running the same calculation on many processors. This works very well except when the memory requirements of a model restrict the number of instances of a calculation that will fit on a machine. To more effectively utilise parallel hardware OpenMP has been used to implement shared memory parallelism in MCBEND. This paper describes the reasoning behind the choice of OpenMP, notes some of the challenges of multi-threading an established code such as MCBEND and assesses the performance of the parallel method implemented in MCBEND.

  6. Parallel workflow tools to facilitate human brain MRI post-processing

    PubMed Central

    Cui, Zaixu; Zhao, Chenxi; Gong, Gaolang

    2015-01-01

    Multi-modal magnetic resonance imaging (MRI) techniques are widely applied in human brain studies. To obtain specific brain measures of interest from MRI datasets, a number of complex image post-processing steps are typically required. Parallel workflow tools have recently been developed, concatenating individual processing steps and enabling fully automated processing of raw MRI data to obtain the final results. These workflow tools are also designed to make optimal use of available computational resources and to support the parallel processing of different subjects or of independent processing steps for a single subject. Automated, parallel MRI post-processing tools can greatly facilitate relevant brain investigations and are being increasingly applied. In this review, we briefly summarize these parallel workflow tools and discuss relevant issues. PMID:26029043

  7. Carpet: Adaptive Mesh Refinement for the Cactus Framework

    NASA Astrophysics Data System (ADS)

    Schnetter, Erik; Hawley, Scott; Hawke, Ian

    2016-11-01

    Carpet is an adaptive mesh refinement and multi-patch driver for the Cactus Framework (ascl:1102.013). Cactus is a software framework for solving time-dependent partial differential equations on block-structured grids, and Carpet acts as driver layer providing adaptive mesh refinement, multi-patch capability, as well as parallelization and efficient I/O.

  8. A Scalable, Parallel Approach for Multi-Point, High-Fidelity Aerostructural Optimization of Aircraft Configurations

    NASA Astrophysics Data System (ADS)

    Kenway, Gaetan K. W.

    This thesis presents new tools and techniques developed to address the challenging problem of high-fidelity aerostructural optimization with respect to large numbers of design variables. A new mesh-movement scheme is developed that is both computationally efficient and sufficiently robust to accommodate large geometric design changes and aerostructural deformations. A fully coupled Newton-Krylov method is presented that accelerates the convergence of aerostructural systems and provides a 20% performance improvement over the traditional nonlinear block Gauss-Seidel approach and can handle more exible structures. A coupled adjoint method is used that efficiently computes derivatives for a gradient-based optimization algorithm. The implementation uses only machine accurate derivative techniques and is verified to yield fully consistent derivatives by comparing against the complex step method. The fully-coupled large-scale coupled adjoint solution method is shown to have 30% better performance than the segregated approach. The parallel scalability of the coupled adjoint technique is demonstrated on an Euler Computational Fluid Dynamics (CFD) model with more than 80 million state variables coupled to a detailed structural finite-element model of the wing with more than 1 million degrees of freedom. Multi-point high-fidelity aerostructural optimizations of a long-range wide-body, transonic transport aircraft configuration are performed using the developed techniques. The aerostructural analysis employs Euler CFD with a 2 million cell mesh and a structural finite element model with 300 000 DOF. Two design optimization problems are solved: one where takeoff gross weight is minimized, and another where fuel burn is minimized. Each optimization uses a multi-point formulation with 5 cruise conditions and 2 maneuver conditions. The optimization problems have 476 design variables are optimal results are obtained within 36 hours of wall time using 435 processors. The TOGW minimization results in a 4.2% reduction in TOGW with a 6.6% fuel burn reduction, while the fuel burn optimization resulted in a 11.2% fuel burn reduction with no change to the takeoff gross weight.

  9. Co-evolution of upstream waves and accelerated ions at parallel shocks

    NASA Astrophysics Data System (ADS)

    Fujimoto, M.; Sugiyama, T.

    2016-12-01

    Shock waves in space plasmas have been considered as the agents for various particle acceleration phenomena. The basic idea behind shock acceleration is that particles are accelerated as they move back-and-forth across a shock front. Detailed studies of ion acceleration at the terrestrial bow shock have been performed, however, the restricted maximum energies attained prevent a straight-forward application of obtained knowledge to more energetic astrophysical situations. Here we show by a large-scale self-consistent particle simulation that the co-evolution of magnetic turbulence and accelerated ion population is the foundation for continuous operation of shock acceleration to ever higher energies. Magnetic turbulence is created by ions reflected back upstream of a parallel shock front. The co-evolution arises because more energetic ions excite waves of longer wavelengths, and because longer wavelength modes are capable of scattering (in the upstream) and reflecting (at the shock front) more energetic ions. Via carefully designed numerical experiments, we show very clearly that this picture is true.

  10. Fano effect dominance over Coulomb blockade in transport properties of parallel coupled quantum dot system

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brogi, Bharat Bhushan, E-mail: brogi-221179@yahoo.in; Ahluwalia, P. K.; Chand, Shyam

    2015-06-24

    Theoretical study of the Coulomb blockade effect on transport properties (Transmission Probability and I-V characteristics) for varied configuration of coupled quantum dot system has been studied by using Non Equilibrium Green Function(NEGF) formalism and Equation of Motion(EOM) method in the presence of magnetic flux. The self consistent approach and intra-dot Coulomb interaction is being taken into account. As the key parameters of the coupled quantum dot system such as dot-lead coupling, inter-dot tunneling and magnetic flux threading through the system can be tuned, the effect of asymmetry parameter and magnetic flux on this tuning is being explored in Coulomb blockademore » regime. The presence of the Coulomb blockade due to on-dot Coulomb interaction decreases the width of transmission peak at energy level ε + U and by adjusting the magnetic flux the swapping effect in the Fano peaks in asymmetric and symmetric parallel configuration sustains despite strong Coulomb blockade effect.« less

  11. Acceleration of Particles Near Earth's Bow Shock

    NASA Astrophysics Data System (ADS)

    Sandroos, A.

    2012-12-01

    Collisionless shock waves, for example, near planetary bodies or driven by coronal mass ejections, are a key source of energetic particles in the heliosphere. When the solar wind hits Earth's bow shock, some of the incident particles get reflected back towards the Sun and are accelerated in the process. Reflected ions are responsible for the creation of a turbulent foreshock in quasi-parallel regions of Earth's bow shock. We present first results of foreshock macroscopic structure and of particle distributions upstream of Earth's bow shock, obtained with a new 2.5-dimensional self-consistent diffusive shock acceleration model. In the model particles' pitch angle scattering rates are calculated from Alfvén wave power spectra using quasilinear theory. Wave power spectra in turn are modified by particles' energy changes due to the scatterings. The new model has been implemented on massively parallel simulation platform Corsair. We have used an earlier version of the model to study ion acceleration in a shock-shock interaction event (Hietala, Sandroos, and Vainio, 2012).

  12. Tempest Neoclassical Simulation of Fusion Edge Plasmas

    NASA Astrophysics Data System (ADS)

    Xu, X. Q.; Xiong, Z.; Cohen, B. I.; Cohen, R. H.; Dorr, M.; Hittinger, J.; Kerbel, G. D.; Nevins, W. M.; Rognlien, T. D.

    2006-04-01

    We are developing a continuum gyrokinetic full-F code, TEMPEST, to simulate edge plasmas. The geometry is that of a fully diverted tokamak and so includes boundary conditions for both closed magnetic flux surfaces and open field lines. The code, presently 4-dimensional (2D2V), includes kinetic ions and electrons, a gyrokinetic Poisson solver for electric field, and the nonlinear Fokker-Planck collision operator. Here we present the simulation results of neoclassical transport with Boltzmann electrons. In a large aspect ratio circular geometry, excellent agreement is found for neoclassical equilibrium with parallel flows in the banana regime without a temperature gradient. In divertor geometry, it is found that the endloss of particles and energy induces pedestal-like density and temperature profiles inside the magnetic separatrix and parallel flow stronger than the neoclassical predictions in the SOL. The impact of the X-point divertor geometry on the self-consistent electric field and geo-acoustic oscillations will be reported. We will also discuss the status of extending TEMPEST into a 5-D code.

  13. Anomalous transport theory for the reversed field pinch

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Terry, P.W.; Hegna, C.C; Sovinec, C.R.

    1996-09-01

    Physically motivated transport models with predictive capabilities and significance beyond the reversed field pinch (RFP) are presented. It is shown that the ambipolar constrained electron heat loss observed in MST can be quantitatively modeled by taking account of the clumping in parallel streaming electrons and the resultant self-consistent interaction with collective modes; that the discrete dynamo process is a relaxation oscillation whose dependence on the tearing instability and profile relaxation physics leads to amplitude and period scaling predictions consistent with experiment; that the Lundquist number scaling in relaxed plasmas driven by magnetic turbulence has a weak S{sup {minus}1/4} scaling; andmore » that radial E{times}B shear flow can lead to large reductions in the edge particle flux with little change in the heat flux, as observed in the RFP and tokamak. 24 refs.« less

  14. The Acceleration of Thermal Protons and Minor Ions at a Quasi-Parallel Interplanetary Shock

    NASA Astrophysics Data System (ADS)

    Giacalone, J.; Lario, D.; Lepri, S. T.

    2017-12-01

    We compare the results from self-consistent hybrid simulations (kinetic ions, massless fluid electrons) and spacecraft observations of a strong, quasi-parallel interplanetary shock that crossed the Advanced Composition Explorer (ACE) on DOY 94, 2001. In our simulations, the un-shocked plasma-frame ion distributions are Maxwellian. Our simulations include protons and minor ions (alphas, 3He++, and C5+). The interplanetary shock crossed both the ACE and the Wind spacecraft, and was associated with significant increases in the flux of > 50 keV/nuc ions. Our simulation uses parameters (ion densities, magnetic field strength, Mach number, etc.) consistent with those observed. Acceleration of the ions by the shock, in a manner similar to that expected from diffusive shock acceleration theory, leads to a high-energy tail in the distribution of the post-shock plasma for all ions we considered. The simulated distributions are directly compared to those observed by ACE/SWICS, EPAM, and ULEIS, and Wind/STICS and 3DP, covering the energy range from below the thermal peak to the suprathermal tail. We conclude from our study that the solar wind is the most significant source of the high-energy ions for this event. Our results have important implications for the physics of the so-called `injection problem', which will be discussed.

  15. The Evaluation of Relationship between Sexual Self-concept and Sexual Dysfunction in Individuals Undergoing Methadone Maintenance Treatment

    PubMed Central

    Rajabizadeh, Ghodratolah; Yazdanpanah, Fatemeh; Ramezani, Mohammad Arash

    2017-01-01

    Background The present study was conducted with the aim of designing a causal model for the evaluation of sexual dysfunctions based on the variables of methadone dosage and sexual self-concept among individuals undergoing methadone maintenance treatment (MMT). Methods The study population of the present study consisted of married men of 20 to 45 years of age with sexual ýrelations and undergoing MMT for a minimum of 8 weeks referring to all MMT clinics of Kerman, Iran, in ýý2015-2016. ýThe subjects were selected through multi-stage cluster sampling (n = 250). Data were collected using the General Health Questionnaire (GHQ-28), Multi-Dimensional Sexual Self-concept Questionnaire (MSSQ), and Internal Index for Erectile Function (IIEF). Data were analyzed using path analysis method and Pearson correlation coefficient. The suggested model was evaluated using structural equation model (SEM), and indirect relationships were assessed using Bootstrap method. Findings The suggested model showed acceptable fitness with the data, and all routes, except methadone use route, to sexual function were significant. The result of the multiple ýindirect route showed that sexual function had a significant relationship with methadone use through ýsexual self-concept. In total, 60% of variance in sexual dysfunction was explained using the variables of the suggested model. Conclusion Further studies are suggested to be conducted regarding psychological factors effective on the sexual dysfunctions among individuals undergoing MMT, such as sexual self-concept. Moreover, more detailed evaluation of each subscale of positive and negative sexual self-concept is recommended to assess the psychological causes of sexual dysfunctions in these individuals and design psychological, behavioral, and cognitive-behavioral treatment interventions for them. PMID:29299211

  16. The Evaluation of Relationship between Sexual Self-concept and Sexual Dysfunction in Individuals Undergoing Methadone Maintenance Treatment.

    PubMed

    Rajabizadeh, Ghodratolah; Yazdanpanah, Fatemeh; Ramezani, Mohammad Arash

    2017-04-01

    The present study was conducted with the aim of designing a causal model for the evaluation of sexual dysfunctions based on the variables of methadone dosage and sexual self-concept among individuals undergoing methadone maintenance treatment (MMT). The study population of the present study consisted of married men of 20 to 45 years of age with sexual ýrelations and undergoing MMT for a minimum of 8 weeks referring to all MMT clinics of Kerman, Iran, in ýý2015-2016. ýThe subjects were selected through multi-stage cluster sampling (n = 250). Data were collected using the General Health Questionnaire (GHQ-28), Multi-Dimensional Sexual Self-concept Questionnaire (MSSQ), and Internal Index for Erectile Function (IIEF). Data were analyzed using path analysis method and Pearson correlation coefficient. The suggested model was evaluated using structural equation model (SEM), and indirect relationships were assessed using Bootstrap method. The suggested model showed acceptable fitness with the data, and all routes, except methadone use route, to sexual function were significant. The result of the multiple ýindirect route showed that sexual function had a significant relationship with methadone use through ýsexual self-concept. In total, 60% of variance in sexual dysfunction was explained using the variables of the suggested model. Further studies are suggested to be conducted regarding psychological factors effective on the sexual dysfunctions among individuals undergoing MMT, such as sexual self-concept. Moreover, more detailed evaluation of each subscale of positive and negative sexual self-concept is recommended to assess the psychological causes of sexual dysfunctions in these individuals and design psychological, behavioral, and cognitive-behavioral treatment interventions for them.

  17. Reflection from a multi-species material and its transmitted effective wavenumber

    NASA Astrophysics Data System (ADS)

    Gower, Artur L.; Smith, Michael J. A.; Parnell, William J.; Abrahams, I. David

    2018-04-01

    We formally deduce closed-form expressions for the transmitted effective wavenumber of a material comprising multiple types of inclusions or particles (multi-species), dispersed in a uniform background medium. The expressions, derived here for the first time, are valid for moderate volume fractions and without restriction on the frequency. We show that the multi-species effective wavenumber is not a straightforward extension of expressions for a single species. Comparisons are drawn with state-of-the-art models in acoustics by presenting numerical results for a concrete and a water-oil emulsion in two dimensions. The limit of when one species is much smaller than the other is also discussed and we determine the background medium felt by the larger species in this limit. Surprisingly, we show that the answer is not the intuitive result predicted by self-consistent multiple scattering theories. The derivation presented here applies to the scalar wave equation with cylindrical or spherical inclusions, with any distribution of sizes, densities and wave speeds. The reflection coefficient associated with a halfspace of multi-species cylindrical inclusions is also formally derived.

  18. Performance of quantum Monte Carlo for calculating molecular bond lengths

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cleland, Deidre M., E-mail: deidre.cleland@csiro.au; Per, Manolo C., E-mail: manolo.per@csiro.au

    2016-03-28

    This work investigates the accuracy of real-space quantum Monte Carlo (QMC) methods for calculating molecular geometries. We present the equilibrium bond lengths of a test set of 30 diatomic molecules calculated using variational Monte Carlo (VMC) and diffusion Monte Carlo (DMC) methods. The effect of different trial wavefunctions is investigated using single determinants constructed from Hartree-Fock (HF) and Density Functional Theory (DFT) orbitals with LDA, PBE, and B3LYP functionals, as well as small multi-configurational self-consistent field (MCSCF) multi-determinant expansions. When compared to experimental geometries, all DMC methods exhibit smaller mean-absolute deviations (MADs) than those given by HF, DFT, and MCSCF.more » The most accurate MAD of 3 ± 2 × 10{sup −3} Å is achieved using DMC with a small multi-determinant expansion. However, the more computationally efficient multi-determinant VMC method has a similar MAD of only 4.0 ± 0.9 × 10{sup −3} Å, suggesting that QMC forces calculated from the relatively simple VMC algorithm may often be sufficient for accurate molecular geometries.« less

  19. Reflection from a multi-species material and its transmitted effective wavenumber

    PubMed Central

    Parnell, William J.; Abrahams, I. David

    2018-01-01

    We formally deduce closed-form expressions for the transmitted effective wavenumber of a material comprising multiple types of inclusions or particles (multi-species), dispersed in a uniform background medium. The expressions, derived here for the first time, are valid for moderate volume fractions and without restriction on the frequency. We show that the multi-species effective wavenumber is not a straightforward extension of expressions for a single species. Comparisons are drawn with state-of-the-art models in acoustics by presenting numerical results for a concrete and a water–oil emulsion in two dimensions. The limit of when one species is much smaller than the other is also discussed and we determine the background medium felt by the larger species in this limit. Surprisingly, we show that the answer is not the intuitive result predicted by self-consistent multiple scattering theories. The derivation presented here applies to the scalar wave equation with cylindrical or spherical inclusions, with any distribution of sizes, densities and wave speeds. The reflection coefficient associated with a halfspace of multi-species cylindrical inclusions is also formally derived. PMID:29740262

  20. A structural model for apolipoprotein C-II amyloid fibrils: experimental characterization and molecular dynamics simulations.

    PubMed

    Teoh, Chai Lean; Pham, Chi L L; Todorova, Nevena; Hung, Andrew; Lincoln, Craig N; Lees, Emma; Lam, Yuen Han; Binger, Katrina J; Thomson, Neil H; Radford, Sheena E; Smith, Trevor A; Müller, Shirley A; Engel, Andreas; Griffin, Michael D W; Yarovsky, Irene; Gooley, Paul R; Howlett, Geoffrey J

    2011-02-04

    The self-assembly of specific proteins to form insoluble amyloid fibrils is a characteristic feature of a number of age-related and debilitating diseases. Lipid-free human apolipoprotein C-II (apoC-II) forms characteristic amyloid fibrils and is one of several apolipoproteins that accumulate in amyloid deposits located within atherosclerotic plaques. X-ray diffraction analysis of aligned apoC-II fibrils indicated a simple cross-β-structure composed of two parallel β-sheets. Examination of apoC-II fibrils using transmission electron microscopy, scanning transmission electron microscopy, and atomic force microscopy indicated that the fibrils are flat ribbons composed of one apoC-II molecule per 4.7-Å rise of the cross-β-structure. Cross-linking results using single-cysteine substitution mutants are consistent with a parallel in-register structural model for apoC-II fibrils. Fluorescence resonance energy transfer analysis of apoC-II fibrils labeled with specific fluorophores provided distance constraints for selected donor-acceptor pairs located within the fibrils. These findings were used to develop a simple 'letter-G-like' β-strand-loop-β-strand model for apoC-II fibrils. Fully solvated all-atom molecular dynamics (MD) simulations showed that the model contained a stable cross-β-core with a flexible connecting loop devoid of persistent secondary structure. The time course of the MD simulations revealed that charge clusters in the fibril rearrange to minimize the effects of same-charge interactions inherent in parallel in-register models. Our structural model for apoC-II fibrils suggests that apoC-II monomers fold and self-assemble to form a stable cross-β-scaffold containing relatively unstructured connecting loops. Copyright © 2010 Elsevier Ltd. All rights reserved.

  1. Adherence predictors in an Internet-based Intervention program for depression.

    PubMed

    Castro, Adoración; López-Del-Hoyo, Yolanda; Peake, Christian; Mayoral, Fermín; Botella, Cristina; García-Campayo, Javier; Baños, Rosa María; Nogueira-Arjona, Raquel; Roca, Miquel; Gili, Margalida

    2018-05-01

    Internet-delivered psychotherapy has been demonstrated to be effective in the treatment of depression. Nevertheless, the study of the adherence in this type of the treatment reported divergent results. The main objective of this study is to analyze predictors of adherence in a primary care Internet-based intervention for depression in Spain. A multi-center, three arm, parallel, randomized controlled trial was conducted with 194 depressive patients, who were allocated in self-guided or supported-guided intervention. Sociodemographic and clinical characteristics were gathered using a case report form. The Mini international neuropsychiatric interview diagnoses major depression. Beck Depression Inventory was used to assess depression severity. The visual analogic scale assesses the respondent's self-rated health and Short Form Health Survey was used to measure the health-related quality of life. Age results a predictor variable for both intervention groups (with and without therapist support). Perceived health is a negative predictor of adherence for the self-guided intervention when change in depression severity was included in the model. Change in depression severity results a predictor of adherence in the support-guided intervention. Our findings demonstrate that in our sample, there are differences in sociodemographic and clinical variables between active and dropout participants and we provide adherence predictors in each intervention condition of this Internet-based program for depression (self-guided and support-guided). It is important to point that further research in this area is essential to improve tailored interventions and to know specific patients groups can benefit from these interventions.

  2. The validity of self-reported leisure time physical activity, and its relationship to serum cholesterol, blood pressure and body mass index. A population based study of 332,182 men and women aged 40-42 years.

    PubMed

    Aires, Nibia; Selmer, Randi; Thelle, Dag

    2003-01-01

    The importance of leisure time physical activity as a health indicator became more obvious after the results of large prospective studies were published. The validity of these results depends upon both the selection of the active individuals and to what extent self-reported physical activity reflects the individual's true activity. The purpose of this paper is to describe the changes in self-reported physical activity, and to assess the relation between this variable and other biological risk factors such as blood lipids, blood pressure and body mass index (BMI). This report also aims at corroborating the validity of self-reported physical activity by assessing the consistency of the associations between these biological risk factors and physical activity during a 25-years period. The basis for this analysis is a long lasting observational study with a questionnaire as the most important research instrument, in addition to physiological and biological factors such as BMI, blood pressure and blood lipids. The study population consists of 332,182 individuals, aged 40-42 from different counties in Norway who were invited to participate in health survey during 1974-1999. The objectives of this study are (1) to describe changes in self-reported physical activity from 1974 to 1999; (2) to assess the relation between physical activity and the biological variables; and (3) to corroborate the validity of the variable physical activity by assessing the consistency of the above analysis. The results of the analyses of association between decade of birth and self-reported physical activity show that physical activity among 40-aged individuals decreased during 1974-1999. This trend is stronger among the men. Multivariate analyses revealed differences in BMI and serum cholesterol between levels of self-reported physical activity, gender, smoking habits and decade of birth. The explained percentage of the total variance ranged from 6% for BMI to 7% for serum cholesterol. The similar shape of serum cholesterol and BMI according to physical activity indicates that the validity of self-reported physical activity has remained stable over these 25 years. Furthermore, the analysis of covariance showed that the slopes relating year of birth and serum cholesterol and BMI are parallel for self-reported physical activity thus the validity of the variable is confirmed.

  3. The Diabetes Self-Management Questionnaire (DSMQ): development and evaluation of an instrument to assess diabetes self-care activities associated with glycaemic control.

    PubMed

    Schmitt, Andreas; Gahr, Annika; Hermanns, Norbert; Kulzer, Bernhard; Huber, Jörg; Haak, Thomas

    2013-08-13

    Though several questionnaires on self-care and regimen adherence have been introduced, the evaluations do not always report consistent and substantial correlations with measures of glycaemic control. Small ability to explain variance in HbA1c constitutes a significant limitation of an instrument's use for scientific purposes as well as clinical practice. In order to assess self-care activities which can predict glycaemic control, the Diabetes Self-Management Questionnaire (DSMQ) was designed. A 16 item questionnaire to assess self-care activities associated with glycaemic control was developed, based on theoretical considerations and a process of empirical improvements. Four subscales, 'Glucose Management' (GM), 'Dietary Control' (DC), 'Physical Activity' (PA), and 'Health-Care Use' (HU), as well as a 'Sum Scale' (SS) as a global measure of self-care were derived. To evaluate its psychometric quality, 261 patients with type 1 or 2 diabetes were assessed with the DSMQ and an established analogous scale, the Summary of Diabetes Self-Care Activities Measure (SDSCA). The DSMQ's item and scale characteristics as well as factorial and convergent validity were analysed, and its convergence with HbA1c was compared to the SDSCA. The items showed appropriate characteristics (mean item-total-correlation: 0.46 ± 0.12; mean correlation with HbA1c: -0.23 ± 0.09). Overall internal consistency (Cronbach's alpha) was good (0.84), consistencies of the subscales were acceptable (GM: 0.77; DC: 0.77; PA: 0.76; HU: 0.60). Principal component analysis indicated a four factor structure and confirmed the designed scale structure. Confirmatory factor analysis indicated appropriate fit of the four factor model. The DSMQ scales showed significant convergent correlations with their parallel SDSCA scales (GM: 0.57; DC: 0.52; PA: 0.58; HU: n/a; SS: 0.57) and HbA1c (GM: -0.39; DC: -0.30; PA: -0.15; HU: -0.22; SS: -0.40). All correlations with HbA1c were significantly stronger than those obtained with the SDSCA. This study provides preliminary evidence that the DSMQ is a reliable and valid instrument and enables an efficient assessment of self-care behaviours associated with glycaemic control. The questionnaire should be valuable for scientific analyses as well as clinical use in both type 1 and type 2 diabetes patients.

  4. Acute psychological benefits of aerobic exercise: a field study into the effects of exercise characteristics.

    PubMed

    Rendi, Mária; Szabo, Attila; Szabó, Tamás; Velenczei, Attila; Kovács, Arpád

    2008-03-01

    Eighty volunteers were tested in their natural exercise environment consisting of a fitness centre they regularly attended. Half of the sample exercised on a stationary bicycle, the other half on a treadmill. All participants filled in the Exercise-Induced Feeling Inventory before and after their 20 min of exercise that was performed at self-selected workload. The results revealed that exercise intensity and the other parallel measures like heart rate, perceived exercise intensity and estimates of burned calories were higher in participants who ran in contrast to those who cycled. There were no differences in self-reports of enjoyment of the exercise sessions and in the psychological improvements from pre- to post-exercise between the groups. It is concluded that significant psychological improvements occur even after a 20-min bout of exercise and these changes are independent of the workload or exercise intensity.

  5. System and method for embedding emotion in logic systems

    NASA Technical Reports Server (NTRS)

    Curtis, Steven A. (Inventor)

    2012-01-01

    A system, method, and computer readable-media for creating a stable synthetic neural system. The method includes training an intellectual choice-driven synthetic neural system (SNS), training an emotional rule-driven SNS by generating emotions from rules, incorporating the rule-driven SNS into the choice-driven SNS through an evolvable interface, and balancing the emotional SNS and the intellectual SNS to achieve stability in a nontrivial autonomous environment with a Stability Algorithm for Neural Entities (SANE). Generating emotions from rules can include coding the rules into the rule-driven SNS in a self-consistent way. Training the emotional rule-driven SNS can occur during a training stage in parallel with training the choice-driven SNS. The training stage can include a self assessment loop which measures performance characteristics of the rule-driven SNS against core genetic code. The method uses a stability threshold to measure stability of the incorporated rule-driven SNS and choice-driven SNS using SANE.

  6. The equilibrium and stability of the gaseous component of the galaxy, 2

    NASA Technical Reports Server (NTRS)

    Kellman, S. A.

    1971-01-01

    A time-independent, linear, plane and axially-symmetric stability analysis was performed on a self-gravitating, plane-parallel, isothermal layer of nonmagnetic, nonrotating gas. The gas layer was immersed in a plane-stratified field isothermal layer of stars which supply a self-consistent gravitational field. Only the gaseous component was perturbed. Expressions were derived for the perturbed gas potential and perturbed gas density that satisfied both the Poisson and hydrostatic equilibrium equations. The equation governing the size of the perturbations in the mid-plane was found to be analogous to the one-dimensional time-independent Schrodinger equation for a particle bound by a potential well, and with similar boundary conditions. The radius of the neutral state was computed numerically and compared with the Jeans' and Ledoux radius. The inclusion of a rigid stellar component increased the Ledoux radius, though only slightly. Isodensity contours of the neutrual or marginally unstable state were constructed.

  7. FleCSPH - a parallel and distributed SPH implementation based on the FleCSI framework

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Junghans, Christoph; Loiseau, Julien

    2017-06-20

    FleCSPH is a multi-physics compact application that exercises FleCSI parallel data structures for tree-based particle methods. In particular, FleCSPH implements a smoothed-particle hydrodynamics (SPH) solver for the solution of Lagrangian problems in astrophysics and cosmology. FleCSPH includes support for gravitational forces using the fast multipole method (FMM).

  8. Does cost-benefit analysis or self-control predict involvement in two forms of aggression?

    PubMed

    Archer, John; Fernández-Fuertes, Andrés A; Thanzami, Van Lal

    2010-01-01

    The main aim of this research was to assess the relative association between physical aggression and (1) self-control and (2) cost-benefit assessment, these variables representing the operation of impulsive and reflective processes. Study 1 involved direct and indirect aggression among young Indian men, and Study 2 physical aggression to dating partners among Spanish adolescents. In Study 1, perceived benefits and costs but not self-control were associated with direct aggression at other men, and the association remained when their close association with indirect aggression was controlled. In Study 2, benefits and self-control showed significant and independent associations (positive for benefits, negative for self-control) with physical aggression at other-sex partners. Although being victimized was also correlated in the same direction with self-control and benefits, perpetration and being victimized were highly correlated, and there was no association between being victimized and these variables when perpetration was controlled. These results support the theory that reflective (cost-benefit analyses) processes and impulsive (self-control) processes operate in parallel in affecting aggression. The finding that male adolescents perceived more costs and fewer benefits from physical aggression to a partner than female adolescents did is consistent with findings indicating greater social disapproval of men hitting women than vice versa, rather than with the view that male violence to women is facilitated by internalized patriarchal values. (c) 2010 Wiley-Liss, Inc.

  9. Multi-hierarchical self-assembly of a collagen mimetic peptide from triple helix to nanofibre and hydrogel

    USDA-ARS?s Scientific Manuscript database

    Replicating the multi-hierarchical self-assembly of collagen has long-attracted scientists, from both the perspective of the fundamental science of supramolecular chemistry and that of potential biomedical applications in tissue engineering. Many approaches to drive the self-assembly of synthetic s...

  10. Body mass index and victimization during adolescence: the mediation role of depressive symptoms and self-esteem.

    PubMed

    Giletta, Matteo; Scholte, Ron H J; Engels, Rutger C M E; Larsen, Junilla K

    2010-12-01

    This study applied a multi-method approach to examine the relationship between body mass index (BMI) and the experience of victimization during adolescence by investigating the role of intrapersonal feelings. The sample consisted of 2051 adolescents (M=13.8 years, S.D.=0.7; 51% male) from seven high schools in the Netherlands. Participants' weight and height were measured and they completed self-report questionnaires on victimization, depressive symptoms and self-esteem. Self-reported and peer-reported measures of victimization were collected and combined to create three different victimization types (i.e., self/peer-identified, self-identified, and peer-identified). Hierarchical logistic regression analyses revealed that higher BMI was associated with both self/peer-identified victimization and self-identified victimization. Intrapersonal feelings (i.e., depressive symptoms and self-esteem) were found to mediate these associations. However, BMI was not associated with peer-identified victimization. These findings suggest that the association between BMI and victimization might be exclusively related to the self-perception of high BMI adolescents. Moreover, the mediation effects indicate that the perception of victimization might be linked to psychological difficulties of adolescents with high BMI. Thus, to fully understand the associations between weight status and victimization, intrapersonal mechanisms need to be examined. Copyright © 2010 Elsevier Inc. All rights reserved.

  11. Design and Development of a Run-Time Monitor for Multi-Core Architectures in Cloud Computing

    PubMed Central

    Kang, Mikyung; Kang, Dong-In; Crago, Stephen P.; Park, Gyung-Leen; Lee, Junghoon

    2011-01-01

    Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM) which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data. PMID:22163811

  12. Design and development of a run-time monitor for multi-core architectures in cloud computing.

    PubMed

    Kang, Mikyung; Kang, Dong-In; Crago, Stephen P; Park, Gyung-Leen; Lee, Junghoon

    2011-01-01

    Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM) which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data.

  13. Programmable diffractive optic for multi-beam processing: applications and limitations

    NASA Astrophysics Data System (ADS)

    Gretzki, Patrick; Gillner, Arnold

    2017-08-01

    In the field of laser ablation, especially in the field of micro-structuring, the current challenge is the improvement of productivity. While many applications, e.g. surface fictionalization and structuring, drilling and thin film ablation, use relatively low pulse energies, industrial laser sources provide considerably higher average powers and pulse energies. The main challenge consist of the effective energy distribution and depositions. There are essential two complementary approaches for the up-scaling of (ultra) short pulse laser processes: Higher repetition frequency or higher pulse energies. Using lasers with high repetition rates in the MHz region can cause thermal issues like overheating, melt production and low ablation quality. In this paper we pursuit the second approach by using diffractive optics for parallel processing. We will discuss, which technologies can be used and which applications will benefit from the multi-beam approach and which increase in productivity can be expected. Additionally we will show, which quality attributes can be used to rate the performance of a diffractive optic and and which limitations and restrictions this technology has.

  14. Radiation from mixed multi-planar wire arrays

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Safronova, A. S.; Kantsyrev, V. L.; Esaulov, A. A.

    2014-03-15

    The study of radiation from different wire materials in wire array Z-pinch plasma is a very challenging topic because it is almost impossible to separate different plasmas at the stagnation. A new approach is suggested based on planar wire array (PWA) loads to assess this problem. Multi-planar wire arrays are implemented that consist of few planes, each with the same number of wires and masses but from different wire materials, arranged in parallel rows. In particular, the experimental results obtained with triple PWAs (TPWAs) on the UNR Zebra generator are analyzed with Wire Ablation Dynamics Model, non-local thermodynamic equilibrium kineticmore » model, and 2D radiation magneto-hydrodynamic to illustrate this new approach. In TPWAs, two wire planes were from mid-atomic-number wire material and another plane was from alloyed Al, placed either in the middle or at the edge of the TPWA. Spatial and temporal properties of K-shell Al and L-shell Cu radiations were analyzed and compared from these two configurations of TPWAs. Advantages of the new approach are demonstrated and future work is discussed.« less

  15. Behavior of Steel-Sheathed Shear Walls Subjected to Seismic and Fire Loads.

    PubMed

    Hoehler, Matthew S; Smith, Christopher M; Hutchinson, Tara C; Wang, Xiang; Meacham, Brian J; Kamath, Praveen

    2017-07-01

    A series of tests was conducted on six 2.7 m × 3.7 m shear wall specimens consisting of cold-formed steel framing sheathed on one side with sheet steel adhered to gypsum board and on the opposite side with plain gypsum board. The specimens were subjected to various sequences of simulated seismic shear deformation and fire exposure to study the influence of multi-hazard interactions on the lateral load resistance of the walls. The test program was designed to complement a parallel effort at the University of California, San Diego to investigate a six-story building subjected to earthquakes and fires. The test results reported here indicate that the fire exposure caused a shift in the failure mode of the walls from local buckling of the sheet steel in cases without fire exposure, to global buckling of the sheet steel with an accompanying 35 % reduction in lateral load capacity after the wall had been exposed to fire. This behavior appears to be predictable, which is encouraging from the standpoint of residual lateral load capacity under these severe multi-hazard actions.

  16. CUDA Optimization Strategies for Compute- and Memory-Bound Neuroimaging Algorithms

    PubMed Central

    Lee, Daren; Dinov, Ivo; Dong, Bin; Gutman, Boris; Yanovsky, Igor; Toga, Arthur W.

    2011-01-01

    As neuroimaging algorithms and technology continue to grow faster than CPU performance in complexity and image resolution, data-parallel computing methods will be increasingly important. The high performance, data-parallel architecture of modern graphical processing units (GPUs) can reduce computational times by orders of magnitude. However, its massively threaded architecture introduces challenges when GPU resources are exceeded. This paper presents optimization strategies for compute- and memory-bound algorithms for the CUDA architecture. For compute-bound algorithms, the registers are reduced through variable reuse via shared memory and the data throughput is increased through heavier thread workloads and maximizing the thread configuration for a single thread block per multiprocessor. For memory-bound algorithms, fitting the data into the fast but limited GPU resources is achieved through reorganizing the data into self-contained structures and employing a multi-pass approach. Memory latencies are reduced by selecting memory resources whose cache performance are optimized for the algorithm's access patterns. We demonstrate the strategies on two computationally expensive algorithms and achieve optimized GPU implementations that perform up to 6× faster than unoptimized ones. Compared to CPU implementations, we achieve peak GPU speedups of 129× for the 3D unbiased nonlinear image registration technique and 93× for the non-local means surface denoising algorithm. PMID:21159404

  17. CUDA optimization strategies for compute- and memory-bound neuroimaging algorithms.

    PubMed

    Lee, Daren; Dinov, Ivo; Dong, Bin; Gutman, Boris; Yanovsky, Igor; Toga, Arthur W

    2012-06-01

    As neuroimaging algorithms and technology continue to grow faster than CPU performance in complexity and image resolution, data-parallel computing methods will be increasingly important. The high performance, data-parallel architecture of modern graphical processing units (GPUs) can reduce computational times by orders of magnitude. However, its massively threaded architecture introduces challenges when GPU resources are exceeded. This paper presents optimization strategies for compute- and memory-bound algorithms for the CUDA architecture. For compute-bound algorithms, the registers are reduced through variable reuse via shared memory and the data throughput is increased through heavier thread workloads and maximizing the thread configuration for a single thread block per multiprocessor. For memory-bound algorithms, fitting the data into the fast but limited GPU resources is achieved through reorganizing the data into self-contained structures and employing a multi-pass approach. Memory latencies are reduced by selecting memory resources whose cache performance are optimized for the algorithm's access patterns. We demonstrate the strategies on two computationally expensive algorithms and achieve optimized GPU implementations that perform up to 6× faster than unoptimized ones. Compared to CPU implementations, we achieve peak GPU speedups of 129× for the 3D unbiased nonlinear image registration technique and 93× for the non-local means surface denoising algorithm. Copyright © 2010 Elsevier Ireland Ltd. All rights reserved.

  18. A Parallel Finite Set Statistical Simulator for Multi-Target Detection and Tracking

    NASA Astrophysics Data System (ADS)

    Hussein, I.; MacMillan, R.

    2014-09-01

    Finite Set Statistics (FISST) is a powerful Bayesian inference tool for the joint detection, classification and tracking of multi-target environments. FISST is capable of handling phenomena such as clutter, misdetections, and target birth and decay. Implicit within the approach are solutions to the data association and target label-tracking problems. Finally, FISST provides generalized information measures that can be used for sensor allocation across different types of tasks such as: searching for new targets, and classification and tracking of known targets. These FISST capabilities have been demonstrated on several small-scale illustrative examples. However, for implementation in a large-scale system as in the Space Situational Awareness problem, these capabilities require a lot of computational power. In this paper, we implement FISST in a parallel environment for the joint detection and tracking of multi-target systems. In this implementation, false alarms and misdetections will be modeled. Target birth and decay will not be modeled in the present paper. We will demonstrate the success of the method for as many targets as we possibly can in a desktop parallel environment. Performance measures will include: number of targets in the simulation, certainty of detected target tracks, computational time as a function of clutter returns and number of targets, among other factors.

  19. Accelerating electrostatic surface potential calculation with multi-scale approximation on graphics processing units.

    PubMed

    Anandakrishnan, Ramu; Scogland, Tom R W; Fenley, Andrew T; Gordon, John C; Feng, Wu-chun; Onufriev, Alexey V

    2010-06-01

    Tools that compute and visualize biomolecular electrostatic surface potential have been used extensively for studying biomolecular function. However, determining the surface potential for large biomolecules on a typical desktop computer can take days or longer using currently available tools and methods. Two commonly used techniques to speed-up these types of electrostatic computations are approximations based on multi-scale coarse-graining and parallelization across multiple processors. This paper demonstrates that for the computation of electrostatic surface potential, these two techniques can be combined to deliver significantly greater speed-up than either one separately, something that is in general not always possible. Specifically, the electrostatic potential computation, using an analytical linearized Poisson-Boltzmann (ALPB) method, is approximated using the hierarchical charge partitioning (HCP) multi-scale method, and parallelized on an ATI Radeon 4870 graphical processing unit (GPU). The implementation delivers a combined 934-fold speed-up for a 476,040 atom viral capsid, compared to an equivalent non-parallel implementation on an Intel E6550 CPU without the approximation. This speed-up is significantly greater than the 42-fold speed-up for the HCP approximation alone or the 182-fold speed-up for the GPU alone. Copyright (c) 2010 Elsevier Inc. All rights reserved.

  20. Role of interbranch pumping on the quantum-statistical behavior of multi-mode magnons in ferromagnetic nanowires

    NASA Astrophysics Data System (ADS)

    Haghshenasfard, Zahra; Cottam, M. G.

    2018-01-01

    Theoretical studies are reported for the quantum-statistical properties of microwave-driven multi-mode magnon systems as represented by ferromagnetic nanowires with a stripe geometry. Effects of both the exchange and the dipole-dipole interactions, as well as a Zeeman term for an external applied field, are included in the magnetic Hamiltonian. The model also contains the time-dependent nonlinear effects due to parallel pumping with an electromagnetic field. Using a coherent magnon state representation in terms of creation and annihilation operators, we investigate the effects of parallel pumping on the temporal evolution of various nonclassical properties of the system. A focus is on the interbranch mixing produced by the pumping field when there are three or more modes. In particular, the occupation magnon number and the multi-mode cross correlations between magnon modes are studied. Manipulation of the collapse and revival phenomena of the average magnon occupation number and the control of the cross correlation between the magnon modes are demonstrated through tuning of the parallel pumping field amplitude and appropriate choices for the coherent magnon states. The cross correlations are a direct consequence of the interbranch pumping effects and do not appear in the corresponding one- or two-mode magnon systems.

Top