Large Scale Document Inversion using a Multi-threaded Computing System
Jung, Sungbo; Chang, Dar-Jen; Park, Juw Won
2018-01-01
Current microprocessor architecture is moving towards multi-core/multi-threaded systems. This trend has led to a surge of interest in using multi-threaded computing devices, such as the Graphics Processing Unit (GPU), for general purpose computing. We can utilize the GPU in computation as a massive parallel coprocessor because the GPU consists of multiple cores. The GPU is also an affordable, attractive, and user-programmable commodity. Nowadays a lot of information has been flooded into the digital domain around the world. Huge volume of data, such as digital libraries, social networking services, e-commerce product data, and reviews, etc., is produced or collected every moment with dramatic growth in size. Although the inverted index is a useful data structure that can be used for full text searches or document retrieval, a large number of documents will require a tremendous amount of time to create the index. The performance of document inversion can be improved by multi-thread or multi-core GPU. Our approach is to implement a linear-time, hash-based, single program multiple data (SPMD), document inversion algorithm on the NVIDIA GPU/CUDA programming platform utilizing the huge computational power of the GPU, to develop high performance solutions for document indexing. Our proposed parallel document inversion system shows 2-3 times faster performance than a sequential system on two different test datasets from PubMed abstract and e-commerce product reviews. CCS Concepts •Information systems➝Information retrieval • Computing methodologies➝Massively parallel and high-performance simulations. PMID:29861701
Large Scale Document Inversion using a Multi-threaded Computing System.
Jung, Sungbo; Chang, Dar-Jen; Park, Juw Won
2017-06-01
Current microprocessor architecture is moving towards multi-core/multi-threaded systems. This trend has led to a surge of interest in using multi-threaded computing devices, such as the Graphics Processing Unit (GPU), for general purpose computing. We can utilize the GPU in computation as a massive parallel coprocessor because the GPU consists of multiple cores. The GPU is also an affordable, attractive, and user-programmable commodity. Nowadays a lot of information has been flooded into the digital domain around the world. Huge volume of data, such as digital libraries, social networking services, e-commerce product data, and reviews, etc., is produced or collected every moment with dramatic growth in size. Although the inverted index is a useful data structure that can be used for full text searches or document retrieval, a large number of documents will require a tremendous amount of time to create the index. The performance of document inversion can be improved by multi-thread or multi-core GPU. Our approach is to implement a linear-time, hash-based, single program multiple data (SPMD), document inversion algorithm on the NVIDIA GPU/CUDA programming platform utilizing the huge computational power of the GPU, to develop high performance solutions for document indexing. Our proposed parallel document inversion system shows 2-3 times faster performance than a sequential system on two different test datasets from PubMed abstract and e-commerce product reviews. •Information systems➝Information retrieval • Computing methodologies➝Massively parallel and high-performance simulations.
A hybrid algorithm for parallel molecular dynamics simulations
NASA Astrophysics Data System (ADS)
Mangiardi, Chris M.; Meyer, R.
2017-10-01
This article describes algorithms for the hybrid parallelization and SIMD vectorization of molecular dynamics simulations with short-range forces. The parallelization method combines domain decomposition with a thread-based parallelization approach. The goal of the work is to enable efficient simulations of very large (tens of millions of atoms) and inhomogeneous systems on many-core processors with hundreds or thousands of cores and SIMD units with large vector sizes. In order to test the efficiency of the method, simulations of a variety of configurations with up to 74 million atoms have been performed. Results are shown that were obtained on multi-core systems with Sandy Bridge and Haswell processors as well as systems with Xeon Phi many-core processors.
Exact diagonalization of quantum lattice models on coprocessors
NASA Astrophysics Data System (ADS)
Siro, T.; Harju, A.
2016-10-01
We implement the Lanczos algorithm on an Intel Xeon Phi coprocessor and compare its performance to a multi-core Intel Xeon CPU and an NVIDIA graphics processor. The Xeon and the Xeon Phi are parallelized with OpenMP and the graphics processor is programmed with CUDA. The performance is evaluated by measuring the execution time of a single step in the Lanczos algorithm. We study two quantum lattice models with different particle numbers, and conclude that for small systems, the multi-core CPU is the fastest platform, while for large systems, the graphics processor is the clear winner, reaching speedups of up to 7.6 compared to the CPU. The Xeon Phi outperforms the CPU with sufficiently large particle number, reaching a speedup of 2.5.
Multi-Core Processor Memory Contention Benchmark Analysis Case Study
NASA Technical Reports Server (NTRS)
Simon, Tyler; McGalliard, James
2009-01-01
Multi-core processors dominate current mainframe, server, and high performance computing (HPC) systems. This paper provides synthetic kernel and natural benchmark results from an HPC system at the NASA Goddard Space Flight Center that illustrate the performance impacts of multi-core (dual- and quad-core) vs. single core processor systems. Analysis of processor design, application source code, and synthetic and natural test results all indicate that multi-core processors can suffer from significant memory subsystem contention compared to similar single-core processors.
Hybrid Parallelism for Volume Rendering on Large-, Multi-, and Many-Core Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Howison, Mark; Bethel, E. Wes; Childs, Hank
2012-01-01
With the computing industry trending towards multi- and many-core processors, we study how a standard visualization algorithm, ray-casting volume rendering, can benefit from a hybrid parallelism approach. Hybrid parallelism provides the best of both worlds: using distributed-memory parallelism across a large numbers of nodes increases available FLOPs and memory, while exploiting shared-memory parallelism among the cores within each node ensures that each node performs its portion of the larger calculation as efficiently as possible. We demonstrate results from weak and strong scaling studies, at levels of concurrency ranging up to 216,000, and with datasets as large as 12.2 trillion cells.more » The greatest benefit from hybrid parallelism lies in the communication portion of the algorithm, the dominant cost at higher levels of concurrency. We show that reducing the number of participants with a hybrid approach significantly improves performance.« less
Efficiency of static core turn-off in a system-on-a-chip with variation
Cher, Chen-Yong; Coteus, Paul W; Gara, Alan; Kursun, Eren; Paulsen, David P; Schuelke, Brian A; Sheets, II, John E; Tian, Shurong
2013-10-29
A processor-implemented method for improving efficiency of a static core turn-off in a multi-core processor with variation, the method comprising: conducting via a simulation a turn-off analysis of the multi-core processor at the multi-core processor's design stage, wherein the turn-off analysis of the multi-core processor at the multi-core processor's design stage includes a first output corresponding to a first multi-core processor core to turn off; conducting a turn-off analysis of the multi-core processor at the multi-core processor's testing stage, wherein the turn-off analysis of the multi-core processor at the multi-core processor's testing stage includes a second output corresponding to a second multi-core processor core to turn off; comparing the first output and the second output to determine if the first output is referring to the same core to turn off as the second output; outputting a third output corresponding to the first multi-core processor core if the first output and the second output are both referring to the same core to turn off.
Ding, Edwin; Lefrancois, Simon; Kutz, Jose Nathan; Wise, Frank W.
2011-01-01
The mode-locking of dissipative soliton fiber lasers using large mode area fiber supporting multiple transverse modes is studied experimentally and theoretically. The averaged mode-locking dynamics in a multi-mode fiber are studied using a distributed model. The co-propagation of multiple transverse modes is governed by a system of coupled Ginzburg–Landau equations. Simulations show that stable and robust mode-locked pulses can be produced. However, the mode-locking can be destabilized by excessive higher-order mode content. Experiments using large core step-index fiber, photonic crystal fiber, and chirally-coupled core fiber show that mode-locking can be significantly disturbed in the presence of higher-order modes, resulting in lower maximum single-pulse energies. In practice, spatial mode content must be carefully controlled to achieve full pulse energy scaling. This paper demonstrates that mode-locking performance is very sensitive to the presence of multiple waveguide modes when compared to systems such as amplifiers and continuous-wave lasers. PMID:21731106
Ding, Edwin; Lefrancois, Simon; Kutz, Jose Nathan; Wise, Frank W
2011-01-01
The mode-locking of dissipative soliton fiber lasers using large mode area fiber supporting multiple transverse modes is studied experimentally and theoretically. The averaged mode-locking dynamics in a multi-mode fiber are studied using a distributed model. The co-propagation of multiple transverse modes is governed by a system of coupled Ginzburg-Landau equations. Simulations show that stable and robust mode-locked pulses can be produced. However, the mode-locking can be destabilized by excessive higher-order mode content. Experiments using large core step-index fiber, photonic crystal fiber, and chirally-coupled core fiber show that mode-locking can be significantly disturbed in the presence of higher-order modes, resulting in lower maximum single-pulse energies. In practice, spatial mode content must be carefully controlled to achieve full pulse energy scaling. This paper demonstrates that mode-locking performance is very sensitive to the presence of multiple waveguide modes when compared to systems such as amplifiers and continuous-wave lasers.
Multi-core processing and scheduling performance in CMS
NASA Astrophysics Data System (ADS)
Hernández, J. M.; Evans, D.; Foulkes, S.
2012-12-01
Commodity hardware is going many-core. We might soon not be able to satisfy the job memory needs per core in the current single-core processing model in High Energy Physics. In addition, an ever increasing number of independent and incoherent jobs running on the same physical hardware not sharing resources might significantly affect processing performance. It will be essential to effectively utilize the multi-core architecture. CMS has incorporated support for multi-core processing in the event processing framework and the workload management system. Multi-core processing jobs share common data in memory, such us the code libraries, detector geometry and conditions data, resulting in a much lower memory usage than standard single-core independent jobs. Exploiting this new processing model requires a new model in computing resource allocation, departing from the standard single-core allocation for a job. The experiment job management system needs to have control over a larger quantum of resource since multi-core aware jobs require the scheduling of multiples cores simultaneously. CMS is exploring the approach of using whole nodes as unit in the workload management system where all cores of a node are allocated to a multi-core job. Whole-node scheduling allows for optimization of the data/workflow management (e.g. I/O caching, local merging) but efficient utilization of all scheduled cores is challenging. Dedicated whole-node queues have been setup at all Tier-1 centers for exploring multi-core processing workflows in CMS. We present the evaluation of the performance scheduling and executing multi-core workflows in whole-node queues compared to the standard single-core processing workflows.
Multi-core processing and scheduling performance in CMS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hernandez, J. M.; Evans, D.; Foulkes, S.
2012-01-01
Commodity hardware is going many-core. We might soon not be able to satisfy the job memory needs per core in the current single-core processing model in High Energy Physics. In addition, an ever increasing number of independent and incoherent jobs running on the same physical hardware not sharing resources might significantly affect processing performance. It will be essential to effectively utilize the multi-core architecture. CMS has incorporated support for multi-core processing in the event processing framework and the workload management system. Multi-core processing jobs share common data in memory, such us the code libraries, detector geometry and conditions data, resultingmore » in a much lower memory usage than standard single-core independent jobs. Exploiting this new processing model requires a new model in computing resource allocation, departing from the standard single-core allocation for a job. The experiment job management system needs to have control over a larger quantum of resource since multi-core aware jobs require the scheduling of multiples cores simultaneously. CMS is exploring the approach of using whole nodes as unit in the workload management system where all cores of a node are allocated to a multi-core job. Whole-node scheduling allows for optimization of the data/workflow management (e.g. I/O caching, local merging) but efficient utilization of all scheduled cores is challenging. Dedicated whole-node queues have been setup at all Tier-1 centers for exploring multi-core processing workflows in CMS. We present the evaluation of the performance scheduling and executing multi-core workflows in whole-node queues compared to the standard single-core processing workflows.« less
Toward GEOS-6, A Global Cloud System Resolving Atmospheric Model
NASA Technical Reports Server (NTRS)
Putman, William M.
2010-01-01
NASA is committed to observing and understanding the weather and climate of our home planet through the use of multi-scale modeling systems and space-based observations. Global climate models have evolved to take advantage of the influx of multi- and many-core computing technologies and the availability of large clusters of multi-core microprocessors. GEOS-6 is a next-generation cloud system resolving atmospheric model that will place NASA at the forefront of scientific exploration of our atmosphere and climate. Model simulations with GEOS-6 will produce a realistic representation of our atmosphere on the scale of typical satellite observations, bringing a visual comprehension of model results to a new level among the climate enthusiasts. In preparation for GEOS-6, the agency's flagship Earth System Modeling Framework [JDl] has been enhanced to support cutting-edge high-resolution global climate and weather simulations. Improvements include a cubed-sphere grid that exposes parallelism; a non-hydrostatic finite volume dynamical core, and algorithm designed for co-processor technologies, among others. GEOS-6 represents a fundamental advancement in the capability of global Earth system models. The ability to directly compare global simulations at the resolution of spaceborne satellite images will lead to algorithm improvements and better utilization of space-based observations within the GOES data assimilation system
Multi-element germanium detectors for synchrotron applications
NASA Astrophysics Data System (ADS)
Rumaiz, A. K.; Kuczewski, A. J.; Mead, J.; Vernon, E.; Pinelli, D.; Dooryhee, E.; Ghose, S.; Caswell, T.; Siddons, D. P.; Miceli, A.; Baldwin, J.; Almer, J.; Okasinski, J.; Quaranta, O.; Woods, R.; Krings, T.; Stock, S.
2018-04-01
We have developed a series of monolithic multi-element germanium detectors, based on sensor arrays produced by the Forschungzentrum Julich, and on Application-specific integrated circuits (ASICs) developed at Brookhaven. Devices have been made with element counts ranging from 64 to 384. These detectors are being used at NSLS-II and APS for a range of diffraction experiments, both monochromatic and energy-dispersive. Compact and powerful readout systems have been developed, based on the new generation of FPGA system-on-chip devices, which provide closely coupled multi-core processors embedded in large gate arrays. We will discuss the technical details of the systems, and present some of the results from them.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Burgett, Eric; Al-Sheikhly, Mohamad; Summers, Christopher
An advanced in-pile multi-parameter reactor monitoring system is being proposed in this funding opportunity. The proposed effort brings cutting edge, high-fidelity optical measurement systems into the reactor environment in an unprecedented fashion, including in-core, in-cladding and in-fuel pellet itself. Unlike instrumented leads, the proposed system provides a unique solution to a multi-parameter monitoring need in core while being minimally intrusive in the reactor core. Detector designs proposed herein can monitor fuel compression and expansion in both the radial and axial dimensions as well as monitor linear power profiles and fission rates during the operation of the reactor. In addition tomore » pressure, stress, strain, compression, neutron flux, neutron spectra, and temperature can be observed inside the fuel bundle and fuel rod using the proposed system. The proposed research aims at developing radiation-hard, harsh-environment multi-parameter systems for insertion into the reactor environment. The proposed research holds the potential to drastically increase the fidelity and precision of in-core instrumentation with little or no impact in the neutron economy in the reactor environment while providing a measurement system capable of operation for entire operating cycles. Significant work has been done over the last few years on the use of nanoparticle-based scintillators. Through the use of metamaterials, the PIs aim to develop planar neutron detectors and large-volume neutron detectors. These detectors will have high efficiencies for neutron detection and will have a high gamma discrimination capability.« less
Pair Formation of Hard Core Bosons in Flat Band Systems
NASA Astrophysics Data System (ADS)
Mielke, Andreas
2018-05-01
Hard core bosons in a large class of one or two dimensional flat band systems have an upper critical density, below which the ground states can be described completely. At the critical density, the ground states are Wigner crystals. If one adds a particle to the system at the critical density, the ground state and the low lying multi particle states of the system can be described as a Wigner crystal with an additional pair of particles. The energy band for the pair is separated from the rest of the multi-particle spectrum. The proofs use a Gerschgorin type of argument for block diagonally dominant matrices. In certain one-dimensional or tree-like structures one can show that the pair is localised, for example in the chequerboard chain. For this one-dimensional system with periodic boundary condition the energy band for the pair is flat, the pair is localised.
NASA Astrophysics Data System (ADS)
Tabik, S.; Romero, L. F.; Mimica, P.; Plata, O.; Zapata, E. L.
2012-09-01
A broad area in astronomy focuses on simulating extragalactic objects based on Very Long Baseline Interferometry (VLBI) radio-maps. Several algorithms in this scope simulate what would be the observed radio-maps if emitted from a predefined extragalactic object. This work analyzes the performance and scaling of this kind of algorithms on multi-socket, multi-core architectures. In particular, we evaluate a sharing approach, a privatizing approach and a hybrid approach on systems with complex memory hierarchy that includes shared Last Level Cache (LLC). In addition, we investigate which manual processes can be systematized and then automated in future works. The experiments show that the data-privatizing model scales efficiently on medium scale multi-socket, multi-core systems (up to 48 cores) while regardless of algorithmic and scheduling optimizations, the sharing approach is unable to reach acceptable scalability on more than one socket. However, the hybrid model with a specific level of data-sharing provides the best scalability over all used multi-socket, multi-core systems.
Earth observing system instrument pointing control modeling for polar orbiting platforms
NASA Technical Reports Server (NTRS)
Briggs, H. C.; Kia, T.; Mccabe, S. A.; Bell, C. E.
1987-01-01
An approach to instrument pointing control performance assessment for large multi-instrument platforms is described. First, instrument pointing requirements and reference platform control systems for the Eos Polar Platforms are reviewed. Performance modeling tools including NASTRAN models of two large platforms, a modal selection procedure utilizing a balanced realization method, and reduced order platform models with core and instrument pointing control loops added are then described. Time history simulations of instrument pointing and stability performance in response to commanded slewing of adjacent instruments demonstrates the limits of tolerable slew activity. Simplified models of rigid body responses are also developed for comparison. Instrument pointing control methods required in addition to the core platform control system to meet instrument pointing requirements are considered.
NASA Astrophysics Data System (ADS)
Aksenov, A. G.; Chechetkin, V. M.
2018-04-01
Most of the energy released in the gravitational collapse of the cores of massive stars is carried away by neutrinos. Neutrinos play a pivotal role in explaining core-collape supernovae. Currently, mathematical models of the gravitational collapse are based on multi-dimensional gas dynamics and thermonuclear reactions, while neutrino transport is considered in a simplified way. Multidimensional gas dynamics is used with neutrino transport in the flux-limited diffusion approximation to study the role of multi-dimensional effects. The possibility of large-scale convection is discussed, which is interesting both for explaining SN II and for setting up observations to register possible high-energy (≳10MeV) neutrinos from the supernova. A new multi-dimensional, multi-temperature gas dynamics method with neutrino transport is presented.
Multi-kW single fiber laser based on an extra large mode area fiber design
NASA Astrophysics Data System (ADS)
Langner, Andreas; Such, Mario; Schötz, Gerhard; Just, Florian; Leich, Martin; Schwuchow, Anka; Grimm, Stephan; Zimer, Hagen; Kozak, Marcin; Wedel, Björn; Rehmann, Georg; Bachert, Charley; Krause, Volker
2012-02-01
The quality of Yb-doped fused bulk silica produced by sintering of Yb-doped fused silica granulates has improved greatly in the past five years [1 - 4]. In particular, the refractive index and doping level homogeneity of such materials are excellent and we achieved excellent background fiber attenuation of the active core material down to about 20 dB/km at 1200 nm. The improvement of the Yb-doped fused bulk silica has enabled the development of multi-kW fiber laser systems based on a single extra large multimode laser fiber (XLMA fiber). When a single active fiber is used in combination with the XLMA multimode fiber of 1200 μm diameter simple and robust high power fiber laser setups without complex fiber coupling and fiber combiner systems become possible. In this papper, we will discuss in detail the development of the core material based on Yb-doped bulk silica and the characterization of Yb-doped fibers with different core compositions. We will also report on the excellent performance of a 4 kW fiber laser based on a single XLMA-fiber and show the first experimental welding results of steel sheets achieved with such a laser.
Neural networks within multi-core optic fibers
Cohen, Eyal; Malka, Dror; Shemer, Amir; Shahmoon, Asaf; Zalevsky, Zeev; London, Michael
2016-01-01
Hardware implementation of artificial neural networks facilitates real-time parallel processing of massive data sets. Optical neural networks offer low-volume 3D connectivity together with large bandwidth and minimal heat production in contrast to electronic implementation. Here, we present a conceptual design for in-fiber optical neural networks. Neurons and synapses are realized as individual silica cores in a multi-core fiber. Optical signals are transferred transversely between cores by means of optical coupling. Pump driven amplification in erbium-doped cores mimics synaptic interactions. We simulated three-layered feed-forward neural networks and explored their capabilities. Simulations suggest that networks can differentiate between given inputs depending on specific configurations of amplification; this implies classification and learning capabilities. Finally, we tested experimentally our basic neuronal elements using fibers, couplers, and amplifiers, and demonstrated that this configuration implements a neuron-like function. Therefore, devices similar to our proposed multi-core fiber could potentially serve as building blocks for future large-scale small-volume optical artificial neural networks. PMID:27383911
Neural networks within multi-core optic fibers.
Cohen, Eyal; Malka, Dror; Shemer, Amir; Shahmoon, Asaf; Zalevsky, Zeev; London, Michael
2016-07-07
Hardware implementation of artificial neural networks facilitates real-time parallel processing of massive data sets. Optical neural networks offer low-volume 3D connectivity together with large bandwidth and minimal heat production in contrast to electronic implementation. Here, we present a conceptual design for in-fiber optical neural networks. Neurons and synapses are realized as individual silica cores in a multi-core fiber. Optical signals are transferred transversely between cores by means of optical coupling. Pump driven amplification in erbium-doped cores mimics synaptic interactions. We simulated three-layered feed-forward neural networks and explored their capabilities. Simulations suggest that networks can differentiate between given inputs depending on specific configurations of amplification; this implies classification and learning capabilities. Finally, we tested experimentally our basic neuronal elements using fibers, couplers, and amplifiers, and demonstrated that this configuration implements a neuron-like function. Therefore, devices similar to our proposed multi-core fiber could potentially serve as building blocks for future large-scale small-volume optical artificial neural networks.
Design and Development of a Run-Time Monitor for Multi-Core Architectures in Cloud Computing
Kang, Mikyung; Kang, Dong-In; Crago, Stephen P.; Park, Gyung-Leen; Lee, Junghoon
2011-01-01
Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM) which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data. PMID:22163811
Design and development of a run-time monitor for multi-core architectures in cloud computing.
Kang, Mikyung; Kang, Dong-In; Crago, Stephen P; Park, Gyung-Leen; Lee, Junghoon
2011-01-01
Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM) which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data.
A highly efficient multi-core algorithm for clustering extremely large datasets
2010-01-01
Background In recent years, the demand for computational power in computational biology has increased due to rapidly growing data sets from microarray and other high-throughput technologies. This demand is likely to increase. Standard algorithms for analyzing data, such as cluster algorithms, need to be parallelized for fast processing. Unfortunately, most approaches for parallelizing algorithms largely rely on network communication protocols connecting and requiring multiple computers. One answer to this problem is to utilize the intrinsic capabilities in current multi-core hardware to distribute the tasks among the different cores of one computer. Results We introduce a multi-core parallelization of the k-means and k-modes cluster algorithms based on the design principles of transactional memory for clustering gene expression microarray type data and categorial SNP data. Our new shared memory parallel algorithms show to be highly efficient. We demonstrate their computational power and show their utility in cluster stability and sensitivity analysis employing repeated runs with slightly changed parameters. Computation speed of our Java based algorithm was increased by a factor of 10 for large data sets while preserving computational accuracy compared to single-core implementations and a recently published network based parallelization. Conclusions Most desktop computers and even notebooks provide at least dual-core processors. Our multi-core algorithms show that using modern algorithmic concepts, parallelization makes it possible to perform even such laborious tasks as cluster sensitivity and cluster number estimation on the laboratory computer. PMID:20370922
Rey-Villamizar, Nicolas; Somasundar, Vinay; Megjhani, Murad; Xu, Yan; Lu, Yanbin; Padmanabhan, Raghav; Trett, Kristen; Shain, William; Roysam, Badri
2014-01-01
In this article, we describe the use of Python for large-scale automated server-based bio-image analysis in FARSIGHT, a free and open-source toolkit of image analysis methods for quantitative studies of complex and dynamic tissue microenvironments imaged by modern optical microscopes, including confocal, multi-spectral, multi-photon, and time-lapse systems. The core FARSIGHT modules for image segmentation, feature extraction, tracking, and machine learning are written in C++, leveraging widely used libraries including ITK, VTK, Boost, and Qt. For solving complex image analysis tasks, these modules must be combined into scripts using Python. As a concrete example, we consider the problem of analyzing 3-D multi-spectral images of brain tissue surrounding implanted neuroprosthetic devices, acquired using high-throughput multi-spectral spinning disk step-and-repeat confocal microscopy. The resulting images typically contain 5 fluorescent channels. Each channel consists of 6000 × 10,000 × 500 voxels with 16 bits/voxel, implying image sizes exceeding 250 GB. These images must be mosaicked, pre-processed to overcome imaging artifacts, and segmented to enable cellular-scale feature extraction. The features are used to identify cell types, and perform large-scale analysis for identifying spatial distributions of specific cell types relative to the device. Python was used to build a server-based script (Dell 910 PowerEdge servers with 4 sockets/server with 10 cores each, 2 threads per core and 1TB of RAM running on Red Hat Enterprise Linux linked to a RAID 5 SAN) capable of routinely handling image datasets at this scale and performing all these processing steps in a collaborative multi-user multi-platform environment. Our Python script enables efficient data storage and movement between computers and storage servers, logs all the processing steps, and performs full multi-threaded execution of all codes, including open and closed-source third party libraries.
Efficient provisioning for multi-core applications with LSF
NASA Astrophysics Data System (ADS)
Dal Pra, Stefano
2015-12-01
Tier-1 sites providing computing power for HEP experiments are usually tightly designed for high throughput performances. This is pursued by reducing the variety of supported use cases and tuning for performances those ones, the most important of which have been that of singlecore jobs. Moreover, the usual workload is saturation: each available core in the farm is in use and there are queued jobs waiting for their turn to run. Enabling multi-core jobs thus requires dedicating a number of hosts where to run, and waiting for them to free the needed number of cores. This drain-time introduces a loss of computing power driven by the number of unusable empty cores. As an increasing demand for multi-core capable resources have emerged, a Task Force have been constituted in WLCG, with the goal to define a simple and efficient multi-core resource provisioning model. This paper details the work done at the INFN Tier-1 to enable multi-core support for the LSF batch system, with the intent of reducing to the minimum the average number of unused cores. The adopted strategy has been that of dedicating to multi-core a dynamic set of nodes, whose dimension is mainly driven by the number of pending multi-core requests and fair-share priority of the submitting user. The node status transition, from single to multi core et vice versa, is driven by a finite state machine which is implemented in a custom multi-core director script, running in the cluster. After describing and motivating both the implementation and the details specific to the LSF batch system, results about performance are reported. Factors having positive and negative impact on the overall efficiency are discussed and solutions to reduce at most the negative ones are proposed.
Energy-aware Thread and Data Management in Heterogeneous Multi-core, Multi-memory Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Su, Chun-Yi
By 2004, microprocessor design focused on multicore scaling—increasing the number of cores per die in each generation—as the primary strategy for improving performance. These multicore processors typically equip multiple memory subsystems to improve data throughput. In addition, these systems employ heterogeneous processors such as GPUs and heterogeneous memories like non-volatile memory to improve performance, capacity, and energy efficiency. With the increasing volume of hardware resources and system complexity caused by heterogeneity, future systems will require intelligent ways to manage hardware resources. Early research to improve performance and energy efficiency on heterogeneous, multi-core, multi-memory systems focused on tuning a single primitivemore » or at best a few primitives in the systems. The key limitation of past efforts is their lack of a holistic approach to resource management that balances the tradeoff between performance and energy consumption. In addition, the shift from simple, homogeneous systems to these heterogeneous, multicore, multi-memory systems requires in-depth understanding of efficient resource management for scalable execution, including new models that capture the interchange between performance and energy, smarter resource management strategies, and novel low-level performance/energy tuning primitives and runtime systems. Tuning an application to control available resources efficiently has become a daunting challenge; managing resources in automation is still a dark art since the tradeoffs among programming, energy, and performance remain insufficiently understood. In this dissertation, I have developed theories, models, and resource management techniques to enable energy-efficient execution of parallel applications through thread and data management in these heterogeneous multi-core, multi-memory systems. I study the effect of dynamic concurrent throttling on the performance and energy of multi-core, non-uniform memory access (NUMA) systems. I use critical path analysis to quantify memory contention in the NUMA memory system and determine thread mappings. In addition, I implement a runtime system that combines concurrent throttling and a novel thread mapping algorithm to manage thread resources and improve energy efficient execution in multi-core, NUMA systems.« less
Multi-element germanium detectors for synchrotron applications
Rumaiz, A. K.; Kuczewski, A. J.; Mead, J.; ...
2018-04-27
In this paper, we have developed a series of monolithic multi-element germanium detectors, based on sensor arrays produced by the Forschungzentrum Julich, and on Application-specific integrated circuits (ASICs) developed at Brookhaven. Devices have been made with element counts ranging from 64 to 384. These detectors are being used at NSLS-II and APS for a range of diffraction experiments, both monochromatic and energy-dispersive. Compact and powerful readout systems have been developed, based on the new generation of FPGA system-on-chip devices, which provide closely coupled multi-core processors embedded in large gate arrays. Finally, we will discuss the technical details of the systems,more » and present some of the results from them.« less
Multi-element germanium detectors for synchrotron applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rumaiz, A. K.; Kuczewski, A. J.; Mead, J.
In this paper, we have developed a series of monolithic multi-element germanium detectors, based on sensor arrays produced by the Forschungzentrum Julich, and on Application-specific integrated circuits (ASICs) developed at Brookhaven. Devices have been made with element counts ranging from 64 to 384. These detectors are being used at NSLS-II and APS for a range of diffraction experiments, both monochromatic and energy-dispersive. Compact and powerful readout systems have been developed, based on the new generation of FPGA system-on-chip devices, which provide closely coupled multi-core processors embedded in large gate arrays. Finally, we will discuss the technical details of the systems,more » and present some of the results from them.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chin, George; Marquez, Andres; Choudhury, Sutanay
2012-09-01
Triadic analysis encompasses a useful set of graph mining methods that is centered on the concept of a triad, which is a subgraph of three nodes and the configuration of directed edges across the nodes. Such methods are often applied in the social sciences as well as many other diverse fields. Triadic methods commonly operate on a triad census that counts the number of triads of every possible edge configuration in a graph. Like other graph algorithms, triadic census algorithms do not scale well when graphs reach tens of millions to billions of nodes. To enable the triadic analysis ofmore » large-scale graphs, we developed and optimized a triad census algorithm to efficiently execute on shared memory architectures. We will retrace the development and evolution of a parallel triad census algorithm. Over the course of several versions, we continually adapted the code’s data structures and program logic to expose more opportunities to exploit parallelism on shared memory that would translate into improved computational performance. We will recall the critical steps and modifications that occurred during code development and optimization. Furthermore, we will compare the performances of triad census algorithm versions on three specific systems: Cray XMT, HP Superdome, and AMD multi-core NUMA machine. These three systems have shared memory architectures but with markedly different hardware capabilities to manage parallelism.« less
Visible CWDM system design for Multi-Gbit/s transmission over SI-POF
NASA Astrophysics Data System (ADS)
Vázquez, Carmen; Pinzón, Plinio Jesús; Pérez, Isabel
2015-01-01
In order to increase the data rates of Multi-Gbit/s links based on large core step index (SI) plastic optical fibers (POF), different modulation scenes have been proposed. Another option is to use multiple optical carriers for parallel transmission of communication channels over the same fiber. Some designs to reach data rates of 14.77 Gb/s in 50 m, with 4 channels have been developed by off line processing. In this work, designs to test the potential of real Multi- Gbit/s transmission systems using commercial products are reported. Special care in designing low insertion loss multiplexers and demultiplexers is carried out to allow for greener solutions in terms of power consumption.
Stack-and-Draw Manufacture Process of a Seven-Core Optical Fiber for Fluorescence Measurements
NASA Astrophysics Data System (ADS)
Samir, Ahmed; Batagelj, Bostjan
2018-01-01
Multi-core, optical-fiber technology is expected to be used in telecommunications and sensory systems in a relatively short amount of time. However, a successful transition from research laboratories to industry applications will only be possible with an optimized design and manufacturing process. The fabrication process is an important aspect in designing and developing new multi-applicable, multi-core fibers, where the best candidate is a seven-core fiber. Here, the basics for designing and manufacturing a single-mode, seven-core fiber using the stack-and-draw process is described for the example of a fluorescence sensory system.
Oryspayev, Dossay; Aktulga, Hasan Metin; Sosonkina, Masha; ...
2015-07-14
In this article, sparse matrix vector multiply (SpMVM) is an important kernel that frequently arises in high performance computing applications. Due to its low arithmetic intensity, several approaches have been proposed in literature to improve its scalability and efficiency in large scale computations. In this paper, our target systems are high end multi-core architectures and we use messaging passing interface + open multiprocessing hybrid programming model for parallelism. We analyze the performance of recently proposed implementation of the distributed symmetric SpMVM, originally developed for large sparse symmetric matrices arising in ab initio nuclear structure calculations. We also study important featuresmore » of this implementation and compare with previously reported implementations that do not exploit underlying symmetry. Our SpMVM implementations leverage the hybrid paradigm to efficiently overlap expensive communications with computations. Our main comparison criterion is the "CPU core hours" metric, which is the main measure of resource usage on supercomputers. We analyze the effects of topology-aware mapping heuristic using simplified network load model. Furthermore, we have tested the different SpMVM implementations on two large clusters with 3D Torus and Dragonfly topology. Our results show that the distributed SpMVM implementation that exploits matrix symmetry and hides communication yields the best value for the "CPU core hours" metric and significantly reduces data movement overheads.« less
NASA Technical Reports Server (NTRS)
Righter, K.; Pando, K.; Danielson, L.
2014-01-01
Numerous geophysical and geochemical studies have suggested the existence of a small metallic lunar core, but the composition of that core is not known. Knowledge of the composition can have a large impact on the thermal evolution of the core, its possible early dynamo creation, and its overall size and fraction of solid and liquid. Thermal models predict that the current temperature at the core-mantle boundary of the Moon is near 1650 K. Re-evaluation of Apollo seismic data has highlighted the need for new data in a broader range of bulk core compositions in the PT range of the lunar core. Geochemical measurements have suggested a more volatile-rich Moon than previously thought. And GRAIL mission data may allow much better constraints on the physical nature of the lunar core. All of these factors have led us to determine new phase equilibria experimental studies in the Fe-Ni-S-C-Si system in the relevant PT range of the lunar core that will help constrain the composition of Moon's core.
Core-to-core uniformity improvement in multi-core fiber Bragg gratings
NASA Astrophysics Data System (ADS)
Lindley, Emma; Min, Seong-Sik; Leon-Saval, Sergio; Cvetojevic, Nick; Jovanovic, Nemanja; Bland-Hawthorn, Joss; Lawrence, Jon; Gris-Sanchez, Itandehui; Birks, Tim; Haynes, Roger; Haynes, Dionne
2014-07-01
Multi-core fiber Bragg gratings (MCFBGs) will be a valuable tool not only in communications but also various astronomical, sensing and industry applications. In this paper we address some of the technical challenges of fabricating effective multi-core gratings by simulating improvements to the writing method. These methods allow a system designed for inscribing single-core fibers to cope with MCFBG fabrication with only minor, passive changes to the writing process. Using a capillary tube that was polished on one side, the field entering the fiber was flattened which improved the coverage and uniformity of all cores.
MIMO signal progressing with RLSCMA algorithm for multi-mode multi-core optical transmission system
NASA Astrophysics Data System (ADS)
Bi, Yuan; Liu, Bo; Zhang, Li-jia; Xin, Xiang-jun; Zhang, Qi; Wang, Yong-jun; Tian, Qing-hua; Tian, Feng; Mao, Ya-ya
2018-01-01
In the process of transmitting signals of multi-mode multi-core fiber, there will be mode coupling between modes. The mode dispersion will also occur because each mode has different transmission speed in the link. Mode coupling and mode dispersion will cause damage to the useful signal in the transmission link, so the receiver needs to deal received signal with digital signal processing, and compensate the damage in the link. We first analyzes the influence of mode coupling and mode dispersion in the process of transmitting signals of multi-mode multi-core fiber, then presents the relationship between the coupling coefficient and dispersion coefficient. Then we carry out adaptive signal processing with MIMO equalizers based on recursive least squares constant modulus algorithm (RLSCMA). The MIMO equalization algorithm offers adaptive equalization taps according to the degree of crosstalk in cores or modes, which eliminates the interference among different modes and cores in space division multiplexing(SDM) transmission system. The simulation results show that the distorted signals are restored efficiently with fast convergence speed.
Micro CT characterization of a coastal mine tailings deposit, Portmán Bay, SE Spain
NASA Astrophysics Data System (ADS)
Frigola, Jaime; Cerdà-Domènech, Marc; Barriuso, Eduardo; Sanchez-Vidal, Anna; Amblas, David; Canals, Miquel
2017-04-01
Scanning of sediment cores by means of high-resolution non-destructive techniques provides researchers with huge amounts of highly valuable data allowing fast and detailed characterization of the materials. In the last decades several devoted instruments have been developed and applied to the study of sedimentary sequences, mainly multi-sensor core loggers (MSCL) for the physical properties and XRF core scanners for the chemical elemental composition. The geoscientific community started using computed tomography (CT) systems about two decades ago. These were mainly medical systems as dedicated instruments were essentially lacking by that time. The resolution of those medical systems was limited to several hundreds of micrometres voxel size. Micro computed tomography (micro-CT) systems have also spread into geoscientific research, although their limited workspace dimensions prevents their use for large objects, such as long sediment cores. Recently, a new micro-CT system, the MultiTom Core X-ray CT, conceived by University of Barcelona (UB) researchers and developed by X-ray Engineering, became operational. It is able of scanning sediment cores up to 1.5 m long, and allows adjustable resolutions from 300 microns down to 3-4 microns. The system is now installed at UB's CORELAB Laboratory for non-destructive analyses of geological materials. Here we present, as an example, the results of MultiTom scans of a set of sediment cores recovered offshore Portmán Bay, SE Spain, in order to characterize at very high-resolution the metal-enriched deposit generated after 33 years of direct discharge into the sea of mine tailings resulting from the exploitation of Pb and Zn ores. In total 52 short cores and 6 long gravity cores from the mine tailings infilled bay were scanned with the MultiTom system at a mean voxel resolution of 125 microns. The integrated study of micro-CT data allowed differentiating the main tailings units from deposits formed after disposal cessation. Tailings units show higher radio-density values, which correspond to metal enrichments. A lower unit consists of highly laminated interbedded low radio-density and very high radio-density layers, while an upper mine tailings unit is more homogeneous and shows intermediate radio-density values. The limit between the tailings and the post-mining deposits is defined by a sharp surface associated with an abrupt decrease in the radio-densities. Post-mining deposits are also characterized by an increment in bioturbation marks, which are practically absent in the tailings units, and an increase in carbonate particles and organic matter patches. Micro CT scans allow observation of very small structures, which are indicative of the complexity of the sedimentation processes involved in the transport and final deposition of the mine tailings. Integration of micro CT scans together with XRF core scanner and MSCL data allows a better characterization of the metal concentrations and their distribution within the deposit, directly demonstrating the great value of non-destructive techniques for actually high-resolution sedimentological studies.
Optical properties of core-shell and multi-shell nanorods
NASA Astrophysics Data System (ADS)
Mokkath, Junais Habeeb; Shehata, Nader
2018-05-01
We report a first-principles time dependent density functional theory study of the optical response modulations in bimetallic core-shell (Na@Al and Al@Na) and multi-shell (Al@Na@Al@Na and Na@Al@Na@Al: concentric shells of Al and Na alternate) nanorods. All of the core-shell and multi-shell configurations display highly enhanced absorption intensity with respect to the pure Al and Na nanorods, showing sensitivity to both composition and chemical ordering. Remarkably large spectral intensity enhancements were found in a couple of core-shell configurations, indicative that optical response averaging based on the individual components can not be considered as true as always in the case of bimetallic core-shell nanorods. We believe that our theoretical results would be useful in promising applications depending on Aluminum-based plasmonic materials such as solar cells and sensors.
In-Pile Instrumentation Multi- Parameter System Utilizing Photonic Fibers and Nanovision
DOE Office of Scientific and Technical Information (OSTI.GOV)
Burgett, Eric
2015-10-13
An advanced in-pile multi-parameter reactor monitoring system is being proposed in this funding opportunity. The proposed effort brings cutting edge, high fidelity optical measurement systems into the reactor environment in an unprecedented fashion, including in-core, in-cladding and in-fuel pellet itself. Unlike instrumented leads, the proposed system provides a unique solution to a multi-parameter monitoring need in core while being minimally intrusive in the reactor core. Detector designs proposed herein can monitor fuel compression and expansion in both the radial and axial dimensions as well as monitor linear power profiles and fission rates during the operation of the reactor. In additionmore » to pressure, stress, strain, compression, neutron flux, neutron spectra, and temperature can be observed inside the fuel bundle and fuel rod using the proposed system. The proposed research aims at developing radiation-hard, harsh-environment multi-parameter systems for insertion into the reactor environment. The proposed research holds the potential to drastically increase the fidelity and precision of in-core instrumentation with little or no impact in the neutron economy in the reactor environment while providing a measurement system capable of operation for entire operating cycles.« less
Multi-view L2-SVM and its multi-view core vector machine.
Huang, Chengquan; Chung, Fu-lai; Wang, Shitong
2016-03-01
In this paper, a novel L2-SVM based classifier Multi-view L2-SVM is proposed to address multi-view classification tasks. The proposed Multi-view L2-SVM classifier does not have any bias in its objective function and hence has the flexibility like μ-SVC in the sense that the number of the yielded support vectors can be controlled by a pre-specified parameter. The proposed Multi-view L2-SVM classifier can make full use of the coherence and the difference of different views through imposing the consensus among multiple views to improve the overall classification performance. Besides, based on the generalized core vector machine GCVM, the proposed Multi-view L2-SVM classifier is extended into its GCVM version MvCVM which can realize its fast training on large scale multi-view datasets, with its asymptotic linear time complexity with the sample size and its space complexity independent of the sample size. Our experimental results demonstrated the effectiveness of the proposed Multi-view L2-SVM classifier for small scale multi-view datasets and the proposed MvCVM classifier for large scale multi-view datasets. Copyright © 2015 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Halimah, B. Z.; Azlina, A.; Sembok, T. M.; Sufian, I.; Sharul Azman, M. N.; Azuraliza, A. B.; Zulaiha, A. O.; Nazlia, O.; Salwani, A.; Sanep, A.; Hailani, M. T.; Zaher, M. Z.; Azizah, J.; Nor Faezah, M. Y.; Choo, W. O.; Abdullah, Chew; Sopian, B.
The Holistic Islamic Banking System (HiCORE), a banking system suitable for virtual banking environment, created based on universityindustry collaboration initiative between Universiti Kebangsaan Malaysia (UKM) and Fuziq Software Sdn Bhd. HiCORE was modeled on a multitiered Simple - Services Oriented Architecture (S-SOA), using the parameterbased semantic approach. HiCORE's existence is timely as the financial world is looking for a new approach to creating banking and financial products that are interest free or based on the Islamic Syariah principles and jurisprudence. An interest free banking system has currently caught the interest of bankers and financiers all over the world. HiCORE's Parameter-based module houses the Customer-information file (CIF), Deposit and Financing components. The Parameter based module represents the third tier of the multi-tiered Simple SOA approach. This paper highlights the multi-tiered parameter- driven approach to the creation of new Islamiic products based on the 'dalil' (Quran), 'syarat' (rules) and 'rukun' (procedures) as required by the syariah principles and jurisprudence reflected by the semantic ontology embedded in the parameter module of the system.
Multiphoton endoscopy based on a mode-filtered single-mode fiber
NASA Astrophysics Data System (ADS)
Moon, Sucbei; Liu, Gangjun; Chen, Zhongping
2011-03-01
We present a new low-nonlinearity fiber of mode-filtered large-core fiber for flexible beam delivery of intense pulsed light aiming at multi-photon endoscopy application. A multimode fiber of a large core diameter (20 μm) equips a mode filtering means in the middle of the fiber link to suppress the high-order modes selectively. A large effective core area of ~200 μm2 has been achieved at 0.8-μm and 1.0-μm bands. This is 8 times larger than the core area of a conventional SMF used for those spectral bands. Various advantages of our large-mode area fiber will be demonstrated and discussed in this report.
Parallel Index and Query for Large Scale Data Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chou, Jerry; Wu, Kesheng; Ruebel, Oliver
2011-07-18
Modern scientific datasets present numerous data management and analysis challenges. State-of-the-art index and query technologies are critical for facilitating interactive exploration of large datasets, but numerous challenges remain in terms of designing a system for process- ing general scientific datasets. The system needs to be able to run on distributed multi-core platforms, efficiently utilize underlying I/O infrastructure, and scale to massive datasets. We present FastQuery, a novel software framework that address these challenges. FastQuery utilizes a state-of-the-art index and query technology (FastBit) and is designed to process mas- sive datasets on modern supercomputing platforms. We apply FastQuery to processing ofmore » a massive 50TB dataset generated by a large scale accelerator modeling code. We demonstrate the scalability of the tool to 11,520 cores. Motivated by the scientific need to search for inter- esting particles in this dataset, we use our framework to reduce search time from hours to tens of seconds.« less
Fast and Accurate Simulation of the Cray XMT Multithreaded Supercomputer
DOE Office of Scientific and Technical Information (OSTI.GOV)
Villa, Oreste; Tumeo, Antonino; Secchi, Simone
Irregular applications, such as data mining and analysis or graph-based computations, show unpredictable memory/network access patterns and control structures. Highly multithreaded architectures with large processor counts, like the Cray MTA-1, MTA-2 and XMT, appear to address their requirements better than commodity clusters. However, the research on highly multithreaded systems is currently limited by the lack of adequate architectural simulation infrastructures due to issues such as size of the machines, memory footprint, simulation speed, accuracy and customization. At the same time, Shared-memory MultiProcessors (SMPs) with multi-core processors have become an attractive platform to simulate large scale machines. In this paper, wemore » introduce a cycle-level simulator of the highly multithreaded Cray XMT supercomputer. The simulator runs unmodified XMT applications. We discuss how we tackled the challenges posed by its development, detailing the techniques introduced to make the simulation as fast as possible while maintaining a high accuracy. By mapping XMT processors (ThreadStorm with 128 hardware threads) to host computing cores, the simulation speed remains constant as the number of simulated processors increases, up to the number of available host cores. The simulator supports zero-overhead switching among different accuracy levels at run-time and includes a network model that takes into account contention. On a modern 48-core SMP host, our infrastructure simulates a large set of irregular applications 500 to 2000 times slower than real time when compared to a 128-processor XMT, while remaining within 10\\% of accuracy. Emulation is only from 25 to 200 times slower than real time.« less
Multi-phase model development to assess RCIC system capabilities under severe accident conditions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kirkland, Karen Vierow; Ross, Kyle; Beeny, Bradley
The Reactor Core Isolation Cooling (RCIC) System is a safety-related system that provides makeup water for core cooling of some Boiling Water Reactors (BWRs) with a Mark I containment. The RCIC System consists of a steam-driven Terry turbine that powers a centrifugal, multi-stage pump for providing water to the reactor pressure vessel. The Fukushima Dai-ichi accidents demonstrated that the RCIC System can play an important role under accident conditions in removing core decay heat. The unexpectedly sustained, good performance of the RCIC System in the Fukushima reactor demonstrates, firstly, that its capabilities are not well understood, and secondly, that themore » system has high potential for extended core cooling in accident scenarios. Better understanding and analysis tools would allow for more options to cope with a severe accident situation and to reduce the consequences. The objectives of this project were to develop physics-based models of the RCIC System, incorporate them into a multi-phase code and validate the models. This Final Technical Report details the progress throughout the project duration and the accomplishments.« less
Adapting wave-front algorithms to efficiently utilize systems with deep communication hierarchies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kerbyson, Darren J; Lang, Michael; Pakin, Scott
2009-01-01
Large-scale systems increasingly exhibit a differential between intra-chip and inter-chip communication performance. Processor-cores on the same socket are able to communicate at lower latencies, and with higher bandwidths, than cores on different sockets either within the same node or between nodes. A key challenge is to efficiently use this communication hierarchy and hence optimize performance. We consider here the class of applications that contain wave-front processing. In these applications data can only be processed after their upstream neighbors have been processed. Similar dependencies result between processors in which communication is required to pass boundary data downstream and whose cost ismore » typically impacted by the slowest communication channel in use. In this work we develop a novel hierarchical wave-front approach that reduces the use of slower communications in the hierarchy but at the cost of additional computation and higher use of on-chip communications. This tradeoff is explored using a performance model and an implementation on the Petascale Roadrunner system demonstrates a 27% performance improvement at full system-scale on a kernel application. The approach is generally applicable to large-scale multi-core and accelerated systems where a differential in system communication performance exists.« less
Design and realization of the real-time spectrograph controller for LAMOST based on FPGA
NASA Astrophysics Data System (ADS)
Wang, Jianing; Wu, Liyan; Zeng, Yizhong; Dai, Songxin; Hu, Zhongwen; Zhu, Yongtian; Wang, Lei; Wu, Zhen; Chen, Yi
2008-08-01
A large Schmitt reflector telescope, Large Sky Area Multi-Object Fiber Spectroscopic Telescope(LAMOST), is being built in China, which has effective aperture of 4 meters and can observe the spectra of as many as 4000 objects simultaneously. To fit such a large amount of observational objects, the dispersion part is composed of a set of 16 multipurpose fiber-fed double-beam Schmidt spectrographs, of which each has about ten of moveable components realtimely accommodated and manipulated by a controller. An industrial Ethernet network connects those 16 spectrograph controllers. The light from stars is fed to the entrance slits of the spectrographs with optical fibers. In this paper, we mainly introduce the design and realization of our real-time controller for the spectrograph, our design using the technique of System On Programmable Chip (SOPC) based on Field Programmable Gate Array (FPGA) and then realizing the control of the spectrographs through NIOSII Soft Core Embedded Processor. We seal the stepper motor controller as intellectual property (IP) cores and reuse it, greatly simplifying the design process and then shortening the development time. Under the embedded operating system μC/OS-II, a multi-tasks control program has been well written to realize the real-time control of the moveable parts of the spectrographs. At present, a number of such controllers have been applied in the spectrograph of LAMOST.
Simulation of DKIST solar adaptive optics system
NASA Astrophysics Data System (ADS)
Marino, Jose; Carlisle, Elizabeth; Schmidt, Dirk
2016-07-01
Solar adaptive optics (AO) simulations are a valuable tool to guide the design and optimization process of current and future solar AO and multi-conjugate AO (MCAO) systems. Solar AO and MCAO systems rely on extended object cross-correlating Shack-Hartmann wavefront sensors to measure the wavefront. Accurate solar AO simulations require computationally intensive operations, which have until recently presented a prohibitive computational cost. We present an update on the status of a solar AO and MCAO simulation tool being developed at the National Solar Observatory. The simulation tool is a multi-threaded application written in the C++ language that takes advantage of current large multi-core CPU computer systems and fast ethernet connections to provide accurate full simulation of solar AO and MCAO systems. It interfaces with KAOS, a state of the art solar AO control software developed by the Kiepenheuer-Institut fuer Sonnenphysik, that provides reliable AO control. We report on the latest results produced by the solar AO simulation tool.
CMS Readiness for Multi-Core Workload Scheduling
DOE Office of Scientific and Technical Information (OSTI.GOV)
Perez-Calero Yzquierdo, A.; Balcas, J.; Hernandez, J.
In the present run of the LHC, CMS data reconstruction and simulation algorithms benefit greatly from being executed as multiple threads running on several processor cores. The complexity of the Run 2 events requires parallelization of the code to reduce the memory-per- core footprint constraining serial execution programs, thus optimizing the exploitation of present multi-core processor architectures. The allocation of computing resources for multi-core tasks, however, becomes a complex problem in itself. The CMS workload submission infrastructure employs multi-slot partitionable pilots, built on HTCondor and GlideinWMS native features, to enable scheduling of single and multi-core jobs simultaneously. This provides amore » solution for the scheduling problem in a uniform way across grid sites running a diversity of gateways to compute resources and batch system technologies. This paper presents this strategy and the tools on which it has been implemented. The experience of managing multi-core resources at the Tier-0 and Tier-1 sites during 2015, along with the deployment phase to Tier-2 sites during early 2016 is reported. The process of performance monitoring and optimization to achieve efficient and flexible use of the resources is also described.« less
CMS readiness for multi-core workload scheduling
NASA Astrophysics Data System (ADS)
Perez-Calero Yzquierdo, A.; Balcas, J.; Hernandez, J.; Aftab Khan, F.; Letts, J.; Mason, D.; Verguilov, V.
2017-10-01
In the present run of the LHC, CMS data reconstruction and simulation algorithms benefit greatly from being executed as multiple threads running on several processor cores. The complexity of the Run 2 events requires parallelization of the code to reduce the memory-per- core footprint constraining serial execution programs, thus optimizing the exploitation of present multi-core processor architectures. The allocation of computing resources for multi-core tasks, however, becomes a complex problem in itself. The CMS workload submission infrastructure employs multi-slot partitionable pilots, built on HTCondor and GlideinWMS native features, to enable scheduling of single and multi-core jobs simultaneously. This provides a solution for the scheduling problem in a uniform way across grid sites running a diversity of gateways to compute resources and batch system technologies. This paper presents this strategy and the tools on which it has been implemented. The experience of managing multi-core resources at the Tier-0 and Tier-1 sites during 2015, along with the deployment phase to Tier-2 sites during early 2016 is reported. The process of performance monitoring and optimization to achieve efficient and flexible use of the resources is also described.
Options for Parallelizing a Planning and Scheduling Algorithm
NASA Technical Reports Server (NTRS)
Clement, Bradley J.; Estlin, Tara A.; Bornstein, Benjamin D.
2011-01-01
Space missions have a growing interest in putting multi-core processors onboard spacecraft. For many missions processing power significantly slows operations. We investigate how continual planning and scheduling algorithms can exploit multi-core processing and outline different potential design decisions for a parallelized planning architecture. This organization of choices and challenges helps us with an initial design for parallelizing the CASPER planning system for a mesh multi-core processor. This work extends that presented at another workshop with some preliminary results.
Nonlinear Light Dynamics in Multi-Core Structures
2017-02-27
be generated in continuous- discrete optical media such as multi-core optical fiber or waveguide arrays; localisation dynamics in a continuous... discrete nonlinear system. Detailed theoretical analysis is presented of the existence and stability of the discrete -continuous light bullets using a very...and pulse compression using wave collapse (self-focusing) energy localisation dynamics in a continuous- discrete nonlinear system, as implemented in a
Software Defined Radio with Parallelized Software Architecture
NASA Technical Reports Server (NTRS)
Heckler, Greg
2013-01-01
This software implements software-defined radio procession over multi-core, multi-CPU systems in a way that maximizes the use of CPU resources in the system. The software treats each processing step in either a communications or navigation modulator or demodulator system as an independent, threaded block. Each threaded block is defined with a programmable number of input or output buffers; these buffers are implemented using POSIX pipes. In addition, each threaded block is assigned a unique thread upon block installation. A modulator or demodulator system is built by assembly of the threaded blocks into a flow graph, which assembles the processing blocks to accomplish the desired signal processing. This software architecture allows the software to scale effortlessly between single CPU/single-core computers or multi-CPU/multi-core computers without recompilation. NASA spaceflight and ground communications systems currently rely exclusively on ASICs or FPGAs. This software allows low- and medium-bandwidth (100 bps to .50 Mbps) software defined radios to be designed and implemented solely in C/C++ software, while lowering development costs and facilitating reuse and extensibility.
Using Multi-Core Systems for Rover Autonomy
NASA Technical Reports Server (NTRS)
Clement, Brad; Estlin, Tara; Bornstein, Benjamin; Springer, Paul; Anderson, Robert C.
2010-01-01
Task Objectives are: (1) Develop and demonstrate key capabilities for rover long-range science operations using multi-core computing, (a) Adapt three rover technologies to execute on SOA multi-core processor (b) Illustrate performance improvements achieved (c) Demonstrate adapted capabilities with rover hardware, (2) Targeting three high-level autonomy technologies (a) Two for onboard data analysis (b) One for onboard command sequencing/planning, (3) Technologies identified as enabling for future missions, (4)Benefits will be measured along several metrics: (a) Execution time / Power requirements (b) Number of data products processed per unit time (c) Solution quality
Data Acquisition System for Multi-Frequency Radar Flight Operations Preparation
NASA Technical Reports Server (NTRS)
Leachman, Jonathan
2010-01-01
A three-channel data acquisition system was developed for the NASA Multi-Frequency Radar (MFR) system. The system is based on a commercial-off-the-shelf (COTS) industrial PC (personal computer) and two dual-channel 14-bit digital receiver cards. The decimated complex envelope representations of the three radar signals are passed to the host PC via the PCI bus, and then processed in parallel by multiple cores of the PC CPU (central processing unit). The innovation is this parallelization of the radar data processing using multiple cores of a standard COTS multi-core CPU. The data processing portion of the data acquisition software was built using autonomous program modules or threads, which can run simultaneously on different cores. A master program module calculates the optimal number of processing threads, launches them, and continually supplies each with data. The benefit of this new parallel software architecture is that COTS PCs can be used to implement increasingly complex processing algorithms on an increasing number of radar range gates and data rates. As new PCs become available with higher numbers of CPU cores, the software will automatically utilize the additional computational capacity.
1998-04-01
The result of the project is a demonstration of the fusion process, the sensors management and the real-time capabilities using simulated sensors...demonstrator (TAD) is a system that demonstrates the core ele- ment of a battlefield ground surveillance system by simulation in near real-time. The core...Management and Sensor/Platform simulation . The surveillance system observes the real world through a non-collocated heterogene- ous multisensory system
Behavior-aware cache hierarchy optimization for low-power multi-core embedded systems
NASA Astrophysics Data System (ADS)
Zhao, Huatao; Luo, Xiao; Zhu, Chen; Watanabe, Takahiro; Zhu, Tianbo
2017-07-01
In modern embedded systems, the increasing number of cores requires efficient cache hierarchies to ensure data throughput, but such cache hierarchies are restricted by their tumid size and interference accesses which leads to both performance degradation and wasted energy. In this paper, we firstly propose a behavior-aware cache hierarchy (BACH) which can optimally allocate the multi-level cache resources to many cores and highly improved the efficiency of cache hierarchy, resulting in low energy consumption. The BACH takes full advantage of the explored application behaviors and runtime cache resource demands as the cache allocation bases, so that we can optimally configure the cache hierarchy to meet the runtime demand. The BACH was implemented on the GEM5 simulator. The experimental results show that energy consumption of a three-level cache hierarchy can be saved from 5.29% up to 27.94% compared with other key approaches while the performance of the multi-core system even has a slight improvement counting in hardware overhead.
Fault-Tolerant, Real-Time, Multi-Core Computer System
NASA Technical Reports Server (NTRS)
Gostelow, Kim P.
2012-01-01
A document discusses a fault-tolerant, self-aware, low-power, multi-core computer for space missions with thousands of simple cores, achieving speed through concurrency. The proposed machine decides how to achieve concurrency in real time, rather than depending on programmers. The driving features of the system are simple hardware that is modular in the extreme, with no shared memory, and software with significant runtime reorganizing capability. The document describes a mechanism for moving ongoing computations and data that is based on a functional model of execution. Because there is no shared memory, the processor connects to its neighbors through a high-speed data link. Messages are sent to a neighbor switch, which in turn forwards that message on to its neighbor until reaching the intended destination. Except for the neighbor connections, processors are isolated and independent of each other. The processors on the periphery also connect chip-to-chip, thus building up a large processor net. There is no particular topology to the larger net, as a function at each processor allows it to forward a message in the correct direction. Some chip-to-chip connections are not necessarily nearest neighbors, providing short cuts for some of the longer physical distances. The peripheral processors also provide the connections to sensors, actuators, radios, science instruments, and other devices with which the computer system interacts.
Asif, Rameez
2016-01-01
Space division multiplexing (SDM), incorporating multi-core fibers (MCFs), has been demonstrated for effectively maximizing the data capacity in an impending capacity crunch. To achieve high spectral-density through multi-carrier encoding while simultaneously maintaining transmission reach, benefits from inter-core crosstalk (XT) and non-linear compensation must be utilized. In this report, we propose a proof-of-concept unified receiver architecture that jointly compensates optical Kerr effects, intra- and inter-core XT in MCFs. The architecture is analysed in multi-channel 512 Gbit/s dual-carrier DP-16QAM system over 800 km 19-core MCF to validate the digital compensation of inter-core XT. Through this architecture: (a) we efficiently compensates the inter-core XT improving Q-factor by 4.82 dB and (b) achieve a momentous gain in transmission reach, increasing the maximum achievable distance from 480 km to 1208 km, via analytical analysis. Simulation results confirm that inter-core XT distortions are more relentless for cores fabricated around the central axis of cladding. Predominantly, XT induced Q-penalty can be suppressed to be less than 1 dB up-to −11.56 dB of inter-core XT over 800 km MCF, offering flexibility to fabricate dense core structures with same cladding diameter. Moreover, this report outlines the relationship between core pitch and forward-error correction (FEC). PMID:27270381
Li, Xiangyu; Xie, Nijie; Tian, Xinyue
2017-01-01
This paper proposes a scheduling and power management solution for energy harvesting heterogeneous multi-core WSN node SoC such that the system continues to operate perennially and uses the harvested energy efficiently. The solution consists of a heterogeneous multi-core system oriented task scheduling algorithm and a low-complexity dynamic workload scaling and configuration optimization algorithm suitable for light-weight platforms. Moreover, considering the power consumption of most WSN applications have the characteristic of data dependent behavior, we introduce branches handling mechanism into the solution as well. The experimental result shows that the proposed algorithm can operate in real-time on a lightweight embedded processor (MSP430), and that it can make a system do more valuable works and make more than 99.9% use of the power budget. PMID:28208730
Li, Xiangyu; Xie, Nijie; Tian, Xinyue
2017-02-08
This paper proposes a scheduling and power management solution for energy harvesting heterogeneous multi-core WSN node SoC such that the system continues to operate perennially and uses the harvested energy efficiently. The solution consists of a heterogeneous multi-core system oriented task scheduling algorithm and a low-complexity dynamic workload scaling and configuration optimization algorithm suitable for light-weight platforms. Moreover, considering the power consumption of most WSN applications have the characteristic of data dependent behavior, we introduce branches handling mechanism into the solution as well. The experimental result shows that the proposed algorithm can operate in real-time on a lightweight embedded processor (MSP430), and that it can make a system do more valuable works and make more than 99.9% use of the power budget.
Efficiently Scheduling Multi-core Guest Virtual Machines on Multi-core Hosts in Network Simulation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yoginath, Srikanth B; Perumalla, Kalyan S
2011-01-01
Virtual machine (VM)-based simulation is a method used by network simulators to incorporate realistic application behaviors by executing actual VMs as high-fidelity surrogates for simulated end-hosts. A critical requirement in such a method is the simulation time-ordered scheduling and execution of the VMs. Prior approaches such as time dilation are less efficient due to the high degree of multiplexing possible when multiple multi-core VMs are simulated on multi-core host systems. We present a new simulation time-ordered scheduler to efficiently schedule multi-core VMs on multi-core real hosts, with a virtual clock realized on each virtual core. The distinguishing features of ourmore » approach are: (1) customizable granularity of the VM scheduling time unit on the simulation time axis, (2) ability to take arbitrary leaps in virtual time by VMs to maximize the utilization of host (real) cores when guest virtual cores idle, and (3) empirically determinable optimality in the tradeoff between total execution (real) time and time-ordering accuracy levels. Experiments show that it is possible to get nearly perfect time-ordered execution, with a slight cost in total run time, relative to optimized non-simulation VM schedulers. Interestingly, with our time-ordered scheduler, it is also possible to reduce the time-ordering error from over 50% of non-simulation scheduler to less than 1% realized by our scheduler, with almost the same run time efficiency as that of the highly efficient non-simulation VM schedulers.« less
1001 Ways to run AutoDock Vina for virtual screening
NASA Astrophysics Data System (ADS)
Jaghoori, Mohammad Mahdi; Bleijlevens, Boris; Olabarriaga, Silvia D.
2016-03-01
Large-scale computing technologies have enabled high-throughput virtual screening involving thousands to millions of drug candidates. It is not trivial, however, for biochemical scientists to evaluate the technical alternatives and their implications for running such large experiments. Besides experience with the molecular docking tool itself, the scientist needs to learn how to run it on high-performance computing (HPC) infrastructures, and understand the impact of the choices made. Here, we review such considerations for a specific tool, AutoDock Vina, and use experimental data to illustrate the following points: (1) an additional level of parallelization increases virtual screening throughput on a multi-core machine; (2) capturing of the random seed is not enough (though necessary) for reproducibility on heterogeneous distributed computing systems; (3) the overall time spent on the screening of a ligand library can be improved by analysis of factors affecting execution time per ligand, including number of active torsions, heavy atoms and exhaustiveness. We also illustrate differences among four common HPC infrastructures: grid, Hadoop, small cluster and multi-core (virtual machine on the cloud). Our analysis shows that these platforms are suitable for screening experiments of different sizes. These considerations can guide scientists when choosing the best computing platform and set-up for their future large virtual screening experiments.
1001 Ways to run AutoDock Vina for virtual screening.
Jaghoori, Mohammad Mahdi; Bleijlevens, Boris; Olabarriaga, Silvia D
2016-03-01
Large-scale computing technologies have enabled high-throughput virtual screening involving thousands to millions of drug candidates. It is not trivial, however, for biochemical scientists to evaluate the technical alternatives and their implications for running such large experiments. Besides experience with the molecular docking tool itself, the scientist needs to learn how to run it on high-performance computing (HPC) infrastructures, and understand the impact of the choices made. Here, we review such considerations for a specific tool, AutoDock Vina, and use experimental data to illustrate the following points: (1) an additional level of parallelization increases virtual screening throughput on a multi-core machine; (2) capturing of the random seed is not enough (though necessary) for reproducibility on heterogeneous distributed computing systems; (3) the overall time spent on the screening of a ligand library can be improved by analysis of factors affecting execution time per ligand, including number of active torsions, heavy atoms and exhaustiveness. We also illustrate differences among four common HPC infrastructures: grid, Hadoop, small cluster and multi-core (virtual machine on the cloud). Our analysis shows that these platforms are suitable for screening experiments of different sizes. These considerations can guide scientists when choosing the best computing platform and set-up for their future large virtual screening experiments.
Separate Poles Mode for Large-Capacity HVDC System
NASA Astrophysics Data System (ADS)
Zhu, Lin; Gao, Qin
2017-05-01
This paper proposes a novel connection mode, separate poles mode (SPM), for large-capacity HVDC systems. The proposed mode focuses on the core issues of HVDC connection in interconnected power grids and principally aims at increasing effective electric distance between poles, which helps to mitigate the interaction problems between AC system and DC system. Receiving end of bipolar HVDC has been divided into different inverter stations under the mode, and thus significantly alleviates difficulties in power transmission and consumption of receiving-end AC grids. By investigating the changes of multi-feed short-circuit ratio (MISCR), finding that HVDC with SPM shows critical impacts upon itself and other HVDC systems with conventional connection mode, which demonstrates that SPM can make balance between MISCR increase and short-circuit current limit.
Development of an extensible dual-core wireless sensing node for cyber-physical systems
NASA Astrophysics Data System (ADS)
Kane, Michael; Zhu, Dapeng; Hirose, Mitsuhito; Dong, Xinjun; Winter, Benjamin; Häckell, Mortiz; Lynch, Jerome P.; Wang, Yang; Swartz, A.
2014-04-01
The introduction of wireless telemetry into the design of monitoring and control systems has been shown to reduce system costs while simplifying installations. To date, wireless nodes proposed for sensing and actuation in cyberphysical systems have been designed using microcontrollers with one computational pipeline (i.e., single-core microcontrollers). While concurrent code execution can be implemented on single-core microcontrollers, concurrency is emulated by splitting the pipeline's resources to support multiple threads of code execution. For many applications, this approach to multi-threading is acceptable in terms of speed and function. However, some applications such as feedback controls demand deterministic timing of code execution and maximum computational throughput. For these applications, the adoption of multi-core processor architectures represents one effective solution. Multi-core microcontrollers have multiple computational pipelines that can execute embedded code in parallel and can be interrupted independent of one another. In this study, a new wireless platform named Martlet is introduced with a dual-core microcontroller adopted in its design. The dual-core microcontroller design allows Martlet to dedicate one core to standard wireless sensor operations while the other core is reserved for embedded data processing and real-time feedback control law execution. Another distinct feature of Martlet is a standardized hardware interface that allows specialized daughter boards (termed wing boards) to be interfaced to the Martlet baseboard. This extensibility opens opportunity to encapsulate specialized sensing and actuation functions in a wing board without altering the design of Martlet. In addition to describing the design of Martlet, a few example wings are detailed, along with experiments showing the Martlet's ability to monitor and control physical systems such as wind turbines and buildings.
Parallel transformation of K-SVD solar image denoising algorithm
NASA Astrophysics Data System (ADS)
Liang, Youwen; Tian, Yu; Li, Mei
2017-02-01
The images obtained by observing the sun through a large telescope always suffered with noise due to the low SNR. K-SVD denoising algorithm can effectively remove Gauss white noise. Training dictionaries for sparse representations is a time consuming task, due to the large size of the data involved and to the complexity of the training algorithms. In this paper, an OpenMP parallel programming language is proposed to transform the serial algorithm to the parallel version. Data parallelism model is used to transform the algorithm. Not one atom but multiple atoms updated simultaneously is the biggest change. The denoising effect and acceleration performance are tested after completion of the parallel algorithm. Speedup of the program is 13.563 in condition of using 16 cores. This parallel version can fully utilize the multi-core CPU hardware resources, greatly reduce running time and easily to transplant in multi-core platform.
Hydrogen bonding in malonaldehyde: a density functional and reparametrized semiempirical approach
NASA Astrophysics Data System (ADS)
Kovačević, Goran; Hrenar, Tomica; Došlić, Nadja
2003-08-01
Intramolecular proton transfer in malonaldehyde (MA) has been investigated by density functional theory (DFT). The DFT results were used for the construction of a high quality semiempirical potential energy surface with a reparametrized PM3 Hamiltonian. A two-step reparameterization procedure is proposed in which (i) the PM3-MAIS core-core functions for the O-H and H-H interactions were used and a new functional form for the O-O correction function was proposed and (ii) a set of specific reaction parameters (SRP) has been obtained via genetic algorithm optimization. The quality of the reparametrized semiempirical potential energy surfaces was tested by calculating the tunneling splitting of vibrational levels and the anharmonic vibrational frequencies of the system. The applicability to multi-dimensional dynamics in large molecular systems is discussed.
Large-Scale Compute-Intensive Analysis via a Combined In-situ and Co-scheduling Workflow Approach
DOE Office of Scientific and Technical Information (OSTI.GOV)
Messer, Bronson; Sewell, Christopher; Heitmann, Katrin
2015-01-01
Large-scale simulations can produce tens of terabytes of data per analysis cycle, complicating and limiting the efficiency of workflows. Traditionally, outputs are stored on the file system and analyzed in post-processing. With the rapidly increasing size and complexity of simulations, this approach faces an uncertain future. Trending techniques consist of performing the analysis in situ, utilizing the same resources as the simulation, and/or off-loading subsets of the data to a compute-intensive analysis system. We introduce an analysis framework developed for HACC, a cosmological N-body code, that uses both in situ and co-scheduling approaches for handling Petabyte-size outputs. An initial inmore » situ step is used to reduce the amount of data to be analyzed, and to separate out the data-intensive tasks handled off-line. The analysis routines are implemented using the PISTON/VTK-m framework, allowing a single implementation of an algorithm that simultaneously targets a variety of GPU, multi-core, and many-core architectures.« less
Lake deposits record evidence of large post-1505 AD earthquakes in western Nepal
NASA Astrophysics Data System (ADS)
Ghazoui, Z.; Bertrand, S.; Vanneste, K.; Yokoyama, Y.; Van Der Beek, P.; Nomade, J.; Gajurel, A.
2016-12-01
According to historical records, the last large earthquake that ruptured the Main Frontal Thrust (MFT) in western Nepal occurred in 1505 AD. Since then, no evidence of other large earthquakes has been found in historical records or geological archives. In view of the catastrophic consequences to millions of inhabitants of Nepal and northern India, intense efforts currently focus on improving our understanding of past earthquake activity and complement the historical data on Himalayan earthquakes. Here we report a new record, based on earthquake-triggered turbidites in lakes. We use lake sediment records from Lake Rara, western Nepal, to reconstruct the occurrence of seismic events. The sediment cores were studied using a multi-proxy approach combining radiocarbon and 210Pb chronologies, physical properties (X-ray computerized axial tomography scan, Geotek multi-sensor core logger), high-resolution grain size, inorganic geochemistry (major elements by ITRAX XRF core scanning) and bulk organic geochemistry (C, N concentrations and stable isotopes). We identified several sequences of dense and layered fine sand mainly composed of mica, which we interpret as earthquake-triggered turbidites. Our results suggest the presence of a synchronous event between the two lake sites correlated with the well-known 1505 AD earthquake. In addition, our sediment records reveal five earthquake-triggered turbidites younger than the 1505 AD event. By comparison with historical archives, we relate one of those to the 1833 AD MFT rupture. The others may reflect successive ruptures of the Western Nepal Fault System. Our study sheds light on events that have not been recorded in historical chronicles. Those five MMI>7 earthquakes permit addressing the problem of missing slip on the MFT in western Nepal and reevaluating the risk of a large earthquake affecting western Nepal and North India.
NASA Technical Reports Server (NTRS)
Putnam, William M.
2011-01-01
Earth system models like the Goddard Earth Observing System model (GEOS-5) have been pushing the limits of large clusters of multi-core microprocessors, producing breath-taking fidelity in resolving cloud systems at a global scale. GPU computing presents an opportunity for improving the efficiency of these leading edge models. A GPU implementation of GEOS-5 will facilitate the use of cloud-system resolving resolutions in data assimilation and weather prediction, at resolutions near 3.5 km, improving our ability to extract detailed information from high-resolution satellite observations and ultimately produce better weather and climate predictions
NASA Technical Reports Server (NTRS)
Go, B. M.; Righter, K.; Danielson, L.; Pando, K.
2015-01-01
Previous geochemical and geophysical experiments have proposed the presence of a small, metallic lunar core, but its composition is still being investigated. Knowledge of core composition can have a significant effect on understanding the thermal history of the Moon, the conditions surrounding the liquid-solid or liquid-liquid field, and siderophile element partitioning between mantle and core. However, experiments on complex bulk core compositions are very limited. One limitation comes from numerous studies that have only considered two or three element systems such as Fe-S or Fe-C, which do not supply a comprehensive understanding for complex systems such as Fe-Ni-S-Si-C. Recent geophysical data suggests the presence of up to 6% lighter elements. Reassessments of Apollo seismological analyses and samples have also shown the need to acquire more data for a broader range of pressures, temperatures, and compositions. This study considers a complex multi-element system (Fe-Ni-S-C) for a relevant pressure and temperature range to the Moon's core conditions.
Data management integration for biomedical core facilities
NASA Astrophysics Data System (ADS)
Zhang, Guo-Qiang; Szymanski, Jacek; Wilson, David
2007-03-01
We present the design, development, and pilot-deployment experiences of MIMI, a web-based, Multi-modality Multi-Resource Information Integration environment for biomedical core facilities. This is an easily customizable, web-based software tool that integrates scientific and administrative support for a biomedical core facility involving a common set of entities: researchers; projects; equipments and devices; support staff; services; samples and materials; experimental workflow; large and complex data. With this software, one can: register users; manage projects; schedule resources; bill services; perform site-wide search; archive, back-up, and share data. With its customizable, expandable, and scalable characteristics, MIMI not only provides a cost-effective solution to the overarching data management problem of biomedical core facilities unavailable in the market place, but also lays a foundation for data federation to facilitate and support discovery-driven research.
NASA Astrophysics Data System (ADS)
Niwase, Hiroaki; Takada, Naoki; Araki, Hiromitsu; Maeda, Yuki; Fujiwara, Masato; Nakayama, Hirotaka; Kakue, Takashi; Shimobaba, Tomoyoshi; Ito, Tomoyoshi
2016-09-01
Parallel calculations of large-pixel-count computer-generated holograms (CGHs) are suitable for multiple-graphics processing unit (multi-GPU) cluster systems. However, it is not easy for a multi-GPU cluster system to accomplish fast CGH calculations when CGH transfers between PCs are required. In these cases, the CGH transfer between the PCs becomes a bottleneck. Usually, this problem occurs only in multi-GPU cluster systems with a single spatial light modulator. To overcome this problem, we propose a simple method using the InfiniBand network. The computational speed of the proposed method using 13 GPUs (NVIDIA GeForce GTX TITAN X) was more than 3000 times faster than that of a CPU (Intel Core i7 4770) when the number of three-dimensional (3-D) object points exceeded 20,480. In practice, we achieved ˜40 tera floating point operations per second (TFLOPS) when the number of 3-D object points exceeded 40,960. Our proposed method was able to reconstruct a real-time movie of a 3-D object comprising 95,949 points.
NASA Astrophysics Data System (ADS)
Tu, Yiyou; Tong, Zhen; Jiang, Jianqing
2013-04-01
The effect of microstructure on clad/core interactions during the brazing of 4343/3005/4343 multi-layer aluminum brazing sheet was investigated employing differential scanning calorimetry (DSC) and electron back-scattering diffraction (EBSD). The thickness of the melted clad layer gradually decreased during the brazing operation. It could be completely removed isothermally as a result of diffusional solidification at the brazing temperature. During the brazing cycle, the rate of loss of the melt in the brazing sheet, with small equiaxed grains' core layer, was higher than that with the core layer consisting of elongated large grains. The difference in microstructure affected the amount of liquid formed during brazing.
Stamatakis, Alexandros; Ott, Michael
2008-12-27
The continuous accumulation of sequence data, for example, due to novel wet-laboratory techniques such as pyrosequencing, coupled with the increasing popularity of multi-gene phylogenies and emerging multi-core processor architectures that face problems of cache congestion, poses new challenges with respect to the efficient computation of the phylogenetic maximum-likelihood (ML) function. Here, we propose two approaches that can significantly speed up likelihood computations that typically represent over 95 per cent of the computational effort conducted by current ML or Bayesian inference programs. Initially, we present a method and an appropriate data structure to efficiently compute the likelihood score on 'gappy' multi-gene alignments. By 'gappy' we denote sampling-induced gaps owing to missing sequences in individual genes (partitions), i.e. not real alignment gaps. A first proof-of-concept implementation in RAXML indicates that this approach can accelerate inferences on large and gappy alignments by approximately one order of magnitude. Moreover, we present insights and initial performance results on multi-core architectures obtained during the transition from an OpenMP-based to a Pthreads-based fine-grained parallelization of the ML function.
The design of multi-core DSP parallel model based on message passing and multi-level pipeline
NASA Astrophysics Data System (ADS)
Niu, Jingyu; Hu, Jian; He, Wenjing; Meng, Fanrong; Li, Chuanrong
2017-10-01
Currently, the design of embedded signal processing system is often based on a specific application, but this idea is not conducive to the rapid development of signal processing technology. In this paper, a parallel processing model architecture based on multi-core DSP platform is designed, and it is mainly suitable for the complex algorithms which are composed of different modules. This model combines the ideas of multi-level pipeline parallelism and message passing, and summarizes the advantages of the mainstream model of multi-core DSP (the Master-Slave model and the Data Flow model), so that it has better performance. This paper uses three-dimensional image generation algorithm to validate the efficiency of the proposed model by comparing with the effectiveness of the Master-Slave and the Data Flow model.
The parallel algorithm for the 2D discrete wavelet transform
NASA Astrophysics Data System (ADS)
Barina, David; Najman, Pavel; Kleparnik, Petr; Kula, Michal; Zemcik, Pavel
2018-04-01
The discrete wavelet transform can be found at the heart of many image-processing algorithms. Until now, the transform on general-purpose processors (CPUs) was mostly computed using a separable lifting scheme. As the lifting scheme consists of a small number of operations, it is preferred for processing using single-core CPUs. However, considering a parallel processing using multi-core processors, this scheme is inappropriate due to a large number of steps. On such architectures, the number of steps corresponds to the number of points that represent the exchange of data. Consequently, these points often form a performance bottleneck. Our approach appropriately rearranges calculations inside the transform, and thereby reduces the number of steps. In other words, we propose a new scheme that is friendly to parallel environments. When evaluating on multi-core CPUs, we consistently overcome the original lifting scheme. The evaluation was performed on 61-core Intel Xeon Phi and 8-core Intel Xeon processors.
Shared Memory Parallelism for 3D Cartesian Discrete Ordinates Solver
NASA Astrophysics Data System (ADS)
Moustafa, Salli; Dutka-Malen, Ivan; Plagne, Laurent; Ponçot, Angélique; Ramet, Pierre
2014-06-01
This paper describes the design and the performance of DOMINO, a 3D Cartesian SN solver that implements two nested levels of parallelism (multicore+SIMD) on shared memory computation nodes. DOMINO is written in C++, a multi-paradigm programming language that enables the use of powerful and generic parallel programming tools such as Intel TBB and Eigen. These two libraries allow us to combine multi-thread parallelism with vector operations in an efficient and yet portable way. As a result, DOMINO can exploit the full power of modern multi-core processors and is able to tackle very large simulations, that usually require large HPC clusters, using a single computing node. For example, DOMINO solves a 3D full core PWR eigenvalue problem involving 26 energy groups, 288 angular directions (S16), 46 × 106 spatial cells and 1 × 1012 DoFs within 11 hours on a single 32-core SMP node. This represents a sustained performance of 235 GFlops and 40:74% of the SMP node peak performance for the DOMINO sweep implementation. The very high Flops/Watt ratio of DOMINO makes it a very interesting building block for a future many-nodes nuclear simulation tool.
NASA Astrophysics Data System (ADS)
Alpers, Matthias; Brüns, Christian; Pillukat, Alexander
2017-11-01
The evolving needs of the meteorological community concerning the EUMETSAT Polar System follow-on satellite mission (Post-EPS) require the development of a high-performance multi-spectral imaging radiometer. Recognizing these needs, Jena Optronik GmbH proposed an innovative instrument concept, which comprises a high flexibility to adapt to user requirements as a very important feature. Core parameters like ground sampling distance (GSD), number and width of spectral channels, signal-to-noise ratio, polarization control and calibration facilities can be chosen in a wide range without changing the basic instrument configuration. Core item of the METimage instrument is a rotating telescope scanner to cover the large swath width of about 2800 km, which all polar platforms need for global coverage. The de-rotated image facilitates use of in-field spectral channel separation, which allows tailoring individual channel GSD (ground sampling distance) and features like TDI (time delay and integration). State-of-the-art detector arrays and readout electronics can easily be employed. Currently, the German DLR Space Agency, Jena- Optronik GmbH and AIM Infrarot Module GmbH work together implementing core assemblies of METimage: the rotating telescope scanner and the infrared detectors. The METimage instrument phase B study was kicked-off in September 2008. Germany intents to provide METimage as an in-kind contribution of the first METimage flight model to the EUMETSAT Post-EPS Programme.
Fault Tolerance Middleware for a Multi-Core System
NASA Technical Reports Server (NTRS)
Some, Raphael R.; Springer, Paul L.; Zima, Hans P.; James, Mark; Wagner, David A.
2012-01-01
Fault Tolerance Middleware (FTM) provides a framework to run on a dedicated core of a multi-core system and handles detection of single-event upsets (SEUs), and the responses to those SEUs, occurring in an application running on multiple cores of the processor. This software was written expressly for a multi-core system and can support different kinds of fault strategies, such as introspection, algorithm-based fault tolerance (ABFT), and triple modular redundancy (TMR). It focuses on providing fault tolerance for the application code, and represents the first step in a plan to eventually include fault tolerance in message passing and the FTM itself. In the multi-core system, the FTM resides on a single, dedicated core, separate from the cores used by the application. This is done in order to isolate the FTM from application faults and to allow it to swap out any application core for a substitute. The structure of the FTM consists of an interface to a fault tolerant strategy module, a responder module, a fault manager module, an error factory, and an error mapper that determines the severity of the error. In the present reference implementation, the only fault tolerant strategy implemented is introspection. The introspection code waits for an application node to send an error notification to it. It then uses the error factory to create an error object, and at this time, a severity level is assigned to the error. The introspection code uses its built-in knowledge base to generate a recommended response to the error. Responses might include ignoring the error, logging it, rolling back the application to a previously saved checkpoint, swapping in a new node to replace a bad one, or restarting the application. The original error and recommended response are passed to the top-level fault manager module, which invokes the response. The responder module also notifies the introspection module of the generated response. This provides additional information to the introspection module that it can use in generating its next response. For example, if the responder triggers an application rollback and errors are still occurring, the introspection module may decide to recommend an application restart.
Benchmarking NWP Kernels on Multi- and Many-core Processors
NASA Astrophysics Data System (ADS)
Michalakes, J.; Vachharajani, M.
2008-12-01
Increased computing power for weather, climate, and atmospheric science has provided direct benefits for defense, agriculture, the economy, the environment, and public welfare and convenience. Today, very large clusters with many thousands of processors are allowing scientists to move forward with simulations of unprecedented size. But time-critical applications such as real-time forecasting or climate prediction need strong scaling: faster nodes and processors, not more of them. Moreover, the need for good cost- performance has never been greater, both in terms of performance per watt and per dollar. For these reasons, the new generations of multi- and many-core processors being mass produced for commercial IT and "graphical computing" (video games) are being scrutinized for their ability to exploit the abundant fine- grain parallelism in atmospheric models. We present results of our work to date identifying key computational kernels within the dynamics and physics of a large community NWP model, the Weather Research and Forecast (WRF) model. We benchmark and optimize these kernels on several different multi- and many-core processors. The goals are to (1) characterize and model performance of the kernels in terms of computational intensity, data parallelism, memory bandwidth pressure, memory footprint, etc. (2) enumerate and classify effective strategies for coding and optimizing for these new processors, (3) assess difficulties and opportunities for tool or higher-level language support, and (4) establish a continuing set of kernel benchmarks that can be used to measure and compare effectiveness of current and future designs of multi- and many-core processors for weather and climate applications.
NASA Astrophysics Data System (ADS)
Chan, YinThai
2016-03-01
Colloidal semiconductor nanocrystals are ideal fluorophores for clinical diagnostics, therapeutics, and highly sensitive biochip applications due to their high photostability, size-tunable color of emission and flexible surface chemistry. The relatively recent development of core-seeded semiconductor nanorods showed that the presence of a rod-like shell can confer even more advantageous physicochemical properties than their spherical counterparts, such as large multi-photon absorption cross-sections and facet-specific chemistry that can be exploited to deposit secondary nanoparticles. It may be envisaged that these highly fluorescent nanorods can be integrated with large scale integrated (LSI) microfluidic systems that allow miniaturization and integration of multiple biochemical processes in a single device at the nanoliter scale, resulting in a highly sensitive and automated detection platform. In this talk, I will describe a LSI microfluidic device that integrates RNA extraction, reverse transcription to cDNA, amplification and target pull-down to detect histidine decarboxylase (HDC) gene directly from human white blood cells samples. When anisotropic colloidal semiconductor nanorods (NRs) were used as the fluorescent readout, the detection limit was found to be 0.4 ng of total RNA, which was much lower than that obtained using spherical quantum dots (QDs) or organic dyes. This was attributed to the large action cross-section of NRs and their high probability of target capture in a pull-down detection scheme. The combination of large scale integrated microfluidics with highly fluorescent semiconductor NRs may find widespread utility in point-of-care devices and multi-target diagnostics.
Adapting Wave-front Algorithms to Efficiently Utilize Systems with Deep Communication Hierarchies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kerbyson, Darren J.; Lang, Michael; Pakin, Scott
2011-09-30
Large-scale systems increasingly exhibit a differential between intra-chip and inter-chip communication performance especially in hybrid systems using accelerators. Processorcores on the same socket are able to communicate at lower latencies, and with higher bandwidths, than cores on different sockets either within the same node or between nodes. A key challenge is to efficiently use this communication hierarchy and hence optimize performance. We consider here the class of applications that contains wavefront processing. In these applications data can only be processed after their upstream neighbors have been processed. Similar dependencies result between processors in which communication is required to pass boundarymore » data downstream and whose cost is typically impacted by the slowest communication channel in use. In this work we develop a novel hierarchical wave-front approach that reduces the use of slower communications in the hierarchy but at the cost of additional steps in the parallel computation and higher use of on-chip communications. This tradeoff is explored using a performance model. An implementation using the Reverse-acceleration programming model on the petascale Roadrunner system demonstrates a 27% performance improvement at full system-scale on a kernel application. The approach is generally applicable to large-scale multi-core and accelerated systems where a differential in system communication performance exists.« less
Parallel Task Management Library for MARTe
NASA Astrophysics Data System (ADS)
Valcarcel, Daniel F.; Alves, Diogo; Neto, Andre; Reux, Cedric; Carvalho, Bernardo B.; Felton, Robert; Lomas, Peter J.; Sousa, Jorge; Zabeo, Luca
2014-06-01
The Multithreaded Application Real-Time executor (MARTe) is a real-time framework with increasing popularity and support in the thermonuclear fusion community. It allows modular code to run in a multi-threaded environment leveraging on the current multi-core processor (CPU) technology. One application that relies on the MARTe framework is the Joint European Torus (JET) tokamak WAll Load Limiter System (WALLS). It calculates and monitors the temperature on metal tiles and plasma facing components (PFCs) that can melt or flake if their temperature gets too high when exposed to power loads. One of the main time consuming tasks in WALLS is the calculation of thermal diffusion models in real-time. These models tend to be described by very large state-space models thus making them perfect candidates for parallelisation. MARTe's traditional approach for task parallelisation is to split the problem into several Real-Time Threads, each responsible for a self-contained sequential execution of an input-to-output chain. This is usually possible, but it might not always be practical for algorithmic or technical reasons. Also, it might not be easily scalable with an increase in the number of available CPU cores. The WorkLibrary introduces a “GPU-like approach” of splitting work among the available cores of modern CPUs that is (i) straightforward to use in an application, (ii) scalable with the availability of cores and all of this (iii) without rewriting or recompiling the source code. The first part of this article explains the motivation behind the library, its architecture and implementation. The second part presents a real application for WALLS, a parallel version of a large state-space model describing the 2D thermal diffusion on a JET tile.
NASA Astrophysics Data System (ADS)
Dave, Gaurav P.; Sureshkumar, N.; Blessy Trencia Lincy, S. S.
2017-11-01
Current trend in processor manufacturing focuses on multi-core architectures rather than increasing the clock speed for performance improvement. Graphic processors have become as commodity hardware for providing fast co-processing in computer systems. Developments in IoT, social networking web applications, big data created huge demand for data processing activities and such kind of throughput intensive applications inherently contains data level parallelism which is more suited for SIMD architecture based GPU. This paper reviews the architectural aspects of multi/many core processors and graphics processors. Different case studies are taken to compare performance of throughput computing applications using shared memory programming in OpenMP and CUDA API based programming.
NASA Astrophysics Data System (ADS)
Fonseca, R. A.; Vieira, J.; Fiuza, F.; Davidson, A.; Tsung, F. S.; Mori, W. B.; Silva, L. O.
2013-12-01
A new generation of laser wakefield accelerators (LWFA), supported by the extreme accelerating fields generated in the interaction of PW-Class lasers and underdense targets, promises the production of high quality electron beams in short distances for multiple applications. Achieving this goal will rely heavily on numerical modelling to further understand the underlying physics and identify optimal regimes, but large scale modelling of these scenarios is computationally heavy and requires the efficient use of state-of-the-art petascale supercomputing systems. We discuss the main difficulties involved in running these simulations and the new developments implemented in the OSIRIS framework to address these issues, ranging from multi-dimensional dynamic load balancing and hybrid distributed/shared memory parallelism to the vectorization of the PIC algorithm. We present the results of the OASCR Joule Metric program on the issue of large scale modelling of LWFA, demonstrating speedups of over 1 order of magnitude on the same hardware. Finally, scalability to over ˜106 cores and sustained performance over ˜2 P Flops is demonstrated, opening the way for large scale modelling of LWFA scenarios.
Performance implications from sizing a VM on multi-core systems: A Data analytic application s view
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lim, Seung-Hwan; Horey, James L; Begoli, Edmon
In this paper, we present a quantitative performance analysis of data analytics applications running on multi-core virtual machines. Such environments form the core of cloud computing. In addition, data analytics applications, such as Cassandra and Hadoop, are becoming increasingly popular on cloud computing platforms. This convergence necessitates a better understanding of the performance and cost implications of such hybrid systems. For example, the very rst step in hosting applications in virtualized environments, requires the user to con gure the number of virtual processors and the size of memory. To understand performance implications of this step, we benchmarked three Yahoo Cloudmore » Serving Benchmark (YCSB) workloads in a virtualized multi-core environment. Our measurements indicate that the performance of Cassandra for YCSB workloads does not heavily depend on the processing capacity of a system, while the size of the data set is critical to performance relative to allocated memory. We also identi ed a strong relationship between the running time of workloads and various hardware events (last level cache loads, misses, and CPU migrations). From this analysis, we provide several suggestions to improve the performance of data analytics applications running on cloud computing environments.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Seol, Yongkoo, E-mail: Yongkoo.Seol@netl.doe.gov; Choi, Jeong-Hoon; Dai, Sheng
With the increase in the interest of producing natural gas from methane hydrates as well as potential risks of massive hydrate dissociation in the context of global warming, studies have recently shifted from pure hydrate crystals to hydrates in sediments. Such a research focus shift requires a series of innovative laboratory devices that are capable of investigating various properties of hydrate-bearing sediments (HBS). This study introduces a newly developed high pressure testing chamber, i.e., multi-property characterization chamber (MPCC), that allows simultaneous investigation of a series of fundamental properties of HBS, including small-strain stiffness (i.e., P- and S-waves), shear strength, large-strainmore » deformation, stress-volume responses, and permeability. The peripheral coolant circulation system of the MPCC permits stable and accurate temperature control, while the core holder body, made of aluminum, enables X-ray computer tomography scanning to be easily employed for structural and morphological characterization of specimens. Samples of hydrate-bearing sediments are held within a rubber sleeve inside the chamber. The thick sleeve is more durable and versatile than thin membranes while also being much softer than oedometer-type chambers that are incapable of enabling flow tests. Bias introduced by the rubber sleeve during large deformation tests are also calibrated both theoretically and experimentally. This system provides insight into full characterization of hydrate-bearing sediments in the laboratory, as well as pressure core technology in the field.« less
Strong-lensing analysis of A2744 with MUSE and Hubble Frontier Fields images
NASA Astrophysics Data System (ADS)
Mahler, G.; Richard, J.; Clément, B.; Lagattuta, D.; Schmidt, K.; Patrício, V.; Soucail, G.; Bacon, R.; Pello, R.; Bouwens, R.; Maseda, M.; Martinez, J.; Carollo, M.; Inami, H.; Leclercq, F.; Wisotzki, L.
2018-01-01
We present an analysis of Multi Unit Spectroscopic Explorer (MUSE) observations obtained on the massive Frontier Fields (FFs) cluster A2744. This new data set covers the entire multiply imaged region around the cluster core. The combined catalogue consists of 514 spectroscopic redshifts (with 414 new identifications). We use this redshift information to perform a strong-lensing analysis revising multiple images previously found in the deep FF images, and add three new MUSE-detected multiply imaged systems with no obvious Hubble Space Telescope counterpart. The combined strong-lensing constraints include a total of 60 systems producing 188 images altogether, out of which 29 systems and 83 images are spectroscopically confirmed, making A2744 one of the most well-constrained clusters to date. Thanks to the large amount of spectroscopic redshifts, we model the influence of substructures at larger radii, using a parametrization including two cluster-scale components in the cluster core and several group scale in the outskirts. The resulting model accurately reproduces all the spectroscopic multiple systems, reaching an rms of 0.67 arcsec in the image plane. The large number of MUSE spectroscopic redshifts gives us a robust model, which we estimate reduces the systematic uncertainty on the 2D mass distribution by up to ∼2.5 times the statistical uncertainty in the cluster core. In addition, from a combination of the parametrization and the set of constraints, we estimate the relative systematic uncertainty to be up to 9 per cent at 200 kpc.
A GIS Approach to Prioritizing Habitat for Restoration Using Neotropical Migrant Songbird Criteria
NASA Astrophysics Data System (ADS)
Holzmueller, Eric J.; Gaskins, Michael D.; Mangun, Jean C.
2011-07-01
Restoration efforts to increase wildlife habitat quality in agricultural landscapes have limited funding and are typically done on a first come, first serve basis. In order to increase the efficiency of these restoration efforts, a prioritized ranking system is needed to obtain the greatest increase in habitat quality possible for the fewest amount of hectares restored. This project examines the use of a GIS based multi-criteria approach to prioritize lands for reforestation along the Kaskaskia River in Illinois. Loss of forested area and corresponding increase in forest fragmentation has decreased songbird habitat quality across the Midwestern United States. We prioritized areas for reforestation based on nine landscape metrics: available agricultural land, forest cover gaps, edge density, proximity to river, 200 m corridor area, total forest core area, fringe core area, distance to primary core value, and primary core area. The multi-criteria analysis revealed that high priority areas for reforestation were most likely to be close to the riparian corridor and existing large blocks of forest. Analysis of simulated reforestation (0, 0.5, 1.0, 5.0 10.0, 25.0, and 50.0% of highest priority parcels reforested) revealed different responses for multiple landscape metrics used to quantify forest fragmentation following reforestation, but indicated that the study area would get the greatest rate of return on reforestation efforts by reforesting 10.0% of the highest priority areas. This project demonstrates how GIS and a multi-criteria analysis approach can be used to increase the efficiency of restoration projects. This approach should be considered by land managers when attempting to identify the location and quantity of area for restoration within a landscape.
A GIS approach to prioritizing habitat for restoration using neotropical migrant songbird criteria.
Holzmueller, Eric J; Gaskins, Michael D; Mangun, Jean C
2011-07-01
Restoration efforts to increase wildlife habitat quality in agricultural landscapes have limited funding and are typically done on a first come, first serve basis. In order to increase the efficiency of these restoration efforts, a prioritized ranking system is needed to obtain the greatest increase in habitat quality possible for the fewest amount of hectares restored. This project examines the use of a GIS based multi-criteria approach to prioritize lands for reforestation along the Kaskaskia River in Illinois. Loss of forested area and corresponding increase in forest fragmentation has decreased songbird habitat quality across the Midwestern United States. We prioritized areas for reforestation based on nine landscape metrics: available agricultural land, forest cover gaps, edge density, proximity to river, 200 m corridor area, total forest core area, fringe core area, distance to primary core value, and primary core area. The multi-criteria analysis revealed that high priority areas for reforestation were most likely to be close to the riparian corridor and existing large blocks of forest. Analysis of simulated reforestation (0, 0.5, 1.0, 5.0 10.0, 25.0, and 50.0% of highest priority parcels reforested) revealed different responses for multiple landscape metrics used to quantify forest fragmentation following reforestation, but indicated that the study area would get the greatest rate of return on reforestation efforts by reforesting 10.0% of the highest priority areas. This project demonstrates how GIS and a multi-criteria analysis approach can be used to increase the efficiency of restoration projects. This approach should be considered by land managers when attempting to identify the location and quantity of area for restoration within a landscape.
NASA Astrophysics Data System (ADS)
Lawry, B. J.; Encarnacao, A.; Hipp, J. R.; Chang, M.; Young, C. J.
2011-12-01
With the rapid growth of multi-core computing hardware, it is now possible for scientific researchers to run complex, computationally intensive software on affordable, in-house commodity hardware. Multi-core CPUs (Central Processing Unit) and GPUs (Graphics Processing Unit) are now commonplace in desktops and servers. Developers today have access to extremely powerful hardware that enables the execution of software that could previously only be run on expensive, massively-parallel systems. It is no longer cost-prohibitive for an institution to build a parallel computing cluster consisting of commodity multi-core servers. In recent years, our research team has developed a distributed, multi-core computing system and used it to construct global 3D earth models using seismic tomography. Traditionally, computational limitations forced certain assumptions and shortcuts in the calculation of tomographic models; however, with the recent rapid growth in computational hardware including faster CPU's, increased RAM, and the development of multi-core computers, we are now able to perform seismic tomography, 3D ray tracing and seismic event location using distributed parallel algorithms running on commodity hardware, thereby eliminating the need for many of these shortcuts. We describe Node Resource Manager (NRM), a system we developed that leverages the capabilities of a parallel computing cluster. NRM is a software-based parallel computing management framework that works in tandem with the Java Parallel Processing Framework (JPPF, http://www.jppf.org/), a third party library that provides a flexible and innovative way to take advantage of modern multi-core hardware. NRM enables multiple applications to use and share a common set of networked computers, regardless of their hardware platform or operating system. Using NRM, algorithms can be parallelized to run on multiple processing cores of a distributed computing cluster of servers and desktops, which results in a dramatic speedup in execution time. NRM is sufficiently generic to support applications in any domain, as long as the application is parallelizable (i.e., can be subdivided into multiple individual processing tasks). At present, NRM has been effective in decreasing the overall runtime of several algorithms: 1) the generation of a global 3D model of the compressional velocity distribution in the Earth using tomographic inversion, 2) the calculation of the model resolution matrix, model covariance matrix, and travel time uncertainty for the aforementioned velocity model, and 3) the correlation of waveforms with archival data on a massive scale for seismic event detection. Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-AC04-94AL85000.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brogan, C. L.; Hunter, T. R.; Indebetouw, R.
2016-12-01
We present Very Large Array and Atacama Large Millimeter/submillimeter Array imaging of the deeply embedded protostellar cluster NGC 6334I from 5 cm to 1.3 mm at angular resolutions as fine as 0.″17 (220 au). The dominant hot core MM1 is resolved into seven components at 1.3 mm, clustered within a radius of 1000 au. Four of the components have brightness temperatures >200 K, radii ∼300 au, minimum luminosities ∼10{sup 4} L {sub ⊙}, and must be centrally heated. We term this new phenomenon a “hot multi-core.” Two of these objects also exhibit compact free–free emission at longer wavelengths, consistent withmore » a hypercompact H ii region (MM1B) and a jet (MM1D). The spatial kinematics of the water maser emission centered on MM1D are consistent with it being the origin of the high-velocity bipolar molecular outflow seen in CO. The close proximity of MM1B and MM1D (440 au) suggests a proto-binary or a transient bound system. Several components of MM1 exhibit steep millimeter spectral energy distributions indicative of either unusual dust spectral properties or time variability. In addition to resolving MM1 and the other hot core (MM2) into multiple components, we detect five new millimeter and two new centimeter sources. Water masers are detected for the first time toward MM4A, confirming its membership in the protocluster. With a 1.3 mm brightness temperature of 97 K coupled with a lack of thermal molecular line emission, MM4A appears to be a highly optically thick 240 L {sub ⊙} dust core, possibly tracing a transient stage of massive protostellar evolution. The nature of the strongest water maser source CM2 remains unclear due to its combination of non-thermal radio continuum and lack of dust emission.« less
NASA Astrophysics Data System (ADS)
Brogan, C. L.; Hunter, T. R.; Cyganowski, C. J.; Chandler, C. J.; Friesen, R.; Indebetouw, R.
2016-12-01
We present Very Large Array and Atacama Large Millimeter/submillimeter Array imaging of the deeply embedded protostellar cluster NGC 6334I from 5 cm to 1.3 mm at angular resolutions as fine as 0.″17 (220 au). The dominant hot core MM1 is resolved into seven components at 1.3 mm, clustered within a radius of 1000 au. Four of the components have brightness temperatures >200 K, radii ˜300 au, minimum luminosities ˜104 L ⊙, and must be centrally heated. We term this new phenomenon a “hot multi-core.” Two of these objects also exhibit compact free-free emission at longer wavelengths, consistent with a hypercompact H II region (MM1B) and a jet (MM1D). The spatial kinematics of the water maser emission centered on MM1D are consistent with it being the origin of the high-velocity bipolar molecular outflow seen in CO. The close proximity of MM1B and MM1D (440 au) suggests a proto-binary or a transient bound system. Several components of MM1 exhibit steep millimeter spectral energy distributions indicative of either unusual dust spectral properties or time variability. In addition to resolving MM1 and the other hot core (MM2) into multiple components, we detect five new millimeter and two new centimeter sources. Water masers are detected for the first time toward MM4A, confirming its membership in the protocluster. With a 1.3 mm brightness temperature of 97 K coupled with a lack of thermal molecular line emission, MM4A appears to be a highly optically thick 240 L ⊙ dust core, possibly tracing a transient stage of massive protostellar evolution. The nature of the strongest water maser source CM2 remains unclear due to its combination of non-thermal radio continuum and lack of dust emission.
Polytopol computing for multi-core and distributed systems
NASA Astrophysics Data System (ADS)
Spaanenburg, Henk; Spaanenburg, Lambert; Ranefors, Johan
2009-05-01
Multi-core computing provides new challenges to software engineering. The paper addresses such issues in the general setting of polytopol computing, that takes multi-core problems in such widely differing areas as ambient intelligence sensor networks and cloud computing into account. It argues that the essence lies in a suitable allocation of free moving tasks. Where hardware is ubiquitous and pervasive, the network is virtualized into a connection of software snippets judiciously injected to such hardware that a system function looks as one again. The concept of polytopol computing provides a further formalization in terms of the partitioning of labor between collector and sensor nodes. Collectors provide functions such as a knowledge integrator, awareness collector, situation displayer/reporter, communicator of clues and an inquiry-interface provider. Sensors provide functions such as anomaly detection (only communicating singularities, not continuous observation), they are generally powered or self-powered, amorphous (not on a grid) with generation-and-attrition, field re-programmable, and sensor plug-and-play-able. Together the collector and the sensor are part of the skeleton injector mechanism, added to every node, and give the network the ability to organize itself into some of many topologies. Finally we will discuss a number of applications and indicate how a multi-core architecture supports the security aspects of the skeleton injector.
Accelerating Large Scale Image Analyses on Parallel, CPU-GPU Equipped Systems
Teodoro, George; Kurc, Tahsin M.; Pan, Tony; Cooper, Lee A.D.; Kong, Jun; Widener, Patrick; Saltz, Joel H.
2014-01-01
The past decade has witnessed a major paradigm shift in high performance computing with the introduction of accelerators as general purpose processors. These computing devices make available very high parallel computing power at low cost and power consumption, transforming current high performance platforms into heterogeneous CPU-GPU equipped systems. Although the theoretical performance achieved by these hybrid systems is impressive, taking practical advantage of this computing power remains a very challenging problem. Most applications are still deployed to either GPU or CPU, leaving the other resource under- or un-utilized. In this paper, we propose, implement, and evaluate a performance aware scheduling technique along with optimizations to make efficient collaborative use of CPUs and GPUs on a parallel system. In the context of feature computations in large scale image analysis applications, our evaluations show that intelligently co-scheduling CPUs and GPUs can significantly improve performance over GPU-only or multi-core CPU-only approaches. PMID:25419545
All-fiber orbital angular momentum mode generation and transmission system
NASA Astrophysics Data System (ADS)
Heng, Xiaobo; Gan, Jiulin; Zhang, Zhishen; Qian, Qi; Xu, Shanhui; Yang, Zhongmin
2017-11-01
We proposed and demonstrated an all-fiber system for generating and transmitting orbital angular momentum (OAM) mode light. A specially designed multi-core fiber (MCF) was used to endow with guide modes different phase change and two tapered transition regions were used for providing low-loss interfaces between different fiber structures. By arranging the refractive index distribution among the multi-cores and controlling the length of MCF, which essentially change the phase difference between the neighboring cores, OAM modes with different topological charge l can be generated selectively. Through two tapered transition regions, the non-OAM mode light can be effectively injected into the MCF and the generated OAM mode light can be easily launched into OAM mode supporting fiber for long distance and high purity transmission. Such an all-fiber OAM mode generation and transmission system owns the merits of flexibility, compactness, portability, and would have practical application value in OAM optical fiber communication systems.
Recommendations for the design and the installation of large laser scanning microscopy systems
NASA Astrophysics Data System (ADS)
Helm, P. Johannes
2012-03-01
Laser Scanning Microscopy (LSM) has since the inventions of the Confocal Scanning Laser Microscope (CLSM) and the Multi Photon Laser Scanning Microscope (MPLSM) developed into an essential tool in contemporary life science and material science. The market provides an increasing number of turn-key and hands-off commercial LSM systems, un-problematic to purchase, set up and integrate even into minor research groups. However, the successful definition, financing, acquisition, installation and effective use of one or more large laser scanning microscopy systems, possibly of core facility character, often requires major efforts by senior staff members of large academic or industrial units. Here, a set of recommendations is presented, which are helpful during the process of establishing large systems for confocal or non-linear laser scanning microscopy as an effective operational resource in the scientific or industrial production process. Besides the description of technical difficulties and possible pitfalls, the article also illuminates some seemingly "less scientific" processes, i.e. the definition of specific laboratory demands, advertisement of the intention to purchase one or more large systems, evaluation of quotations, establishment of contracts and preparation of the local environment and laboratory infrastructure.
MetAlign 3.0: performance enhancement by efficient use of advances in computer hardware.
Lommen, Arjen; Kools, Harrie J
2012-08-01
A new, multi-threaded version of the GC-MS and LC-MS data processing software, metAlign, has been developed which is able to utilize multiple cores on one PC. This new version was tested using three different multi-core PCs with different operating systems. The performance of noise reduction, baseline correction and peak-picking was 8-19 fold faster compared to the previous version on a single core machine from 2008. The alignment was 5-10 fold faster. Factors influencing the performance enhancement are discussed. Our observations show that performance scales with the increase in processor core numbers we currently see in consumer PC hardware development.
Accelerating 3D Elastic Wave Equations on Knights Landing based Intel Xeon Phi processors
NASA Astrophysics Data System (ADS)
Sourouri, Mohammed; Birger Raknes, Espen
2017-04-01
In advanced imaging methods like reverse-time migration (RTM) and full waveform inversion (FWI) the elastic wave equation (EWE) is numerically solved many times to create the seismic image or the elastic parameter model update. Thus, it is essential to optimize the solution time for solving the EWE as this will have a major impact on the total computational cost in running RTM or FWI. From a computational point of view applications implementing EWEs are associated with two major challenges. The first challenge is the amount of memory-bound computations involved, while the second challenge is the execution of such computations over very large datasets. So far, multi-core processors have not been able to tackle these two challenges, which eventually led to the adoption of accelerators such as Graphics Processing Units (GPUs). Compared to conventional CPUs, GPUs are densely populated with many floating-point units and fast memory, a type of architecture that has proven to map well to many scientific computations. Despite its architectural advantages, full-scale adoption of accelerators has yet to materialize. First, accelerators require a significant programming effort imposed by programming models such as CUDA or OpenCL. Second, accelerators come with a limited amount of memory, which also require explicit data transfers between the CPU and the accelerator over the slow PCI bus. The second generation of the Xeon Phi processor based on the Knights Landing (KNL) architecture, promises the computational capabilities of an accelerator but require the same programming effort as traditional multi-core processors. The high computational performance is realized through many integrated cores (number of cores and tiles and memory varies with the model) organized in tiles that are connected via a 2D mesh based interconnect. In contrary to accelerators, KNL is a self-hosted system, meaning explicit data transfers over the PCI bus are no longer required. However, like most accelerators, KNL sports a memory subsystem consisting of low-level caches and 16GB of high-bandwidth MCDRAM memory. For capacity computing, up to 400GB of conventional DDR4 memory is provided. Such a strict hierarchical memory layout means that data locality is imperative if the true potential of this product is to be harnessed. In this work, we study a series of optimizations specifically targeting KNL for our EWE based application to reduce the time-to-solution time for the following 3D model sizes in grid points: 1283, 2563 and 5123. We compare the results with an optimized version for multi-core CPUs running on a dual-socket Xeon E5 2680v3 system using OpenMP. Our initial naive implementation on the KNL is roughly 20% faster than the multi-core version, but by using only one thread per core and careful memory placement using the memkind library, we could achieve higher speedups. Additionally, by using the MCDRAM as cache for problem sizes that are smaller than 16 GB further performance improvements were unlocked. Depending on the problem size, our overall results indicate that the KNL based system is approximately 2.2x faster than the 24-core Xeon E5 2680v3 system, with only modest changes to the code.
Developing core-shell upconversion nanoparticles for optical encoding
NASA Astrophysics Data System (ADS)
Huang, Kai
Lanthanide-doped upconversion nanoparticles (UCNPs) are an emerging class of luminescent materials that emit UV or visible light under near infra-red (NIR) excitations, thereby possessing a large anti-Stokes shift property. Also considering their sharp emission bands, excellent photo- and chemical stability, and almost zero auto-fluorescence of their NIR excitation, UCNPs are advantageous for optical encoding. Fabricating core-shell structured UCNPs provides a promising strategy to tune and enhance their upconverting luminescence. However, the energy transfer between core and shell had been rarely studied. Moreover, this strategy had been limited by the difficulty of coating thick shells onto the large cores of UCNPs. To overcome these constraints, the overall aim of this project is to study the inter-layers energy transfer in core-shell UCNPs and to develop an approach for coating thicker shell onto the core UCNPs, in order to fabricate UCNPs with enhanced and tunable luminescence for optical encoding. The strategy for encapsulating UCNPs into hydrogel droplet to fabricate multi-color bead barcodes has also been developed. Firstly, to study the inter-layers energy transfer between the core and shell of coreshell UCNPs, the activator and sensitizer ions were separately doped in the core or shell by fabricating NaYF4:Er NaYF4:Yb and NaYF4:Yb NaYF4:Er UCNPs. This eliminated the intra-layer energy transfer, resulting in a luminescence that is solely based on the energy transfer between layers, which facilitated the study of inter-layers energy transfer. The results demonstrated that the NaYF4:Yb NaYF4:Er structure, with sensitizer ions doped in the core, was preferable because of the strong luminescence, through minimizing the cross relaxations between Er3+ and Yb3+ and the surface quenching. Based on these information, a strategy of enhancing and tuning upconversion luminescence of core-shell UCNPs by accumulating sensitizer in the core has been developed. Next, a strategy of coating a thick shell by lutetium doping has been developed. With a smaller ion radius compared to Y3+, when Lu3+ partially replace Y3+ in the NaYF4 UCNPs during nanoparticle synthesis, nucleation process is suppressed and the growth process is promoted, which are favorable for increasing the nanoparticle size and coating a thicker shell onto the core UCNPs. Through the rational doping of Lu3+, core UCNPs with bigger sizes and enhanced luminescence were produced. Using NaLuF4 as the shell material, shells with tremendous thickness were coated onto core UCNPs, with the shell/core ratio of up to 10:1. This led to the fabrication of multi-color UCNPs with well-designed core-shell structures with multiple layers and controllable thicknesses. Finally, a strategy of encapsulating these UCNPs to produce optically encoded micro-beads through high-throughput microfluidics has been developed. The hydrophobic UCNPs were first modified with Pluronic F127 to render them hydrophilic and uniformly distributed in the poly (ethylene glycol) diacrylate (PEGDA) hydrogel precursor. Droplets of the hydrogel precursor were formed in a microfluidic device and cross-linked into micro-beads under UV irradiation. Through encapsulation of multi-color UCNPs and by controlling their ratio, optically encoded multi-color micro-beads have been easily fabricated. These multi-color UCNPs and micro-bead barcodes have great potential for use in multiplexed bioimaging and detection.
Scalable and Power Efficient Data Analytics for Hybrid Exascale Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Choudhary, Alok; Samatova, Nagiza; Wu, Kesheng
This project developed a generic and optimized set of core data analytics functions. These functions organically consolidate a broad constellation of high performance analytical pipelines. As the architectures of emerging HPC systems become inherently heterogeneous, there is a need to design algorithms for data analysis kernels accelerated on hybrid multi-node, multi-core HPC architectures comprised of a mix of CPUs, GPUs, and SSDs. Furthermore, the power-aware trend drives the advances in our performance-energy tradeoff analysis framework which enables our data analysis kernels algorithms and software to be parameterized so that users can choose the right power-performance optimizations.
PyNEST: A Convenient Interface to the NEST Simulator.
Eppler, Jochen Martin; Helias, Moritz; Muller, Eilif; Diesmann, Markus; Gewaltig, Marc-Oliver
2008-01-01
The neural simulation tool NEST (http://www.nest-initiative.org) is a simulator for heterogeneous networks of point neurons or neurons with a small number of compartments. It aims at simulations of large neural systems with more than 10(4) neurons and 10(7) to 10(9) synapses. NEST is implemented in C++ and can be used on a large range of architectures from single-core laptops over multi-core desktop computers to super-computers with thousands of processor cores. Python (http://www.python.org) is a modern programming language that has recently received considerable attention in Computational Neuroscience. Python is easy to learn and has many extension modules for scientific computing (e.g. http://www.scipy.org). In this contribution we describe PyNEST, the new user interface to NEST. PyNEST combines NEST's efficient simulation kernel with the simplicity and flexibility of Python. Compared to NEST's native simulation language SLI, PyNEST makes it easier to set up simulations, generate stimuli, and analyze simulation results. We describe how PyNEST connects NEST and Python and how it is implemented. With a number of examples, we illustrate how it is used.
PyNEST: A Convenient Interface to the NEST Simulator
Eppler, Jochen Martin; Helias, Moritz; Muller, Eilif; Diesmann, Markus; Gewaltig, Marc-Oliver
2008-01-01
The neural simulation tool NEST (http://www.nest-initiative.org) is a simulator for heterogeneous networks of point neurons or neurons with a small number of compartments. It aims at simulations of large neural systems with more than 104 neurons and 107 to 109 synapses. NEST is implemented in C++ and can be used on a large range of architectures from single-core laptops over multi-core desktop computers to super-computers with thousands of processor cores. Python (http://www.python.org) is a modern programming language that has recently received considerable attention in Computational Neuroscience. Python is easy to learn and has many extension modules for scientific computing (e.g. http://www.scipy.org). In this contribution we describe PyNEST, the new user interface to NEST. PyNEST combines NEST's efficient simulation kernel with the simplicity and flexibility of Python. Compared to NEST's native simulation language SLI, PyNEST makes it easier to set up simulations, generate stimuli, and analyze simulation results. We describe how PyNEST connects NEST and Python and how it is implemented. With a number of examples, we illustrate how it is used. PMID:19198667
DOE Office of Scientific and Technical Information (OSTI.GOV)
Harrison, Cyrus; Larsen, Matt; Brugger, Eric
Strawman is a system designed to explore the in situ visualization and analysis needs of simulation code teams running multi-physics calculations on many-core HPC architectures. It porvides rendering pipelines that can leverage both many-core CPUs and GPUs to render images of simulation meshes.
Using OpenMP vs. Threading Building Blocks for Medical Imaging on Multi-cores
NASA Astrophysics Data System (ADS)
Kegel, Philipp; Schellmann, Maraike; Gorlatch, Sergei
We compare two parallel programming approaches for multi-core systems: the well-known OpenMP and the recently introduced Threading Building Blocks (TBB) library by Intel®. The comparison is made using the parallelization of a real-world numerical algorithm for medical imaging. We develop several parallel implementations, and compare them w.r.t. programming effort, programming style and abstraction, and runtime performance. We show that TBB requires a considerable program re-design, whereas with OpenMP simple compiler directives are sufficient. While TBB appears to be less appropriate for parallelizing existing implementations, it fosters a good programming style and higher abstraction level for newly developed parallel programs. Our experimental measurements on a dual quad-core system demonstrate that OpenMP slightly outperforms TBB in our implementation.
Improvement of optical damage in specialty fiber at 266 nm wavelength
NASA Astrophysics Data System (ADS)
Tobisch, T.; Ohlmeyer, H.; Zimmermann, H.; Prein, S.; Kirchhof, J.; Unger, S.; Belz, M.; Klein, K.-F.
2014-02-01
Improved multimode UV-fibers with core diameters ranging from 70 to 600 μm diameter have been manufactured based on novel preform modifications and fiber processing techniques. Only E'-centers at 214 nm and NBOHC at 260 nm are generated in these fibers. A new generation of inexpensive laser-systems have entered the market and generated a multitude of new and attractive applications in the bio-life science, chemical and material processing field. However, for example pulsed 355 nm Nd:YAG lasers generate significant UV-damages in commercially available fibers. For lower wavelengths, no results on suitable multi-mode or low-mode fibers with high UV resistance at 266 nm wavelength (pulsed 4th harmonic Nd:YAG laser) have been published. In this report, double-clad fibers with 70 μm or 100 μm core diameter and a large claddingto- core ratio will be recommended. Laser-induced UV-damages will be compared between these new fiber type and traditional UV fibers with similar core sizes. Finally, experimental results will be cross compared against broadband cw deuterium lamp damage standards.
Towards multi-decadal to multi-millennial ice core records from coastal west Greenland ice caps
NASA Astrophysics Data System (ADS)
Das, Sarah B.; Osman, Matthew B.; Trusel, Luke D.; McConnell, Joseph R.; Smith, Ben E.; Evans, Matthew J.; Frey, Karen E.; Arienzo, Monica; Chellman, Nathan
2017-04-01
The Arctic region, and Greenland in particular, is undergoing dramatic change as characterized by atmospheric warming, decreasing sea ice, shifting ocean circulation patterns, and rapid ice sheet mass loss, but longer records are needed to put these changes into context. Ice core records from the Greenland ice sheet have yielded invaluable insight into past climate change both regionally and globally, and provided important constraints on past surface mass balance more directly, but these ice cores are most often from the interior ice sheet accumulation zone, at high altitude and hundreds of kilometers from the coast. Coastal ice caps, situated around the margins of Greenland, have the potential to provide novel high-resolution records of local and regional maritime climate and sea surface conditions, as well as contemporaneous glaciological changes (such as accumulation and surface melt history). But obtaining these records is extremely challenging. Most of these ice caps are unexplored, and thus their thickness, age, stratigraphy, and utility as sites of new and unique paleoclimate records is largely unknown. Access is severely limited due to their high altitude, steep relief, small surface area, and inclement weather. Furthermore, their relatively low elevation and marine moderated climate can contribute to significant surface melting and degradation of the ice stratigraphy. We recently targeted areas near the Disko Bay region of central west Greenland where maritime ice caps are prevalent but unsampled, as potential sites for new multi-decadal to multi-millennial ice core records. In 2014 & 2015 we identified two promising ice caps, one on Disko Island (1250 m. asl) and one on Nuussuaq Peninsula (1980 m. asl) based on airborne and ground-based geophysical observations and physical and glaciochemical stratigraphy from shallow firn cores. In spring 2015 we collected ice cores at both sites using the Badger-Eclipse electromechanical drill, transported by a medley of small fixed wing and helicopter aircraft, and working out of small tent camps. On Disko Island, despite high accumulation rates and ice thickness of 250 meters, drilling was halted twice due to the encounter of liquid water at depths ranging from 18-20 meters, limiting the depth of the final core to 21 m, providing a multi-decadal record (1980-2015.) On Nuussuaq Peninsula, we collected a 138 m ice core, almost to bedrock, representing a 2500 year record. The ice cores were subsequently analyzed using a continuous flow analysis system (CFA). Age-depth profiles and accumulation histories were determined by combining annual layer counting and an ice flow thinning model, both constrained by glaciochemical tie points to other well-dated Greenland ice core records (e.g. volcanic horizons and continuous heavy metal records). Here we will briefly provide an overview of the project and the new sites, and the novel dating methodology, and describe the latest stratigraphic, isotopic and glaciochemical results. We will also provide a particular focus on new regional climatological insight gained from our records during three climatically sensitive time periods: the late 20th & early 21st centuries; the Little Ice Age; and the Medieval Climate Anomaly.
Spectral efficiency in crosstalk-impaired multi-core fiber links
NASA Astrophysics Data System (ADS)
Luís, Ruben S.; Puttnam, Benjamin J.; Rademacher, Georg; Klaus, Werner; Agrell, Erik; Awaji, Yoshinari; Wada, Naoya
2018-02-01
We review the latest advances on ultra-high throughput transmission using crosstalk-limited single-mode multicore fibers and compare these with the theoretical spectral efficiency of such systems. We relate the crosstalkimposed spectral efficiency limits with fiber parameters, such as core diameter, core pitch, and trench design. Furthermore, we investigate the potential of techniques such as direction interleaving and high-order MIMO to improve the throughput or reach of these systems when using various modulation formats.
NASA Astrophysics Data System (ADS)
Homma, Yuto; Moriwaki, Hiroyuki; Ohki, Shigeo; Ikeda, Kazumi
2014-06-01
This paper deals with verification of three dimensional triangular prismatic discrete ordinates transport calculation code ENSEMBLE-TRIZ by comparison with multi-group Monte Carlo calculation code GMVP in a large fast breeder reactor. The reactor is a 750 MWe electric power sodium cooled reactor. Nuclear characteristics are calculated at beginning of cycle of an initial core and at beginning and end of cycle of equilibrium core. According to the calculations, the differences between the two methodologies are smaller than 0.0002 Δk in the multi-plication factor, relatively about 1% in the control rod reactivity, and 1% in the sodium void reactivity.
Focal ratio degradation in lightly fused hexabundles
NASA Astrophysics Data System (ADS)
Bryant, J. J.; Bland-Hawthorn, J.; Fogarty, L. M. R.; Lawrence, J. S.; Croom, S. M.
2014-02-01
We are now moving into an era where multi-object wide-field surveys, which traditionally use single fibres to observe many targets simultaneously, can exploit compact integral field units (IFUs) in place of single fibres. Current multi-object integral field instruments such as Sydney-AAO Multi-object Integral field spectrograph have driven the development of new imaging fibre bundles (hexabundles) for multi-object spectrographs. We have characterized the performance of hexabundles with different cladding thicknesses and compared them to that of the same type of bare fibre, across the range of fill fractions and input f-ratios likely in an IFU instrument. Hexabundles with 7-cores and 61-cores were tested for focal ratio degradation (FRD), throughput and cross-talk when fed with inputs from F/3.4 to >F/8. The five 7-core bundles have cladding thickness ranging from 1 to 8 μm, and the 61-core bundles have 5 μm cladding. As expected, the FRD improves as the input focal ratio decreases. We find that the FRD and throughput of the cores in the hexabundles match the performance of single fibres of the same material at low input f-ratios. The performance results presented can be used to set a limit on the f-ratio of a system based on the maximum loss allowable for a planned instrument. Our results confirm that hexabundles are a successful alternative for fibre imaging devices for multi-object spectroscopy on wide-field telescopes and have prompted further development of hexabundle designs with hexagonal packing and square cores.
NASA Technical Reports Server (NTRS)
Hinchey, Michael G. (Inventor); Rash, James L. (Inventor); Pena, Joaquin (Inventor)
2011-01-01
Systems, methods and apparatus are provided through which an evolutionary system is managed and viewed as a software product line. In some embodiments, the core architecture is a relatively unchanging part of the system, and each version of the system is viewed as a product from the product line. Each software product is generated from the core architecture with some agent-based additions. The result may be a multi-agent system software product line.
Heath, Caroline; Posner, Mareike G; Aass, Hans C; Upadhyay, Abhishek; Scott, David J; Hough, David W; Danson, Michael J
2007-10-01
The aerobic archaea possess four closely spaced, adjacent genes that encode proteins showing significant sequence identities with the bacterial and eukaryal components comprising the 2-oxoacid dehydrogenase multi-enzyme complexes. However, catalytic activities of such complexes have never been detected in the archaea, although 2-oxoacid ferredoxin oxidoreductases that catalyze the equivalent metabolic reactions are present. In the current paper, we clone and express the four genes from the thermophilic archaeon, Thermoplasma acidophilum, and demonstrate that the recombinant enzymes are active and assemble into a large (M(r) = 5 x 10(6)) multi-enzyme complex. The post-translational incorporation of lipoic acid into the transacylase component of the complex is demonstrated, as is the assembly of this enzyme into a 24-mer core to which the other components bind to give the functional multi-enzyme system. This assembled complex is shown to catalyze the oxidative decarboxylation of branched-chain 2-oxoacids and pyruvate to their corresponding acyl-CoA derivatives. Our data constitute the first proof that the archaea possess a functional 2-oxoacid dehydrogenase complex.
Characterizing the galaxy populations within different environments in the RCS2319 supercluster
NASA Astrophysics Data System (ADS)
Delahaye, Anna; Webb, Tracy
We present the results of a multi-wavelength photometric study of the high redshift supercluster RCS2319+00. RCS2319+00 is a high-redshift (z ~ 0.9) supercluster comprising three spectroscopically confrmed cluster cores discovered in the Red Sequence Cluster Survey (RCS) (Gladders & Yee 2005). Core proximities and merger rates estimate coalescence into a 1015 M ⊙ cluster by z ~ 0.5 (Gilbank et al. 2008). Spectroscopic studies of the system have revealed over 300 supercluster members located in the cores and several infalling groups (Faloon et al. 2013). RCS2319 presents a diverse range of dynamical systems and densities making it an ideal laboratory in which to study the effects of environment on galaxy properties. Imaging in optical and near infrared (griz' from MegaCam, JK s from WIRCam, both at CFHT), as well as 3.6 μm and 4.5μm from IRAC have enabled the assembly of a large photometric catalogue. Coupled with an extensive spectroscopic survey (Faloon et al. 2013) providing nearly 2400 redshifts across the field, photometric redshifts were determined using the template fitting code EAZY (Brammer et al. 2008). Nearly 80 000 photometric redshifts were measured providing a sample of nearly 3000 cluster members. To investigate effects of global environment, analysis was done utilizing a friend-of-friends group finding algorithm identifying several large and small infalling groups along with the three cluster cores. The cores are found to be dominated by massive, red galaxies and the field galaxies are populated by low mass, blue galaxies, as is the case in the local universe. Interestingly, the large groups exhibit intermediate properties between field and core populations, suggesting possible pre-processing as they are being accreted into the core halos. Relative fifth-nearest neighbour overdensity, log(1+δ5), is used as a proxy for local environment to investigate environmental dependence on galaxy colour. While there is an overall dependence of colour on local density, when controlled for stellar mass the dependence largely disappears. Indeed, galaxy mass is the dominant factor in determining colour, with local density a secondary effect only noticeable in lower mass galaxies at the 3 σ level for both colour and red fraction. RCS2319+00 presents a rare opportunity to probe many different densities and environments all located within the same object. We're able to investigate how galaxy evolution is affected by the environment, from field galaxies to infalling to groups to dense cluster cores, as well as the different density regions within each environment.
Multicore Architectures for Multiple Independent Levels of Security Applications
2012-09-01
to bolster the MILS effort. However, current MILS operating systems are not designed for multi-core platforms. They do not have the hardware support...current MILS operating systems are not designed for multi‐core platforms. They do not have the hardware support to ensure that the separation...the availability of information at different security classification levels while increasing the overall security of the computing system . Due to the
Plasmon-Enhanced Multi-Carrier Photocatalysis.
Shaik, Firdoz; Peer, Imanuel; Jain, Prashant K; Amirav, Lilac
2018-06-22
Conversion of solar energy into liquid fuel often relies on multi-electron redox processes that include highly reactive intermediates, with back reaction routes that hinder the overall efficiency of the process. Here we reveal that these undesirable reaction pathways can be minimized, rendering the photocatalytic reactions more efficient, when charge carriers are harvested from a multi-excitonic state of a semiconductor photocatalyst. A plasmonic antenna, comprised of Au nanoprisms, was employed to accomplish feasible levels of multiple carrier excitations in semiconductor nanocrystal-based photocatalytic systems (CdSe@CdS core-shell quantum dots and CdSe@CdS seeded nanorods). The antenna's near-field amplifies the otherwise inherently weak two-photon absorption in the semiconductor. The two-electron photoreduction of Pt and Pd metal precursors served as model reactions. In the presence of the plasmonic antenna, these photocatalyzed two-electron reactions exhibited enhanced yields and kinetics. This work uniquely relies on a non-linear enhancement that has potential for large amplification of photocatalytic activity in the presence of a plasmonic near-field.
From Theoretical Physics to Cancer Diagnostics
NASA Astrophysics Data System (ADS)
Tang, Cha-Mei
The core technologies of our company are techniques in high aspect ratio microfabrication. We are currently applying them to the development of x-ray phase contrast imaging systems and blood test for early detection of solid tumors. Good research does not always mean marketable products. Large market size is critical for commercialization. Excellent and accurate science is important, but not sufficient. We are learning that commercialization requires different expertise, experienced advisors, well known board of directors, partnerships, and investments. Our technologies have the potential of reaching multi-billion dollar sales, and we hope we will achieve that goal.
Dietrich, Philipp-Immanuel; Harris, Robert J; Blaicher, Matthias; Corrigan, Mark K; Morris, Tim M; Freude, Wolfgang; Quirrenbach, Andreas; Koos, Christian
2017-07-24
Coupling of light into multi-core fibers (MCF) for spatially resolved spectroscopy is of great importance to astronomical instrumentation. To achieve high coupling efficiencies along with fill-fractions close to unity, micro-optical elements are required to concentrate the incoming light to the individual cores of the MCF. In this paper we demonstrate facet-attached lens arrays (LA) fabricated by two-photon polymerization. The LA provide close to 100% fill-fraction along with efficiencies of up to 73% (down to 1.4 dB loss) for coupling of light from free space into an MCF core. We show the viability of the concept for astrophotonic applications by integrating an MCF-LA assembly in an adaptive-optics test bed and by assessing its performance as a tip/tilt sensor.
Wang, Kang; Gu, Huaxi; Yang, Yintang; Wang, Kun
2015-08-10
With the number of cores increasing, there is an emerging need for a high-bandwidth low-latency interconnection network, serving core-to-memory communication. In this paper, aiming at the goal of simultaneous access to multi-rank memory, we propose an optical interconnection network for core-to-memory communication. In the proposed network, the wavelength usage is delicately arranged so that cores can communicate with different ranks at the same time and broadcast for flow control can be achieved. A distributed memory controller architecture that works in a pipeline mode is also designed for efficient optical communication and transaction address processes. The scaling method and wavelength assignment for the proposed network are investigated. Compared with traditional electronic bus-based core-to-memory communication, the simulation results based on the PARSEC benchmark show that the bandwidth enhancement and latency reduction are apparent.
Thin disk lasers: history and prospects
NASA Astrophysics Data System (ADS)
Speiser, Jochen
2016-04-01
During the early 1990s, collaboration between the German Aerospace Center and the University of Stuttgart started to work on the Thin Disk concept. The core idea behind the thin disk design is the use of a thin, disk-shaped active medium that is cooled through one of the flat faces of the disk. This ensures a large surface-to-volume ratio and therefore provides very efficient thermal management. Today, the thin disk concept is used in various commercial lasers - ranging from compact, efficient low power systems to multi-kW lasers, including cw lasers and also pulsed (femtosecond to nanosecond) oscillators and amplifiers. The whole development of the Thin Disk laser was and will be accompanied by numerical modeling and optimization of the thermal and thermo-mechanic behavior of the disk and also the heat sink structure, mostly based on finite element models. For further increasing the energy and efficiency of pulsed Thin Disk lasers, the effects of amplified spontaneous emission (ASE) are a core issue. Actual efforts are oriented towards short pulse and ultra-short pulse amplifiers with (multi-)kW average power or Joule-class Thin Disk amplifiers, but also on new designs for cw thin disk MOPA designs.
Interactive high-resolution isosurface ray casting on multicore processors.
Wang, Qin; JaJa, Joseph
2008-01-01
We present a new method for the interactive rendering of isosurfaces using ray casting on multi-core processors. This method consists of a combination of an object-order traversal that coarsely identifies possible candidate 3D data blocks for each small set of contiguous pixels, and an isosurface ray casting strategy tailored for the resulting limited-size lists of candidate 3D data blocks. While static screen partitioning is widely used in the literature, our scheme performs dynamic allocation of groups of ray casting tasks to ensure almost equal loads among the different threads running on multi-cores while maintaining spatial locality. We also make careful use of memory management environment commonly present in multi-core processors. We test our system on a two-processor Clovertown platform, each consisting of a Quad-Core 1.86-GHz Intel Xeon Processor, for a number of widely different benchmarks. The detailed experimental results show that our system is efficient and scalable, and achieves high cache performance and excellent load balancing, resulting in an overall performance that is superior to any of the previous algorithms. In fact, we achieve an interactive isosurface rendering on a 1024(2) screen for all the datasets tested up to the maximum size of the main memory of our platform.
Phase Equilibria of a S- and C-Poor Lunar Core
NASA Technical Reports Server (NTRS)
Righter, K.; Pando, K.; Go, B. M.; Danielson, L. R.; Habermann, M.
2016-01-01
The composition of the lunar core can have a large impact on its thermal evolution, possible early dynamo creation, and physical state. Geochemical measurements have placed better constraints on the S and C content of the lunar mantle. In this study we have carried out phase equilibrium studies of geochemically plausible S- and C-poor lunar core compositions in the Fe-Ni-S-C system, and apply them to the early history of the Moon. We chose two bulk core compositions, with differing S and C content based on geochemical analyses of S and C trapped melts in Apollo samples, and on the partitioning of S and C between metal and silicate. This approach allowed calculation of core S and C contents - 90% Fe, 9% Ni, 0.5% C, and 0.375% S by weight; a second composition contained 1% each of S and C. Experiments were carried out from 1473K to 1973K and 1 GPa to 5 GPa, in piston cylinder and multi- anvil apparatuses. Combination of the thermal model of with our results, shows that a solid inner core (and therefore initiation of a dynamo) may have been possible in the earliest history of the Moon (approximately 4.2 Ga ago), in agreement with. Thus a volatile poor lunar core may explain the thermal and magnetic history of the Moon.
Equalizer: a scalable parallel rendering framework.
Eilemann, Stefan; Makhinya, Maxim; Pajarola, Renato
2009-01-01
Continuing improvements in CPU and GPU performances as well as increasing multi-core processor and cluster-based parallelism demand for flexible and scalable parallel rendering solutions that can exploit multipipe hardware accelerated graphics. In fact, to achieve interactive visualization, scalable rendering systems are essential to cope with the rapid growth of data sets. However, parallel rendering systems are non-trivial to develop and often only application specific implementations have been proposed. The task of developing a scalable parallel rendering framework is even more difficult if it should be generic to support various types of data and visualization applications, and at the same time work efficiently on a cluster with distributed graphics cards. In this paper we introduce a novel system called Equalizer, a toolkit for scalable parallel rendering based on OpenGL which provides an application programming interface (API) to develop scalable graphics applications for a wide range of systems ranging from large distributed visualization clusters and multi-processor multipipe graphics systems to single-processor single-pipe desktop machines. We describe the system architecture, the basic API, discuss its advantages over previous approaches, present example configurations and usage scenarios as well as scalability results.
A study of the parallel algorithm for large-scale DC simulation of nonlinear systems
NASA Astrophysics Data System (ADS)
Cortés Udave, Diego Ernesto; Ogrodzki, Jan; Gutiérrez de Anda, Miguel Angel
Newton-Raphson DC analysis of large-scale nonlinear circuits may be an extremely time consuming process even if sparse matrix techniques and bypassing of nonlinear models calculation are used. A slight decrease in the time required for this task may be enabled on multi-core, multithread computers if the calculation of the mathematical models for the nonlinear elements as well as the stamp management of the sparse matrix entries are managed through concurrent processes. This numerical complexity can be further reduced via the circuit decomposition and parallel solution of blocks taking as a departure point the BBD matrix structure. This block-parallel approach may give a considerable profit though it is strongly dependent on the system topology and, of course, on the processor type. This contribution presents the easy-parallelizable decomposition-based algorithm for DC simulation and provides a detailed study of its effectiveness.
Software Defined Radio with Parallelized Software Architecture
NASA Technical Reports Server (NTRS)
Heckler, Greg
2013-01-01
This software implements software-defined radio procession over multicore, multi-CPU systems in a way that maximizes the use of CPU resources in the system. The software treats each processing step in either a communications or navigation modulator or demodulator system as an independent, threaded block. Each threaded block is defined with a programmable number of input or output buffers; these buffers are implemented using POSIX pipes. In addition, each threaded block is assigned a unique thread upon block installation. A modulator or demodulator system is built by assembly of the threaded blocks into a flow graph, which assembles the processing blocks to accomplish the desired signal processing. This software architecture allows the software to scale effortlessly between single CPU/single-core computers or multi-CPU/multi-core computers without recompilation. NASA spaceflight and ground communications systems currently rely exclusively on ASICs or FPGAs. This software allows low- and medium-bandwidth (100 bps to approx.50 Mbps) software defined radios to be designed and implemented solely in C/C++ software, while lowering development costs and facilitating reuse and extensibility.
2009-06-01
to floating point , to multi-level logic. 2 Overview Self-aware computation can be distinguished from existing computational models which are...systems have advanced to the point that the time is ripe to realize such a system. To illustrate, let us examine each of the key aspects of self...servers for each service, there are no single points of failure in the system. If an OS or user core has a failure, one of several introspection cores
A study of core Thomson scattering measurements in ITER using a multi-laser approach
NASA Astrophysics Data System (ADS)
Kurskiev, G. S.; Sdvizhenskii, P. A.; Bassan, M.; Andrew, P.; Bazhenov, A. N.; Bukreev, I. M.; Chernakov, P. V.; Kochergin, M. M.; Kukushkin, A. B.; Kukushkin, A. S.; Mukhin, E. E.; Razdobarin, A. G.; Samsonov, D. S.; Semenov, V. V.; Tolstyakov, S. Yu.; Kajita, S.; Masyukevich, S. V.
2015-05-01
The electron component is the main channel for anomalous power loss and the main indicator of transient processes in the tokamak plasma. The electron temperature and density profiles mainly determine the operational mode of the machine. This imposes demanding requirements on the precision and on the spatial and temporal resolution of the Thomson scattering (TS) measurements. Measurements of such high electron temperature with good accuracy in a large fusion device such as ITER using TS encounter a number of physical problems. The 40 keV TS spectrum has a significant blue shift. Due to the transmission functions of the fibres and to their darkening that can occur under a strong neutron irradiation, the operational wavelength range is bounded on the blue side. For example, high temperature measurements become impossible with the 1064 nm probing wavelength since the TS signal within the boundaries of the operational window weakly depends on Te. The second problem is connected with the TS calibration. The TS system for a large fusion machine like ITER will have a set of optical components inaccessible for maintenance, and their spectral characteristics may change with time. Since the present concept of the TS system for ITER relies on the classical approach to measuring the shape of the scattered spectra using wide spectral channels, the diagnostic will be very sensitive to the changes in the optical transmission. The third complication is connected with the deviation of the electron velocity distribution function from a Maxwellian that can happen under a strong ECRH/ECCD, and it may additionally hamper the measurements. This paper analyses the advantages of a ‘multi-laser approach’ implementation for the current design of the core TS system. Such an approach assumes simultaneous plasma probing with different wavelengths that allows the measurement accuracy to be improved significantly and to perform the spectral calibration of the TS system. Comparative analysis of the conservative and advanced approaches is given.
Stochastic Local Search for Core Membership Checking in Hedonic Games
NASA Astrophysics Data System (ADS)
Keinänen, Helena
Hedonic games have emerged as an important tool in economics and show promise as a useful formalism to model multi-agent coalition formation in AI as well as group formation in social networks. We consider a coNP-complete problem of core membership checking in hedonic coalition formation games. No previous algorithms to tackle the problem have been presented. In this work, we overcome this by developing two stochastic local search algorithms for core membership checking in hedonic games. We demonstrate the usefulness of the algorithms by showing experimentally that they find solutions efficiently, particularly for large agent societies.
An embedded multi-core parallel model for real-time stereo imaging
NASA Astrophysics Data System (ADS)
He, Wenjing; Hu, Jian; Niu, Jingyu; Li, Chuanrong; Liu, Guangyu
2018-04-01
The real-time processing based on embedded system will enhance the application capability of stereo imaging for LiDAR and hyperspectral sensor. The task partitioning and scheduling strategies for embedded multiprocessor system starts relatively late, compared with that for PC computer. In this paper, aimed at embedded multi-core processing platform, a parallel model for stereo imaging is studied and verified. After analyzing the computing amount, throughout capacity and buffering requirements, a two-stage pipeline parallel model based on message transmission is established. This model can be applied to fast stereo imaging for airborne sensors with various characteristics. To demonstrate the feasibility and effectiveness of the parallel model, a parallel software was designed using test flight data, based on the 8-core DSP processor TMS320C6678. The results indicate that the design performed well in workload distribution and had a speed-up ratio up to 6.4.
Genten: Software for Generalized Tensor Decompositions v. 1.0.0
DOE Office of Scientific and Technical Information (OSTI.GOV)
Phipps, Eric T.; Kolda, Tamara G.; Dunlavy, Daniel
Tensors, or multidimensional arrays, are a powerful mathematical means of describing multiway data. This software provides computational means for decomposing or approximating a given tensor in terms of smaller tensors of lower dimension, focusing on decomposition of large, sparse tensors. These techniques have applications in many scientific areas, including signal processing, linear algebra, computer vision, numerical analysis, data mining, graph analysis, neuroscience and more. The software is designed to take advantage of parallelism present emerging computer architectures such has multi-core CPUs, many-core accelerators such as the Intel Xeon Phi, and computation-oriented GPUs to enable efficient processing of large tensors.
Misra, Sanchit; Pamnany, Kiran; Aluru, Srinivas
2015-01-01
Construction of whole-genome networks from large-scale gene expression data is an important problem in systems biology. While several techniques have been developed, most cannot handle network reconstruction at the whole-genome scale, and the few that can, require large clusters. In this paper, we present a solution on the Intel Xeon Phi coprocessor, taking advantage of its multi-level parallelism including many x86-based cores, multiple threads per core, and vector processing units. We also present a solution on the Intel® Xeon® processor. Our solution is based on TINGe, a fast parallel network reconstruction technique that uses mutual information and permutation testing for assessing statistical significance. We demonstrate the first ever inference of a plant whole genome regulatory network on a single chip by constructing a 15,575 gene network of the plant Arabidopsis thaliana from 3,137 microarray experiments in only 22 minutes. In addition, our optimization for parallelizing mutual information computation on the Intel Xeon Phi coprocessor holds out lessons that are applicable to other domains.
The microstructure and magnetic properties of Cu/CuO/Ni core/multi-shell nanowire arrays
NASA Astrophysics Data System (ADS)
Yang, Feng; Shi, Jie; Zhang, Xiaofeng; Hao, Shijie; Liu, Yinong; Feng, Chun; Cui, Lishan
2018-04-01
Multifunctional metal/oxide/metal core/multi-shell nanowire arrays were prepared mostly by physical or chemical vapor deposition. In our study, the Cu/CuO/Ni core/multi-shell nanowire arrays were prepared by AAO template-electrodeposition and oxidation processes. The Cu/Ni core/shell nanowire arrays were prepared by AAO template-electrodeposition method. The microstructure and chemical compositions of the core/multi-shell nanowires and core/shell nanowires have been characterized using transmission electron microscopy with HADDF-STEM and X-ray diffraction. Magnetization measurements revealed that the Cu/CuO/Ni and Cu/Ni nanowire arrays have high coercivity and remanence ratio.
Many-core computing for space-based stereoscopic imaging
NASA Astrophysics Data System (ADS)
McCall, Paul; Torres, Gildo; LeGrand, Keith; Adjouadi, Malek; Liu, Chen; Darling, Jacob; Pernicka, Henry
The potential benefits of using parallel computing in real-time visual-based satellite proximity operations missions are investigated. Improvements in performance and relative navigation solutions over single thread systems can be achieved through multi- and many-core computing. Stochastic relative orbit determination methods benefit from the higher measurement frequencies, allowing them to more accurately determine the associated statistical properties of the relative orbital elements. More accurate orbit determination can lead to reduced fuel consumption and extended mission capabilities and duration. Inherent to the process of stereoscopic image processing is the difficulty of loading, managing, parsing, and evaluating large amounts of data efficiently, which may result in delays or highly time consuming processes for single (or few) processor systems or platforms. In this research we utilize the Single-Chip Cloud Computer (SCC), a fully programmable 48-core experimental processor, created by Intel Labs as a platform for many-core software research, provided with a high-speed on-chip network for sharing information along with advanced power management technologies and support for message-passing. The results from utilizing the SCC platform for the stereoscopic image processing application are presented in the form of Performance, Power, Energy, and Energy-Delay-Product (EDP) metrics. Also, a comparison between the SCC results and those obtained from executing the same application on a commercial PC are presented, showing the potential benefits of utilizing the SCC in particular, and any many-core platforms in general for real-time processing of visual-based satellite proximity operations missions.
Architecture of optical sensor for recognition of multiple toxic metal ions from water.
Shenashen, M A; El-Safty, S A; Elshehy, E A
2013-09-15
Here, we designed novel optical sensor based on the wormhole hexagonal mesoporous core/multi-shell silica nanoparticles that enabled the selective recognition and removal of these extremely toxic metals from drinking water. The surface-coating process of a mesoporous core/double-shell silica platforms by several consequence decorations using a cationic surfactant with double alkyl tails (CS-DAT) and then a synthesized dicarboxylate 1,5-diphenyl-3-thiocarbazone (III) signaling probe enabled us to create a unique hierarchical multi-shell sensor. In this design, the high loading capacity and wrapping of the CS-DAT and III organic moieties could be achieved, leading to the formation of silica core with multi-shells that formed from double-silica, CS-DAT, and III dressing layers. In this sensing system, notable changes in color and reflectance intensity of the multi-shelled sensor for Cu(2+), Co(2+), Cd(2+), and Hg(2+) ions, were observed at pH 2, 8, 9.5 and 11.5, respectively. The multi-shelled sensor is added to enable accessibility for continuous monitoring of several different toxic metal ions and efficient multi-ion sensing and removal capabilities with respect to reversibility, selectivity, and signal stability. Copyright © 2013 Elsevier B.V. All rights reserved.
LU Factorization with Partial Pivoting for a Multi-CPU, Multi-GPU Shared Memory System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kurzak, Jakub; Luszczek, Pitior; Faverge, Mathieu
2012-03-01
LU factorization with partial pivoting is a canonical numerical procedure and the main component of the High Performance LINPACK benchmark. This article presents an implementation of the algorithm for a hybrid, shared memory, system with standard CPU cores and GPU accelerators. Performance in excess of one TeraFLOPS is achieved using four AMD Magny Cours CPUs and four NVIDIA Fermi GPUs.
End-To-End performance test of the LINC-NIRVANA Wavefront-Sensor system.
NASA Astrophysics Data System (ADS)
Berwein, Juergen; Bertram, Thomas; Conrad, Al; Briegel, Florian; Kittmann, Frank; Zhang, Xiangyu; Mohr, Lars
2011-09-01
LINC-NIRVANA is an imaging Fizeau interferometer, for use in near infrared wavelengths, being built for the Large Binocular Telescope. Multi-conjugate adaptive optics (MCAO) increases the sky coverage and the field of view over which diffraction limited images can be obtained. For its MCAO implementation, Linc-Nirvana utilizes four total wavefront sensors; each of the two beams is corrected by both a ground-layer wavefront sensor (GWS) and a high-layer wavefront sensor (HWS). The GWS controls the adaptive secondary deformable mirror (DM), which is based on an DSP slope computing unit. Whereas the HWS controls an internal DM via computations provided by an off-the-shelf multi-core Linux system. Using wavefront sensor data collected from a prior lab experiment, we have shown via simulation that the Linux based system is sufficient to operate at 1kHz, with jitter well below the needs of the final system. Based on that setup we tested the end-to-end performance and latency through all parts of the system which includes the camera, the wavefront controller, and the deformable mirror. We will present our loop control structure and the results of those performance tests.
A Survey of Recent MARTe Based Systems
NASA Astrophysics Data System (ADS)
Neto, André C.; Alves, Diogo; Boncagni, Luca; Carvalho, Pedro J.; Valcarcel, Daniel F.; Barbalace, Antonio; De Tommasi, Gianmaria; Fernandes, Horácio; Sartori, Filippo; Vitale, Enzo; Vitelli, Riccardo; Zabeo, Luca
2011-08-01
The Multithreaded Application Real-Time executor (MARTe) is a data driven framework environment for the development and deployment of real-time control algorithms. The main ideas which led to the present version of the framework were to standardize the development of real-time control systems, while providing a set of strictly bounded standard interfaces to the outside world and also accommodating a collection of facilities which promote the speed and ease of development, commissioning and deployment of such systems. At the core of every MARTe based application, is a set of independent inter-communicating software blocks, named Generic Application Modules (GAM), orchestrated by a real-time scheduler. The platform independence of its core library provides MARTe the necessary robustness and flexibility for conveniently testing applications in different environments including non-real-time operating systems. MARTe is already being used in several machines, each with its own peculiarities regarding hardware interfacing, supervisory control configuration, operating system and target control application. This paper presents and compares the most recent results of systems using MARTe: the JET Vertical Stabilization system, which uses the Real Time Application Interface (RTAI) operating system on Intel multi-core processors; the COMPASS plasma control system, driven by Linux RT also on Intel multi-core processors; ISTTOK real-time tomography equilibrium reconstruction which shares the same support configuration of COMPASS; JET error field correction coils based on VME, PowerPC and VxWorks; FTU LH reflected power system running on VME, Intel with RTAI.
Designing Next Generation Massively Multithreaded Architectures for Irregular Applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tumeo, Antonino; Secchi, Simone; Villa, Oreste
Irregular applications, such as data mining or graph-based computations, show unpredictable memory/network access patterns and control structures. Massively multi-threaded architectures with large node count, like the Cray XMT, have been shown to address their requirements better than commodity clusters. In this paper we present the approaches that we are currently pursuing to design future generations of these architectures. First, we introduce the Cray XMT and compare it to other multithreaded architectures. We then propose an evolution of the architecture, integrating multiple cores per node and next generation network interconnect. We advocate the use of hardware support for remote memory referencemore » aggregation to optimize network utilization. For this evaluation we developed a highly parallel, custom simulation infrastructure for multi-threaded systems. Our simulator executes unmodified XMT binaries with very large datasets, capturing effects due to contention and hot-spotting, while predicting execution times with greater than 90% accuracy. We also discuss the FPGA prototyping approach that we are employing to study efficient support for irregular applications in next generation manycore processors.« less
NASA Astrophysics Data System (ADS)
Hassan, Rania A.
In the design of complex large-scale spacecraft systems that involve a large number of components and subsystems, many specialized state-of-the-art design tools are employed to optimize the performance of various subsystems. However, there is no structured system-level concept-architecting process. Currently, spacecraft design is heavily based on the heritage of the industry. Old spacecraft designs are modified to adapt to new mission requirements, and feasible solutions---rather than optimal ones---are often all that is achieved. During the conceptual phase of the design, the choices available to designers are predominantly discrete variables describing major subsystems' technology options and redundancy levels. The complexity of spacecraft configurations makes the number of the system design variables that need to be traded off in an optimization process prohibitive when manual techniques are used. Such a discrete problem is well suited for solution with a Genetic Algorithm, which is a global search technique that performs optimization-like tasks. This research presents a systems engineering framework that places design requirements at the core of the design activities and transforms the design paradigm for spacecraft systems to a top-down approach rather than the current bottom-up approach. To facilitate decision-making in the early phases of the design process, the population-based search nature of the Genetic Algorithm is exploited to provide computationally inexpensive---compared to the state-of-the-practice---tools for both multi-objective design optimization and design optimization under uncertainty. In terms of computational cost, those tools are nearly on the same order of magnitude as that of standard single-objective deterministic Genetic Algorithm. The use of a multi-objective design approach provides system designers with a clear tradeoff optimization surface that allows them to understand the effect of their decisions on all the design objectives under consideration simultaneously. Incorporating uncertainties avoids large safety margins and unnecessary high redundancy levels. The focus on low computational cost for the optimization tools stems from the objective that improving the design of complex systems should not be achieved at the expense of a costly design methodology.
A multi-core fiber based interferometer for high temperature sensing
NASA Astrophysics Data System (ADS)
Zhou, Song; Huang, Bo; Shu, Xuewen
2017-04-01
In this paper, we have verified and implemented a Mach-Zehnder interferometer based on seven-core fiber for high temperature sensing application. This proposed structure is based on a multi-mode-multi-core-multi-mode fiber structure sandwiched by a single mode fiber. Between the single-mode and multi-core fiber, a 3 mm long multi-mode fiber is formed for lead-in and lead-out light. The basic operation principle of this device is the use of multi-core modes, single-mode and multi-mode interference coupling is also utilized. Experimental results indicate that this interferometer sensor is capable of accurate measurements of temperatures up to 800 °C, and the temperature sensitivity of the proposed sensor is as high as 170.2 pm/°C, which is much higher than the current existing MZI based temperature sensors (109 pm/°C). This type of sensor is promising for practical high temperature applications due to its advantages including high sensitivity, simple fabrication process, low cost and compactness.
Adaptive multi-resolution 3D Hartree-Fock-Bogoliubov solver for nuclear structure
NASA Astrophysics Data System (ADS)
Pei, J. C.; Fann, G. I.; Harrison, R. J.; Nazarewicz, W.; Shi, Yue; Thornton, S.
2014-08-01
Background: Complex many-body systems, such as triaxial and reflection-asymmetric nuclei, weakly bound halo states, cluster configurations, nuclear fragments produced in heavy-ion fusion reactions, cold Fermi gases, and pasta phases in neutron star crust, are all characterized by large sizes and complex topologies in which many geometrical symmetries characteristic of ground-state configurations are broken. A tool of choice to study such complex forms of matter is an adaptive multi-resolution wavelet analysis. This method has generated much excitement since it provides a common framework linking many diversified methodologies across different fields, including signal processing, data compression, harmonic analysis and operator theory, fractals, and quantum field theory. Purpose: To describe complex superfluid many-fermion systems, we introduce an adaptive pseudospectral method for solving self-consistent equations of nuclear density functional theory in three dimensions, without symmetry restrictions. Methods: The numerical method is based on the multi-resolution and computational harmonic analysis techniques with a multi-wavelet basis. The application of state-of-the-art parallel programming techniques include sophisticated object-oriented templates which parse the high-level code into distributed parallel tasks with a multi-thread task queue scheduler for each multi-core node. The internode communications are asynchronous. The algorithm is variational and is capable of solving coupled complex-geometric systems of equations adaptively, with functional and boundary constraints, in a finite spatial domain of very large size, limited by existing parallel computer memory. For smooth functions, user-defined finite precision is guaranteed. Results: The new adaptive multi-resolution Hartree-Fock-Bogoliubov (HFB) solver madness-hfb is benchmarked against a two-dimensional coordinate-space solver hfb-ax that is based on the B-spline technique and a three-dimensional solver hfodd that is based on the harmonic-oscillator basis expansion. Several examples are considered, including the self-consistent HFB problem for spin-polarized trapped cold fermions and the Skyrme-Hartree-Fock (+BCS) problem for triaxial deformed nuclei. Conclusions: The new madness-hfb framework has many attractive features when applied to nuclear and atomic problems involving many-particle superfluid systems. Of particular interest are weakly bound nuclear configurations close to particle drip lines, strongly elongated and dinuclear configurations such as those present in fission and heavy-ion fusion, and exotic pasta phases that appear in neutron star crust.
NASA Astrophysics Data System (ADS)
Zhai, Yanling; Zhu, Zhijun; Zhu, Chengzhou; Zhu, Jinbo; Ren, Jiangtao; Wang, Erkang; Dong, Shaojun
2013-05-01
Reversible three-state fluorescence switches triggered by light, electricity and chemical inputs based on ``sponges'' of Pyronin Y-doped silica nanoparticles (PYDS) and polyoxometalate K14[Na(H2O)P5W30O110] (Na-POMs) core-shell nanostructures were realized. Under one or two signal inputs, the system exhibited distinct three-state interconvertible automaton, achieving reversible ``on'' and ``off'' luminescence switches via the related luminescence quenching effect. The features of the system correspond to the equivalent circuitry of an IMPLICATION logic gate performing the Boolean operation by using potential and chemical as inputs. Such a multi-chromic device with novel structure possesses several advantages, such as relative low operation voltage, large reproducibility and reversibility, apparent fluorescence contrast, and long-time stability, which make it a suitable candidate for nonvolatile memory devices. In addition, the current protocol for the hybrid film fabrication can be easily extended from the polyoxometalate and organic dyes to other novel nanostructures matched multifunctional stimulus-responsive species and fluorescence materials in the future.Reversible three-state fluorescence switches triggered by light, electricity and chemical inputs based on ``sponges'' of Pyronin Y-doped silica nanoparticles (PYDS) and polyoxometalate K14[Na(H2O)P5W30O110] (Na-POMs) core-shell nanostructures were realized. Under one or two signal inputs, the system exhibited distinct three-state interconvertible automaton, achieving reversible ``on'' and ``off'' luminescence switches via the related luminescence quenching effect. The features of the system correspond to the equivalent circuitry of an IMPLICATION logic gate performing the Boolean operation by using potential and chemical as inputs. Such a multi-chromic device with novel structure possesses several advantages, such as relative low operation voltage, large reproducibility and reversibility, apparent fluorescence contrast, and long-time stability, which make it a suitable candidate for nonvolatile memory devices. In addition, the current protocol for the hybrid film fabrication can be easily extended from the polyoxometalate and organic dyes to other novel nanostructures matched multifunctional stimulus-responsive species and fluorescence materials in the future. Electronic supplementary information (ESI) available: Experimental details and instrumentation; electrochemical, fluorescence and absorption spectra characterizations of hybrid films. See DOI: 10.1039/c3nr00254c
Enhancing Image Processing Performance for PCID in a Heterogeneous Network of Multi-code Processors
NASA Astrophysics Data System (ADS)
Linderman, R.; Spetka, S.; Fitzgerald, D.; Emeny, S.
The Physically-Constrained Iterative Deconvolution (PCID) image deblurring code is being ported to heterogeneous networks of multi-core systems, including Intel Xeons and IBM Cell Broadband Engines. This paper reports results from experiments using the JAWS supercomputer at MHPCC (60 TFLOPS of dual-dual Xeon nodes linked with Infiniband) and the Cell Cluster at AFRL in Rome, NY. The Cell Cluster has 52 TFLOPS of Playstation 3 (PS3) nodes with IBM Cell Broadband Engine multi-cores and 15 dual-quad Xeon head nodes. The interconnect fabric includes Infiniband, 10 Gigabit Ethernet and 1 Gigabit Ethernet to each of the 336 PS3s. The results compare approaches to parallelizing FFT executions across the Xeons and the Cell's Synergistic Processing Elements (SPEs) for frame-level image processing. The experiments included Intel's Performance Primitives and Math Kernel Library, FFTW3.2, and Carnegie Mellon's SPIRAL. Optimization of FFTs in the PCID code led to a decrease in relative processing time for FFTs. Profiling PCID version 6.2, about one year ago, showed the 13 functions that accounted for the highest percentage of processing were all FFT processing functions. They accounted for over 88% of processing time in one run on Xeons. FFT optimizations led to improvement in the current PCID version 8.0. A recent profile showed that only two of the 19 functions with the highest processing time were FFT processing functions. Timing measurements showed that FFT processing for PCID version 8.0 has been reduced to less than 19% of overall processing time. We are working toward a goal of scaling to 200-400 cores per job (1-2 imagery frames/core). Running a pair of cores on each set of frames reduces latency by implementing parallel FFT processing. Our current results show scaling well out to 100 pairs of cores. These results support the next higher level of parallelism in PCID, where groups of several hundred frames each producing one resolved image are sent to cliques of several hundred cores in a round robin fashion. Current efforts toward further performance enhancement for PCID are shifting toward using the Playstations in conjunction with the Xeons to take advantage of outstanding price/performance as well as the Flops/Watt cost advantage. We are fine-tuning the PCID parallization strategy to balance processing over Xeons and Cell BEs to find an optimal partitioning of PCID over the heterogeneous processors. A high performance information management system that exploits native Infiniband multicast is used to improve latency among the head nodes. Using a publication/subscription oriented information management system to implement a unified communications platform makes runs on large HPCs with thousands of intercommunicating cores more flexible and more fault tolerant. It features a loose couplingof publishers to subscribers through intervening brokers. We are also working on enhancing performance for both Xeons and Cell BEs, buy moving selected operations to single precision. Techniques for adapting the code to single precision and performance results are reported.
xQTL workbench: a scalable web environment for multi-level QTL analysis.
Arends, Danny; van der Velde, K Joeri; Prins, Pjotr; Broman, Karl W; Möller, Steffen; Jansen, Ritsert C; Swertz, Morris A
2012-04-01
xQTL workbench is a scalable web platform for the mapping of quantitative trait loci (QTLs) at multiple levels: for example gene expression (eQTL), protein abundance (pQTL), metabolite abundance (mQTL) and phenotype (phQTL) data. Popular QTL mapping methods for model organism and human populations are accessible via the web user interface. Large calculations scale easily on to multi-core computers, clusters and Cloud. All data involved can be uploaded and queried online: markers, genotypes, microarrays, NGS, LC-MS, GC-MS, NMR, etc. When new data types come available, xQTL workbench is quickly customized using the Molgenis software generator. xQTL workbench runs on all common platforms, including Linux, Mac OS X and Windows. An online demo system, installation guide, tutorials, software and source code are available under the LGPL3 license from http://www.xqtl.org. m.a.swertz@rug.nl.
xQTL workbench: a scalable web environment for multi-level QTL analysis
Arends, Danny; van der Velde, K. Joeri; Prins, Pjotr; Broman, Karl W.; Möller, Steffen; Jansen, Ritsert C.; Swertz, Morris A.
2012-01-01
Summary: xQTL workbench is a scalable web platform for the mapping of quantitative trait loci (QTLs) at multiple levels: for example gene expression (eQTL), protein abundance (pQTL), metabolite abundance (mQTL) and phenotype (phQTL) data. Popular QTL mapping methods for model organism and human populations are accessible via the web user interface. Large calculations scale easily on to multi-core computers, clusters and Cloud. All data involved can be uploaded and queried online: markers, genotypes, microarrays, NGS, LC-MS, GC-MS, NMR, etc. When new data types come available, xQTL workbench is quickly customized using the Molgenis software generator. Availability: xQTL workbench runs on all common platforms, including Linux, Mac OS X and Windows. An online demo system, installation guide, tutorials, software and source code are available under the LGPL3 license from http://www.xqtl.org. Contact: m.a.swertz@rug.nl PMID:22308096
2012-10-01
REPORT 3. DATES COVERED (From - To) MAR 2010 – APR 2012 4 . TITLE AND SUBTITLE IMPLICATIONS OF MULT-CORE ARCHITECTURES ON THE DEVELOPMENT OF...Framework for Multicore Information Flow Analysis ...................................... 23 4 4.1 A Hypothetical Reference Architecture... 4 Figure 2: Pentium II Block Diagram
ERIC Educational Resources Information Center
Education Commission of the States, 2015
2015-01-01
Colleges and postsecondary systems across the nation have demonstrated remarkable progress since "Core Principles for Transforming Remediation" was published in 2012. States and institutions are phasing out stand alone or multi-course remediation sequences, resulting in tens of thousands of students more quickly enrolling in and…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Schunert, Sebastian; Schwen, Daniel; Ghassemi, Pedram
This work presents a multi-physics, multi-scale approach to modeling the Transient Test Reactor (TREAT) currently prepared for restart at the Idaho National Laboratory. TREAT fuel is made up of microscopic fuel grains (r ˜ 20µm) dispersed in a graphite matrix. The novelty of this work is in coupling a binary collision Monte-Carlo (BCMC) model to the Finite Element based code Moose for solving a microsopic heat-conduction problem whose driving source is provided by the BCMC model tracking fission fragment energy deposition. This microscopic model is driven by a transient, engineering scale neutronics model coupled to an adiabatic heating model. Themore » macroscopic model provides local power densities and neutron energy spectra to the microscpic model. Currently, no feedback from the microscopic to the macroscopic model is considered. TREAT transient 15 is used to exemplify the capabilities of the multi-physics, multi-scale model, and it is found that the average fuel grain temperature differs from the average graphite temperature by 80 K despite the low-power transient. The large temperature difference has strong implications on the Doppler feedback a potential LEU TREAT core would see, and it underpins the need for multi-physics, multi-scale modeling of a TREAT LEU core.« less
A Bandwidth-Optimized Multi-Core Architecture for Irregular Applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Secchi, Simone; Tumeo, Antonino; Villa, Oreste
This paper presents an architecture template for next-generation high performance computing systems specifically targeted to irregular applications. We start our work by considering that future generation interconnection and memory bandwidth full-system numbers are expected to grow by a factor of 10. In order to keep up with such a communication capacity, while still resorting to fine-grained multithreading as the main way to tolerate unpredictable memory access latencies of irregular applications, we show how overall performance scaling can benefit from the multi-core paradigm. At the same time, we also show how such an architecture template must be coupled with specific techniquesmore » in order to optimize bandwidth utilization and achieve the maximum scalability. We propose a technique based on memory references aggregation, together with the related hardware implementation, as one of such optimization techniques. We explore the proposed architecture template by focusing on the Cray XMT architecture and, using a dedicated simulation infrastructure, validate the performance of our template with two typical irregular applications. Our experimental results prove the benefits provided by both the multi-core approach and the bandwidth optimization reference aggregation technique.« less
Compact Hybrid Laser Rod and Laser System
NASA Technical Reports Server (NTRS)
Pierrottet, Diego F. (Inventor); Busch, George E. (Inventor); Amzajerdian, Farzin (Inventor)
2017-01-01
A hybrid fiber rod includes a fiber core and inner and outer cladding layers. The core is doped with an active element. The inner cladding layer surrounds the core, and has a refractive index substantially equal to that of the core. The outer cladding layer surrounds the inner cladding layer, and has a refractive index less than that of the core and inner cladding layer. The core length is about 30 to 2000 times the core diameter. A hybrid fiber rod laser system includes an oscillator laser, modulating device, the rod, and pump laser diode(s) energizing the rod from opposite ends. The rod acts as a waveguide for pump radiation but allows for free-space propagation of laser radiation. The rod may be used in a laser resonator. The core length is less than about twice the Rayleigh range. Degradation from single-mode to multi-mode beam propagation is thus avoided.
Design of ultrahigh brightness solar-pumped disk laser.
Liang, Dawei; Almeida, Joana
2012-09-10
To significantly improve the solar-pumped laser beam brightness, a multi-Fresnel lens scheme is proposed for side-pumping either a single-crystal Nd:YAG or a core-doped ceramic Sm(3+) Nd:YAG disk. Optimum laser system parameters are found through ZEMAX and LASCAD numerical analysis. An ultrahigh laser beam figure of merit B of 53 W is numerically calculated, corresponding to a significant enhancement of more than 180 times over the previous record. 17.7 W/m(2) collection efficiency is also numerically attained. The strong thermal effects that have hampered present-day rod-type solar-pumped lasers can also be largely alleviated.
Multi-Kepler GPU vs. multi-Intel MIC for spin systems simulations
NASA Astrophysics Data System (ADS)
Bernaschi, M.; Bisson, M.; Salvadore, F.
2014-10-01
We present and compare the performances of two many-core architectures: the Nvidia Kepler and the Intel MIC both in a single system and in cluster configuration for the simulation of spin systems. As a benchmark we consider the time required to update a single spin of the 3D Heisenberg spin glass model by using the Over-relaxation algorithm. We present data also for a traditional high-end multi-core architecture: the Intel Sandy Bridge. The results show that although on the two Intel architectures it is possible to use basically the same code, the performances of a Intel MIC change dramatically depending on (apparently) minor details. Another issue is that to obtain a reasonable scalability with the Intel Phi coprocessor (Phi is the coprocessor that implements the MIC architecture) in a cluster configuration it is necessary to use the so-called offload mode which reduces the performances of the single system. As to the GPU, the Kepler architecture offers a clear advantage with respect to the previous Fermi architecture maintaining exactly the same source code. Scalability of the multi-GPU implementation remains very good by using the CPU as a communication co-processor of the GPU. All source codes are provided for inspection and for double-checking the results.
LASSIM-A network inference toolbox for genome-wide mechanistic modeling.
Magnusson, Rasmus; Mariotti, Guido Pio; Köpsén, Mattias; Lövfors, William; Gawel, Danuta R; Jörnsten, Rebecka; Linde, Jörg; Nordling, Torbjörn E M; Nyman, Elin; Schulze, Sylvie; Nestor, Colm E; Zhang, Huan; Cedersund, Gunnar; Benson, Mikael; Tjärnberg, Andreas; Gustafsson, Mika
2017-06-01
Recent technological advancements have made time-resolved, quantitative, multi-omics data available for many model systems, which could be integrated for systems pharmacokinetic use. Here, we present large-scale simulation modeling (LASSIM), which is a novel mathematical tool for performing large-scale inference using mechanistically defined ordinary differential equations (ODE) for gene regulatory networks (GRNs). LASSIM integrates structural knowledge about regulatory interactions and non-linear equations with multiple steady state and dynamic response expression datasets. The rationale behind LASSIM is that biological GRNs can be simplified using a limited subset of core genes that are assumed to regulate all other gene transcription events in the network. The LASSIM method is implemented as a general-purpose toolbox using the PyGMO Python package to make the most of multicore computers and high performance clusters, and is available at https://gitlab.com/Gustafsson-lab/lassim. As a method, LASSIM works in two steps, where it first infers a non-linear ODE system of the pre-specified core gene expression. Second, LASSIM in parallel optimizes the parameters that model the regulation of peripheral genes by core system genes. We showed the usefulness of this method by applying LASSIM to infer a large-scale non-linear model of naïve Th2 cell differentiation, made possible by integrating Th2 specific bindings, time-series together with six public and six novel siRNA-mediated knock-down experiments. ChIP-seq showed significant overlap for all tested transcription factors. Next, we performed novel time-series measurements of total T-cells during differentiation towards Th2 and verified that our LASSIM model could monitor those data significantly better than comparable models that used the same Th2 bindings. In summary, the LASSIM toolbox opens the door to a new type of model-based data analysis that combines the strengths of reliable mechanistic models with truly systems-level data. We demonstrate the power of this approach by inferring a mechanistically motivated, genome-wide model of the Th2 transcription regulatory system, which plays an important role in several immune related diseases.
Programmability of nanowire networks
NASA Astrophysics Data System (ADS)
Bellew, A. T.; Bell, A. P.; McCarthy, E. K.; Fairfield, J. A.; Boland, J. J.
2014-07-01
Electrical connectivity in networks of nanoscale junctions must be better understood if nanowire devices are to be scaled up from single wires to functional material systems. We show that the natural connectivity behaviour found in random nanowire networks presents a new paradigm for creating multi-functional, programmable materials. In devices made from networks of Ni/NiO core-shell nanowires at different length scales, we discover the emergence of distinct behavioural regimes when networks are electrically stressed. We show that a small network, with few nanowire-nanowire junctions, acts as a unipolar resistive switch, demonstrating very high ON/OFF current ratios (>105). However, large networks of nanowires distribute an applied bias across a large number of junctions, and thus respond not by switching but instead by evolving connectivity. We demonstrate that these emergent properties lead to fault-tolerant materials whose resistance may be tuned, and which are capable of adaptively reconfiguring under stress. By combining these two behavioural regimes, we demonstrate that the same nanowire network may be programmed to act both as a metallic interconnect, and a resistive switch device with high ON/OFF ratio. These results enable the fabrication of programmable, multi-functional materials from random nanowire networks.Electrical connectivity in networks of nanoscale junctions must be better understood if nanowire devices are to be scaled up from single wires to functional material systems. We show that the natural connectivity behaviour found in random nanowire networks presents a new paradigm for creating multi-functional, programmable materials. In devices made from networks of Ni/NiO core-shell nanowires at different length scales, we discover the emergence of distinct behavioural regimes when networks are electrically stressed. We show that a small network, with few nanowire-nanowire junctions, acts as a unipolar resistive switch, demonstrating very high ON/OFF current ratios (>105). However, large networks of nanowires distribute an applied bias across a large number of junctions, and thus respond not by switching but instead by evolving connectivity. We demonstrate that these emergent properties lead to fault-tolerant materials whose resistance may be tuned, and which are capable of adaptively reconfiguring under stress. By combining these two behavioural regimes, we demonstrate that the same nanowire network may be programmed to act both as a metallic interconnect, and a resistive switch device with high ON/OFF ratio. These results enable the fabrication of programmable, multi-functional materials from random nanowire networks. Electronic supplementary information (ESI) available: Nanowire statistics (length, diameter statistics, and oxide thickness) are provided. Forming curves for single junctions and networks. Passive voltage contrast image demonstrating selectivity of conductive pathways in 100 μm network. See DOI: 10.1039/c4nr02338b
Scalable Performance Measurement and Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gamblin, Todd
2009-01-01
Concurrency levels in large-scale, distributed-memory supercomputers are rising exponentially. Modern machines may contain 100,000 or more microprocessor cores, and the largest of these, IBM's Blue Gene/L, contains over 200,000 cores. Future systems are expected to support millions of concurrent tasks. In this dissertation, we focus on efficient techniques for measuring and analyzing the performance of applications running on very large parallel machines. Tuning the performance of large-scale applications can be a subtle and time-consuming task because application developers must measure and interpret data from many independent processes. While the volume of the raw data scales linearly with the number ofmore » tasks in the running system, the number of tasks is growing exponentially, and data for even small systems quickly becomes unmanageable. Transporting performance data from so many processes over a network can perturb application performance and make measurements inaccurate, and storing such data would require a prohibitive amount of space. Moreover, even if it were stored, analyzing the data would be extremely time-consuming. In this dissertation, we present novel methods for reducing performance data volume. The first draws on multi-scale wavelet techniques from signal processing to compress systemwide, time-varying load-balance data. The second uses statistical sampling to select a small subset of running processes to generate low-volume traces. A third approach combines sampling and wavelet compression to stratify performance data adaptively at run-time and to reduce further the cost of sampled tracing. We have integrated these approaches into Libra, a toolset for scalable load-balance analysis. We present Libra and show how it can be used to analyze data from large scientific applications scalably.« less
Katouda, Michio; Naruse, Akira; Hirano, Yukihiko; Nakajima, Takahito
2016-11-15
A new parallel algorithm and its implementation for the RI-MP2 energy calculation utilizing peta-flop-class many-core supercomputers are presented. Some improvements from the previous algorithm (J. Chem. Theory Comput. 2013, 9, 5373) have been performed: (1) a dual-level hierarchical parallelization scheme that enables the use of more than 10,000 Message Passing Interface (MPI) processes and (2) a new data communication scheme that reduces network communication overhead. A multi-node and multi-GPU implementation of the present algorithm is presented for calculations on a central processing unit (CPU)/graphics processing unit (GPU) hybrid supercomputer. Benchmark results of the new algorithm and its implementation using the K computer (CPU clustering system) and TSUBAME 2.5 (CPU/GPU hybrid system) demonstrate high efficiency. The peak performance of 3.1 PFLOPS is attained using 80,199 nodes of the K computer. The peak performance of the multi-node and multi-GPU implementation is 514 TFLOPS using 1349 nodes and 4047 GPUs of TSUBAME 2.5. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
ERIC Educational Resources Information Center
Beachboard, John; Aytes, Kregg
2011-01-01
Core MBA IT courses have tended to be survey courses that cover important topics but often do not sufficiently engage students. The result is that many top-ranked MBA programs have not found such courses useful enough to include in their core MBA requirements. In this paper, we present a design of an MBA course emphasizing information technology…
The design of multi temperature and humidity monitoring system for incubator
NASA Astrophysics Data System (ADS)
Yu, Junyu; Xu, Peng; Peng, Zitao; Qiang, Haonan; Shen, Xiaoyan
2017-01-01
Currently, there is only one monitor of the temperature and humidity in an incubator, which may cause inaccurate or unreliable data, and even endanger the life safety of the baby. In order to solve this problem,we designed a multi-point temperature and humidity monitoring system for incubators. The system uses the STC12C5A60S2 microcontrollers as the sender core chip which is connected to four AM2321 temperature and humidity sensors. We select STM32F103ZET6 core development board as the receiving end,cooperating with Zigbee wireless transmitting and receiving module to realize data acquisition and transmission. This design can realize remote real-time observation data on the computer by communicating with PC via Ethernet. Prototype tests show that the system can effectively collect and display the information of temperature and humidity of multiple incubators at the same time and there are four monitors in each incubator.
Application of Advanced Multi-Core Processor Technologies to Oceanographic Research
2013-09-30
STM32 NXP LPC series No Proprietary Microchip PIC32/DSPIC No > 500 mW; < 5 W ARM Cortex TI OMAP TI Sitara Broadcom BCM2835 Varies FPGA...1 DISTRIBUTION STATEMENT A. Approved for public release; distribution is unlimited. Application of Advanced Multi-Core Processor Technologies...state-of-the-art information processing architectures. OBJECTIVES Next-generation processor architectures (multi-core, multi-threaded) hold the
High-resolution Observations of Hα Spectra with a Subtractive Double Pass
NASA Astrophysics Data System (ADS)
Beck, C.; Rezaei, R.; Choudhary, D. P.; Gosain, S.; Tritschler, A.; Louis, R. E.
2018-02-01
High-resolution imaging spectroscopy in solar physics has relied on Fabry-Pérot interferometers (FPIs) in recent years. FPI systems, however, become technically challenging and expensive for telescopes larger than the 1 m class. A conventional slit spectrograph with a diffraction-limited performance over a large field of view (FOV) can be built at much lower cost and effort. It can be converted into an imaging spectro(polari)meter using the concept of a subtractive double pass (SDP). We demonstrate that an SDP system can reach a similar performance as FPI-based systems with a high spatial and moderate spectral resolution across a FOV of 100^'' ×100^' ' with a spectral coverage of 1 nm. We use Hα spectra taken with an SDP system at the Dunn Solar Telescope and complementary full-disc data to infer the properties of small-scale superpenumbral filaments. We find that the majority of all filaments end in patches of opposite-polarity fields. The internal fine-structure in the line-core intensity of Hα at spatial scales of about 0.5'' exceeds that in other parameters such as the line width, indicating small-scale opacity effects in a larger-scale structure with common properties. We conclude that SDP systems in combination with (multi-conjugate) adaptive optics are a valid alternative to FPI systems when high spatial resolution and a large FOV are required. They can also reach a cadence that is comparable to that of FPI systems, while providing a much larger spectral range and a simultaneous multi-line capability.
An Energy-Aware Runtime Management of Multi-Core Sensory Swarms.
Kim, Sungchan; Yang, Hoeseok
2017-08-24
In sensory swarms, minimizing energy consumption under performance constraint is one of the key objectives. One possible approach to this problem is to monitor application workload that is subject to change at runtime, and to adjust system configuration adaptively to satisfy the performance goal. As today's sensory swarms are usually implemented using multi-core processors with adjustable clock frequency, we propose to monitor the CPU workload periodically and adjust the task-to-core allocation or clock frequency in an energy-efficient way in response to the workload variations. In doing so, we present an online heuristic that determines the most energy-efficient adjustment that satisfies the performance requirement. The proposed method is based on a simple yet effective energy model that is built upon performance prediction using IPC (instructions per cycle) measured online and power equation derived empirically. The use of IPC accounts for memory intensities of a given workload, enabling the accurate prediction of execution time. Hence, the model allows us to rapidly and accurately estimate the effect of the two control knobs, clock frequency adjustment and core allocation. The experiments show that the proposed technique delivers considerable energy saving of up to 45%compared to the state-of-the-art multi-core energy management technique.
An Energy-Aware Runtime Management of Multi-Core Sensory Swarms
Kim, Sungchan
2017-01-01
In sensory swarms, minimizing energy consumption under performance constraint is one of the key objectives. One possible approach to this problem is to monitor application workload that is subject to change at runtime, and to adjust system configuration adaptively to satisfy the performance goal. As today’s sensory swarms are usually implemented using multi-core processors with adjustable clock frequency, we propose to monitor the CPU workload periodically and adjust the task-to-core allocation or clock frequency in an energy-efficient way in response to the workload variations. In doing so, we present an online heuristic that determines the most energy-efficient adjustment that satisfies the performance requirement. The proposed method is based on a simple yet effective energy model that is built upon performance prediction using IPC (instructions per cycle) measured online and power equation derived empirically. The use of IPC accounts for memory intensities of a given workload, enabling the accurate prediction of execution time. Hence, the model allows us to rapidly and accurately estimate the effect of the two control knobs, clock frequency adjustment and core allocation. The experiments show that the proposed technique delivers considerable energy saving of up to 45%compared to the state-of-the-art multi-core energy management technique. PMID:28837094
An interactive display system for large-scale 3D models
NASA Astrophysics Data System (ADS)
Liu, Zijian; Sun, Kun; Tao, Wenbing; Liu, Liman
2018-04-01
With the improvement of 3D reconstruction theory and the rapid development of computer hardware technology, the reconstructed 3D models are enlarging in scale and increasing in complexity. Models with tens of thousands of 3D points or triangular meshes are common in practical applications. Due to storage and computing power limitation, it is difficult to achieve real-time display and interaction with large scale 3D models for some common 3D display software, such as MeshLab. In this paper, we propose a display system for large-scale 3D scene models. We construct the LOD (Levels of Detail) model of the reconstructed 3D scene in advance, and then use an out-of-core view-dependent multi-resolution rendering scheme to realize the real-time display of the large-scale 3D model. With the proposed method, our display system is able to render in real time while roaming in the reconstructed scene and 3D camera poses can also be displayed. Furthermore, the memory consumption can be significantly decreased via internal and external memory exchange mechanism, so that it is possible to display a large scale reconstructed scene with over millions of 3D points or triangular meshes in a regular PC with only 4GB RAM.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Powalka, Mathieu; Lançon, Ariane; Duc, Pierre-Alain
Large samples of globular clusters (GC) with precise multi-wavelength photometry are becoming increasingly available and can be used to constrain the formation history of galaxies. We present the results of an analysis of Milky Way (MW) and Virgo core GCs based on 5 optical-near-infrared colors and 10 synthetic stellar population models. For the MW GCs, the models tend to agree on photometric ages and metallicities, with values similar to those obtained with previous studies. When used with Virgo core GCs, for which photometry is provided by the Next Generation Virgo cluster Survey (NGVS), the same models generically return younger ages.more » This is a consequence of the systematic differences observed between the locus occupied by Virgo core GCs and models in panchromatic color space. Only extreme fine-tuning of the adjustable parameters available to us can make the majority of the best-fit ages old. Although we cannot exclude that the formation history of the Virgo core may lead to more conspicuous populations of relatively young GCs than in other environments, we emphasize that the intrinsic properties of the Virgo GCs are likely to differ systematically from those assumed in the models. Thus, the large wavelength coverage and photometric quality of modern GC samples, such as those used here, is not by itself sufficient to better constrain the GC formation histories. Models matching the environment-dependent characteristics of GCs in multi-dimensional color space are needed to improve the situation.« less
Seekhao, Nuttiiya; Shung, Caroline; JaJa, Joseph; Mongeau, Luc; Li-Jessen, Nicole Y K
2016-05-01
We present an efficient and scalable scheme for implementing agent-based modeling (ABM) simulation with In Situ visualization of large complex systems on heterogeneous computing platforms. The scheme is designed to make optimal use of the resources available on a heterogeneous platform consisting of a multicore CPU and a GPU, resulting in minimal to no resource idle time. Furthermore, the scheme was implemented under a client-server paradigm that enables remote users to visualize and analyze simulation data as it is being generated at each time step of the model. Performance of a simulation case study of vocal fold inflammation and wound healing with 3.8 million agents shows 35× and 7× speedup in execution time over single-core and multi-core CPU respectively. Each iteration of the model took less than 200 ms to simulate, visualize and send the results to the client. This enables users to monitor the simulation in real-time and modify its course as needed.
Matrix Algebra for GPU and Multicore Architectures (MAGMA) for Large Petascale Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dongarra, Jack J.; Tomov, Stanimire
2014-03-24
The goal of the MAGMA project is to create a new generation of linear algebra libraries that achieve the fastest possible time to an accurate solution on hybrid Multicore+GPU-based systems, using all the processing power that future high-end systems can make available within given energy constraints. Our efforts at the University of Tennessee achieved the goals set in all of the five areas identified in the proposal: 1. Communication optimal algorithms; 2. Autotuning for GPU and hybrid processors; 3. Scheduling and memory management techniques for heterogeneity and scale; 4. Fault tolerance and robustness for large scale systems; 5. Building energymore » efficiency into software foundations. The University of Tennessee’s main contributions, as proposed, were the research and software development of new algorithms for hybrid multi/many-core CPUs and GPUs, as related to two-sided factorizations and complete eigenproblem solvers, hybrid BLAS, and energy efficiency for dense, as well as sparse, operations. Furthermore, as proposed, we investigated and experimented with various techniques targeting the five main areas outlined.« less
Data Parallel Bin-Based Indexing for Answering Queries on Multi-Core Architectures
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gosink, Luke; Wu, Kesheng; Bethel, E. Wes
2009-06-02
The multi-core trend in CPUs and general purpose graphics processing units (GPUs) offers new opportunities for the database community. The increase of cores at exponential rates is likely to affect virtually every server and client in the coming decade, and presents database management systems with a huge, compelling disruption that will radically change how processing is done. This paper presents a new parallel indexing data structure for answering queries that takes full advantage of the increasing thread-level parallelism emerging in multi-core architectures. In our approach, our Data Parallel Bin-based Index Strategy (DP-BIS) first bins the base data, and then partitionsmore » and stores the values in each bin as a separate, bin-based data cluster. In answering a query, the procedures for examining the bin numbers and the bin-based data clusters offer the maximum possible level of concurrency; each record is evaluated by a single thread and all threads are processed simultaneously in parallel. We implement and demonstrate the effectiveness of DP-BIS on two multi-core architectures: a multi-core CPU and a GPU. The concurrency afforded by DP-BIS allows us to fully utilize the thread-level parallelism provided by each architecture--for example, our GPU-based DP-BIS implementation simultaneously evaluates over 12,000 records with an equivalent number of concurrently executing threads. In comparing DP-BIS's performance across these architectures, we show that the GPU-based DP-BIS implementation requires significantly less computation time to answer a query than the CPU-based implementation. We also demonstrate in our analysis that DP-BIS provides better overall performance than the commonly utilized CPU and GPU-based projection index. Finally, due to data encoding, we show that DP-BIS accesses significantly smaller amounts of data than index strategies that operate solely on a column's base data; this smaller data footprint is critical for parallel processors that possess limited memory resources (e.g., GPUs).« less
Tiled architecture of a CNN-mostly IP system
NASA Astrophysics Data System (ADS)
Spaanenburg, Lambert; Malki, Suleyman
2009-05-01
Multi-core architectures have been popularized with the advent of the IBM CELL. On a finer grain the problems in scheduling multi-cores have already existed in the tiled architectures, such as the EPIC and Da Vinci. It is not easy to evaluate the performance of a schedule on such architecture as historical data are not available. One solution is to compile algorithms for which an optimal schedule is known by analysis. A typical example is an algorithm that is already defined in terms of many collaborating simple nodes, such as a Cellular Neural Network (CNN). A simple node with a local register stack together with a 'rotating wheel' internal communication mechanism has been proposed. Though the basic CNN allows for a tiled implementation of a tiled algorithm on a tiled structure, a practical CNN system will have to disturb this regularity by the additional need for arithmetical and logical operations. Arithmetic operations are needed for instance to accommodate for low-level image processing, while logical operations are needed to fork and merge different data streams without use of the external memory. It is found that the 'rotating wheel' internal communication mechanism still handles such mechanisms without the need for global control. Overall the CNN system provides for a practical network size as implemented on a FPGA, can be easily used as embedded IP and provides a clear benchmark for a multi-core compiler.
Lohan, S B; Icken, N; Teutloff, C; Saeidpour, S; Bittl, R; Lademann, J; Fleige, E; Haag, R; Haag, S F; Meinke, M C
2016-03-30
Dendritic core-multi shell (CMS) particles are polymer based systems consisting of a dendritic polar polyglycerol polymer core surrounded by a two-layer shell of nonpolar C18 alkyl chains and hydrophilic polyethylene glycol. Belonging to nanotransport systems (NTS) they allow the transport and storage of molecules with different chemical characters. Their amphipihilic character CMS-NTS permits good solubility in aqueous and organic solutions. We showed by multifrequency electron paramagnetic resonance (EPR) spectroscopy that spin-labeled 5-doxyl stearic acid (5DSA) can be loaded into the CMS-NTS. Furthermore, the release of 5DSA from the carrier into the stratum corneum of porcine skin was monitored ex vivo by EPR spectroscopy. Additionally, the penetration of the CMS-NTS into the skin was analyzed by fluorescence microscopy using indocarbocyanine (ICC) covalently bound to the nanocarrier. Thereby, no transport into the viable skin was observed, whereas the CMS-NTS had penetrated into the hair follicles down to a depth of 340 μm ± 82 μm. Thus, it could be shown that the combined application of fluorescence microscopy and multi-frequency EPR spectroscopy can be an efficient tool for investigating the loading of spin labeled drugs to nanocarrier systems, drug release and penetration into the skin as well as the localization of the NTS in the skin. Copyright © 2016 Elsevier B.V. All rights reserved.
pyPaSWAS: Python-based multi-core CPU and GPU sequence alignment.
Warris, Sven; Timal, N Roshan N; Kempenaar, Marcel; Poortinga, Arne M; van de Geest, Henri; Varbanescu, Ana L; Nap, Jan-Peter
2018-01-01
Our previously published CUDA-only application PaSWAS for Smith-Waterman (SW) sequence alignment of any type of sequence on NVIDIA-based GPUs is platform-specific and therefore adopted less than could be. The OpenCL language is supported more widely and allows use on a variety of hardware platforms. Moreover, there is a need to promote the adoption of parallel computing in bioinformatics by making its use and extension more simple through more and better application of high-level languages commonly used in bioinformatics, such as Python. The novel application pyPaSWAS presents the parallel SW sequence alignment code fully packed in Python. It is a generic SW implementation running on several hardware platforms with multi-core systems and/or GPUs that provides accurate sequence alignments that also can be inspected for alignment details. Additionally, pyPaSWAS support the affine gap penalty. Python libraries are used for automated system configuration, I/O and logging. This way, the Python environment will stimulate further extension and use of pyPaSWAS. pyPaSWAS presents an easy Python-based environment for accurate and retrievable parallel SW sequence alignments on GPUs and multi-core systems. The strategy of integrating Python with high-performance parallel compute languages to create a developer- and user-friendly environment should be considered for other computationally intensive bioinformatics algorithms.
Square-core bundles for astronomical imaging
NASA Astrophysics Data System (ADS)
Bryant, Julia J.; Bland-Hawthorn, Joss
2012-09-01
Optical fibre imaging bundles (hexabundles) are proving to be the next logical step for large galaxy surveys as they offer spatially-resolved spectroscopy of galaxies and can be used with conventional fibre positioners. Hexabundles have been effectively demonstrated in the Sydney-AAO Multi-object IFS (SAMI) instrument at the Anglo- Australian Telescope[5]. Based on the success of hexabundles that have circular cores, we have characterised a bundle made instead from square-core fibres. Square cores naturally pack more evenly, which reduces the interstitial holes and can increase the covering, or filling fraction. Furthermore the regular packing simplifies the process of combining and dithering the final images. We discuss the relative issues of filling fraction, focal ratio degradation (FRD), and cross-talk, and find that square-core bundles perform well enough to warrant further development as a format for imaging fibre bundles.
Multi-Threaded DNA Tag/Anti-Tag Library Generator for Multi-Core Platforms
2009-05-01
base pair) Watson ‐ Crick strand pairs that bind perfectly within pairs, but poorly across pairs. A variety of DNA strand hybridization metrics...AFRL-RI-RS-TR-2009-131 Final Technical Report May 2009 MULTI-THREADED DNA TAG/ANTI-TAG LIBRARY GENERATOR FOR MULTI-CORE PLATFORMS...TYPE Final 3. DATES COVERED (From - To) Jun 08 – Feb 09 4. TITLE AND SUBTITLE MULTI-THREADED DNA TAG/ANTI-TAG LIBRARY GENERATOR FOR MULTI-CORE
Global Climate Change: Valuable Insights from Concordant and Discordant Ice Core Histories
NASA Astrophysics Data System (ADS)
Mosley-Thompson, E.; Thompson, L. G.; Porter, S. E.; Goodwin, B. P.; Wilson, A. B.
2014-12-01
Earth's ice cover is responding to the ongoing large-scale warming driven in part by anthropogenic forces. The highest tropical and subtropical ice fields are dramatically shrinking and/or thinning and unique climate histories archived therein are now threatened, compromised or lost. Many ice fields in higher latitudes are also experiencing and recording climate system changes although these are often manifested in less evident and spectacular ways. The Antarctic Peninsula (AP) has experienced a rapid, widespread and dramatic warming over the last 60 years. Carefully selected ice fields in the AP allow reconstruction of long histories of key climatic variables. As more proxy climate records are recovered it is clear they reflect a combination of expected and unexpected responses to seemingly similar climate forcings. Recently acquired temperature and precipitation histories from the Bruce Plateau are examined within the context provided by other cores recently collected in the AP. Understanding the differences and similarities among these records provides a better understanding of the forces driving climate variability in the AP over the last century. The Arctic is also rapidly warming. The δ18O records from the Bona-Churchill and Mount Logan ice cores from southeast Alaska and southwest Yukon Territory, respectively, do not record this strong warming. The Aleutian Low strongly influences moisture transport to this geographically complex region, yet its interannual variability is preserved differently in these cores located just 110 km apart. Mount Logan is very sensitive to multi-decadal to multi-centennial climate shifts in the tropical Pacific while low frequency variability on Bona-Churchill is more strongly connected to Western Arctic sea ice extent. There is a natural tendency to focus more strongly on commonalities among records, particularly on regional scales. However, it is also important to investigate seemingly poorly correlated records, particularly those from geographically complex settings that appear to be dominated by similar large-scale climatological processes. Better understanding of the spatially and temporally diverse responses in such regions will expand our understanding of the mechanisms forcing climate variability in meteorologically complex environments.
Kumarapeli, Pushpa; de Lusignan, Simon
2013-06-01
Electronic patient record (EPR) systems are widely used. This study explores the context and use of systems to provide insights into improving their use in clinical practice. We used video to observe 163 consultations by 16 clinicians using four EPR brands. We made a visual study of the consultation room and coded interactions between clinician, patient, and computer. Few patients (6.9%, n=12) declined to participate. Patients looked at the computer twice as much (47.6 s vs 20.6 s, p<0.001) when it was within their gaze. A quarter of consultations were interrupted (27.6%, n=45); and in half the clinician left the room (12.3%, n=20). The core consultation takes about 87% of the total session time; 5% of time is spent pre-consultation, reading the record and calling the patient in; and 8% of time is spent post-consultation, largely entering notes. Consultations with more than one person and where prescribing took place were longer (R(2) adj=22.5%, p<0.001). The core consultation can be divided into 61% of direct clinician-patient interaction, of which 15% is examination, 25% computer use with no patient involvement, and 14% simultaneous clinician-computer-patient interplay. The proportions of computer use are similar between consultations (mean=40.6%, SD=13.7%). There was more data coding in problem-orientated EPR systems, though clinicians often used vague codes. The EPR system is used for a consistent proportion of the consultation and should be designed to facilitate multi-tasking. Clinicians who want to promote screen sharing should change their consulting room layout.
Kumarapeli, Pushpa; de Lusignan, Simon
2013-01-01
Background and objective Electronic patient record (EPR) systems are widely used. This study explores the context and use of systems to provide insights into improving their use in clinical practice. Methods We used video to observe 163 consultations by 16 clinicians using four EPR brands. We made a visual study of the consultation room and coded interactions between clinician, patient, and computer. Few patients (6.9%, n=12) declined to participate. Results Patients looked at the computer twice as much (47.6 s vs 20.6 s, p<0.001) when it was within their gaze. A quarter of consultations were interrupted (27.6%, n=45); and in half the clinician left the room (12.3%, n=20). The core consultation takes about 87% of the total session time; 5% of time is spent pre-consultation, reading the record and calling the patient in; and 8% of time is spent post-consultation, largely entering notes. Consultations with more than one person and where prescribing took place were longer (R2 adj=22.5%, p<0.001). The core consultation can be divided into 61% of direct clinician–patient interaction, of which 15% is examination, 25% computer use with no patient involvement, and 14% simultaneous clinician–computer–patient interplay. The proportions of computer use are similar between consultations (mean=40.6%, SD=13.7%). There was more data coding in problem-orientated EPR systems, though clinicians often used vague codes. Conclusions The EPR system is used for a consistent proportion of the consultation and should be designed to facilitate multi-tasking. Clinicians who want to promote screen sharing should change their consulting room layout. PMID:23242763
System Framework for a Multi-Band, Multi-Mode Software Defined Radio
2014-06-01
detection, while the VITA Radio Transport ( VRT ) protocol over Gigabit Ethernet (GIGE) is implemented for the data interface. In addition to the SoC...CTRL VGA CTRL C2 GPP C2 CORE SW ARM0 RX SYN CTRL PL MEMORY MAP DR CTRL GENERIC INTERRUPT CONTROLLER DR GPP VITERBI ALGORITHM & VRT INTERFACE ARM1
Luminescence emission from nonpolar Al0.3Ga0.7N/GaN core-shell and core-multi-shell nanowires
NASA Astrophysics Data System (ADS)
Namvari, E.; Shojaei, S.; Asgari, A.
2017-12-01
In the present work, we theoretically study the possibility of luminescence emission from two systems of nonpolar Al0.3Ga0.7N/GaN Core-shell and core-multi-shell c-axis oriented nanowires with hexagonal cross section. To obtain energy levels and wave functions through the solution of Schrodinger-Poisson equations, numerical Self-consistent procedure has been employed. N-type doping has been considered to investigate the two-dimensional electron gas formation and its effect on luminescence. The detailed analysis of the results as a function of the various structural parameters has been carried out. The results presents an examination of the band to band luminescence feature and its changes with involved parameters. We found that the size of the system determines the feature of luminescence emission. As main finding, our calculations show that the intensity of luminescence spectrum in facet to facet route of NW cross section is significant than that of corner to corner route. In addition, no shift of the peak position is observed with changing the amount of doping. Our numerical calculations give more insights into the luminescence emission of nonpolar GaN/AlGaN core/shell nanowire and have many implications in experiment.
Network Coding on Heterogeneous Multi-Core Processors for Wireless Sensor Networks
Kim, Deokho; Park, Karam; Ro, Won W.
2011-01-01
While network coding is well known for its efficiency and usefulness in wireless sensor networks, the excessive costs associated with decoding computation and complexity still hinder its adoption into practical use. On the other hand, high-performance microprocessors with heterogeneous multi-cores would be used as processing nodes of the wireless sensor networks in the near future. To this end, this paper introduces an efficient network coding algorithm developed for the heterogenous multi-core processors. The proposed idea is fully tested on one of the currently available heterogeneous multi-core processors referred to as the Cell Broadband Engine. PMID:22164053
DPM — efficient storage in diverse environments
NASA Astrophysics Data System (ADS)
Hellmich, Martin; Furano, Fabrizio; Smith, David; Brito da Rocha, Ricardo; Álvarez Ayllón, Alejandro; Manzi, Andrea; Keeble, Oliver; Calvet, Ivan; Regala, Miguel Antonio
2014-06-01
Recent developments, including low power devices, cluster file systems and cloud storage, represent an explosion in the possibilities for deploying and managing grid storage. In this paper we present how different technologies can be leveraged to build a storage service with differing cost, power, performance, scalability and reliability profiles, using the popular storage solution Disk Pool Manager (DPM/dmlite) as the enabling technology. The storage manager DPM is designed for these new environments, allowing users to scale up and down as they need it, and optimizing their computing centers energy efficiency and costs. DPM runs on high-performance machines, profiting from multi-core and multi-CPU setups. It supports separating the database from the metadata server, the head node, largely reducing its hard disk requirements. Since version 1.8.6, DPM is released in EPEL and Fedora, simplifying distribution and maintenance, but also supporting the ARM architecture beside i386 and x86_64, allowing it to run the smallest low-power machines such as the Raspberry Pi or the CuBox. This usage is facilitated by the possibility to scale horizontally using a main database and a distributed memcached-powered namespace cache. Additionally, DPM supports a variety of storage pools in the backend, most importantly HDFS, S3-enabled storage, and cluster file systems, allowing users to fit their DPM installation exactly to their needs. In this paper, we investigate the power-efficiency and total cost of ownership of various DPM configurations. We develop metrics to evaluate the expected performance of a setup both in terms of namespace and disk access considering the overall cost including equipment, power consumptions, or data/storage fees. The setups tested range from the lowest scale using Raspberry Pis with only 700MHz single cores and a 100Mbps network connections, over conventional multi-core servers to typical virtual machine instances in cloud settings. We evaluate the combinations of different name server setups, for example load-balanced clusters, with different storage setups, from using a classic local configuration to private and public clouds.
Mesomorphic properties of multi-arm chenodeoxycholic acid-derived liquid crystals
NASA Astrophysics Data System (ADS)
Dong, Liang; Yao, Miao; Wu, Shuang-jie; Yao, Dan-Shu; Hu, Jian-She; He, Xiao-zhi; Tian, Mei
2017-12-01
Four multi-arm liquid crystals (LCs) based on chenodeoxycholic acid, termed as 2G-PD, 2G-IB, 2G-BD and 5G-GC, respectively, have been synthesised by convergent method, which nematic LC, 6-(4-((4-ethoxybenzoyl)oxy)phenoxy)-6-oxohexanoic acid, was used as side arm, and chenodeoxycholic acid (CDCA) was used as the first core, 1,2-propanediol (PD), isosorbide (IB), 4,4‧-biphenyldiol (BD) and glucose (GC) were used as the second core, respectively. The first generation product, CDCA2EA, displayed cholesteric phase. The second generation products 2G-BD and 5G-GC displayed cholesteric phase, while 2G-PD and 2G-IB exhibited nematic phase. The multi-arm LC, 2G-IB, did not display cholesteric phase although the two cores were all chiral ones. The result indicated that chirality of the second core sometimes made the multi-arm LCs display nematic phase when cholesteric CDCA-derivative were introduced into the second core. Some attention should be paid on molecular conformation besides the introduction of chiral cores for multi-chiral-core LCs to obtain cholesteric phase.
Experimental Analysis of File Transfer Rates over Wide-Area Dedicated Connections
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rao, Nageswara S.; Liu, Qiang; Sen, Satyabrata
2016-12-01
File transfers over dedicated connections, supported by large parallel file systems, have become increasingly important in high-performance computing and big data workflows. It remains a challenge to achieve peak rates for such transfers due to the complexities of file I/O, host, and network transport subsystems, and equally importantly, their interactions. We present extensive measurements of disk-to-disk file transfers using Lustre and XFS file systems mounted on multi-core servers over a suite of 10 Gbps emulated connections with 0-366 ms round trip times. Our results indicate that large buffer sizes and many parallel flows do not always guarantee high transfer rates.more » Furthermore, large variations in the measured rates necessitate repeated measurements to ensure confidence in inferences based on them. We propose a new method to efficiently identify the optimal joint file I/O and network transport parameters using a small number of measurements. We show that for XFS and Lustre with direct I/O, this method identifies configurations achieving 97% of the peak transfer rate while probing only 12% of the parameter space.« less
Investigation of Large Scale Cortical Models on Clustered Multi-Core Processors
2013-02-01
with the bias node ( gray ) denoted as ww and the weights associated with the remaining first layer nodes (black) denoted as W. In forming the overall...Implementation of RBF network on GPU Platform 3.5.1 The Cholesky decomposition algorithm We need to invert the matrix multiplication GTG to
Scaling GDL for Multi-cores to Process Planck HFI Beams Monte Carlo on HPC
NASA Astrophysics Data System (ADS)
Coulais, A.; Schellens, M.; Duvert, G.; Park, J.; Arabas, S.; Erard, S.; Roudier, G.; Hivon, E.; Mottet, S.; Laurent, B.; Pinter, M.; Kasradze, N.; Ayad, M.
2014-05-01
After reviewing the majors progress done in GDL -now in 0.9.4- on performance and plotting capabilities since ADASS XXI paper (Coulais et al. 2012), we detail how a large code for Planck HFI beams Monte Carlo was successfully transposed from IDL to GDL on HPC.
Clinical terminology support for a national ambulatory practice outcomes research network.
Ricciardi, Thomas N; Lieberman, Michael I; Kahn, Michael G; Masarie, F E
2005-01-01
The Medical Quality Improvement Consortium (MQIC) is a nationwide collaboration of 74 healthcare delivery systems, consisting of 3755 clinicians, who contribute de-identified clinical data from the same commercial electronic medical record (EMR) for quality reporting, outcomes research and clinical research in public health and practice benchmarking. Despite the existence of a common, centrally-managed, shared terminology for core concepts (medications, problem lists, observation names), a substantial "back-end" information management process is required to ensure terminology and data harmonization for creating multi-facility clinically-acceptable queries and comparable results. We describe the information architecture created to support terminology harmonization across this data-sharing consortium and discuss the implications for large scale data sharing envisioned by proponents for the national adoption of ambulatory EMR systems.
Clinical Terminology Support for a National Ambulatory Practice Outcomes Research Network
Ricciardi, Thomas N.; Lieberman, Michael I.; Kahn, Michael G.; Masarie, F.E. “Chip”
2005-01-01
The Medical Quality Improvement Consortium (MQIC) is a nationwide collaboration of 74 healthcare delivery systems, consisting of 3755 clinicians, who contribute de-identified clinical data from the same commercial electronic medical record (EMR) for quality reporting, outcomes research and clinical research in public health and practice benchmarking. Despite the existence of a common, centrally-managed, shared terminology for core concepts (medications, problem lists, observation names), a substantial “back-end” information management process is required to ensure terminology and data harmonization for creating multi-facility clinically-acceptable queries and comparable results. We describe the information architecture created to support terminology harmonization across this data-sharing consortium and discuss the implications for large scale data sharing envisioned by proponents for the national adoption of ambulatory EMR systems. PMID:16779116
NASA Astrophysics Data System (ADS)
Ryu, Hoon; Jeong, Yosang; Kang, Ji-Hoon; Cho, Kyu Nam
2016-12-01
Modelling of multi-million atomic semiconductor structures is important as it not only predicts properties of physically realizable novel materials, but can accelerate advanced device designs. This work elaborates a new Technology-Computer-Aided-Design (TCAD) tool for nanoelectronics modelling, which uses a sp3d5s∗ tight-binding approach to describe multi-million atomic structures, and simulate electronic structures with high performance computing (HPC), including atomic effects such as alloy and dopant disorders. Being named as Quantum simulation tool for Advanced Nanoscale Devices (Q-AND), the tool shows nice scalability on traditional multi-core HPC clusters implying the strong capability of large-scale electronic structure simulations, particularly with remarkable performance enhancement on latest clusters of Intel Xeon PhiTM coprocessors. A review of the recent modelling study conducted to understand an experimental work of highly phosphorus-doped silicon nanowires, is presented to demonstrate the utility of Q-AND. Having been developed via Intel Parallel Computing Center project, Q-AND will be open to public to establish a sound framework of nanoelectronics modelling with advanced HPC clusters of a many-core base. With details of the development methodology and exemplary study of dopant electronics, this work will present a practical guideline for TCAD development to researchers in the field of computational nanoelectronics.
Bornstein, Marc H.; Hahn, Chun-Shin; Putnick, Diane L.; Suwalsky, Joan T. D.
2014-01-01
This four-wave prospective longitudinal study evaluated stability of language in 324 children from early childhood to adolescence. Structural equation modeling supported loadings of multiple age-appropriate multi-source measures of child language on single-factor core language skills at 20 months and 4, 10, and 14 years. Large stability coefficients (standardized indirect effect = .46) were obtained between language latent variables from early childhood to adolescence and accounting for child nonverbal intelligence and social competence and maternal verbal intelligence, education, speech, and social desirability. Stability coefficients were similar for girls and boys. Stability of core language skill was stronger from 4 to 10 to 14 years than from 20 months to 4 years, so early intervention to improve lagging language is recommended. PMID:25165797
Electrosprayed Multi-Core Alginate Microcapsules as Novel Self-Healing Containers
NASA Astrophysics Data System (ADS)
Hia, Iee Lee; Pasbakhsh, Pooria; Chan, Eng-Seng; Chai, Siang-Piao
2016-10-01
Alginate microcapsules containing epoxy resin were developed through electrospraying method and embedded into epoxy matrix to produce a capsule-based self-healing composite system. These formaldehyde free alginate/epoxy microcapsules were characterized via light microscope, field emission scanning electron microscope, fourier transform infrared spectroscopy and thermogravimetric analysis. Results showed that epoxy resin was successfully encapsulated within alginate matrix to form porous (multi-core) microcapsules with pore size ranged from 5-100 μm. The microcapsules had an average size of 320 ± 20 μm with decomposition temperature at 220 °C. The loading capacity of these capsules was estimated to be 79%. Under in situ healing test, impact specimens showed healing efficiency as high as 86% and the ability to heal up to 3 times due to the multi-core capsule structure and the high impact energy test that triggered the released of epoxy especially in the second and third healings. TDCB specimens showed one-time healing only with the highest healing efficiency of 76%. The single healing event was attributed by the constant crack propagation rate of TDCB fracture test. For the first time, a cost effective, environmentally benign and sustainable capsule-based self-healing system with multiple healing capabilities and high healing performance was developed.
Electrosprayed Multi-Core Alginate Microcapsules as Novel Self-Healing Containers.
Hia, Iee Lee; Pasbakhsh, Pooria; Chan, Eng-Seng; Chai, Siang-Piao
2016-10-03
Alginate microcapsules containing epoxy resin were developed through electrospraying method and embedded into epoxy matrix to produce a capsule-based self-healing composite system. These formaldehyde free alginate/epoxy microcapsules were characterized via light microscope, field emission scanning electron microscope, fourier transform infrared spectroscopy and thermogravimetric analysis. Results showed that epoxy resin was successfully encapsulated within alginate matrix to form porous (multi-core) microcapsules with pore size ranged from 5-100 μm. The microcapsules had an average size of 320 ± 20 μm with decomposition temperature at 220 °C. The loading capacity of these capsules was estimated to be 79%. Under in situ healing test, impact specimens showed healing efficiency as high as 86% and the ability to heal up to 3 times due to the multi-core capsule structure and the high impact energy test that triggered the released of epoxy especially in the second and third healings. TDCB specimens showed one-time healing only with the highest healing efficiency of 76%. The single healing event was attributed by the constant crack propagation rate of TDCB fracture test. For the first time, a cost effective, environmentally benign and sustainable capsule-based self-healing system with multiple healing capabilities and high healing performance was developed.
Electrosprayed Multi-Core Alginate Microcapsules as Novel Self-Healing Containers
Hia, Iee Lee; Pasbakhsh, Pooria; Chan, Eng-Seng; Chai, Siang-Piao
2016-01-01
Alginate microcapsules containing epoxy resin were developed through electrospraying method and embedded into epoxy matrix to produce a capsule-based self-healing composite system. These formaldehyde free alginate/epoxy microcapsules were characterized via light microscope, field emission scanning electron microscope, fourier transform infrared spectroscopy and thermogravimetric analysis. Results showed that epoxy resin was successfully encapsulated within alginate matrix to form porous (multi-core) microcapsules with pore size ranged from 5–100 μm. The microcapsules had an average size of 320 ± 20 μm with decomposition temperature at 220 °C. The loading capacity of these capsules was estimated to be 79%. Under in situ healing test, impact specimens showed healing efficiency as high as 86% and the ability to heal up to 3 times due to the multi-core capsule structure and the high impact energy test that triggered the released of epoxy especially in the second and third healings. TDCB specimens showed one-time healing only with the highest healing efficiency of 76%. The single healing event was attributed by the constant crack propagation rate of TDCB fracture test. For the first time, a cost effective, environmentally benign and sustainable capsule-based self-healing system with multiple healing capabilities and high healing performance was developed. PMID:27694922
NASA Technical Reports Server (NTRS)
Lamarque, J.-F.; Dentener, F.; McConnell, J.; Ro, C.-U.; Shaw, M.; Vet, R.; Bergmann, D.; Cameron-Smith, P.; Doherty, R.; Faluvegi, G.;
2013-01-01
We present multi-model global datasets of nitrogen and sulfate deposition covering time periods from 1850 to 2100, calculated within the Atmospheric Chemistry and Climate Model Intercomparison Project (ACCMIP). The computed deposition fluxes are compared to surface wet deposition and ice-core measurements. We use a new dataset of wet deposition for 2000-2002 based on critical assessment of the quality of existing regional network data. We show that for present-day (year 2000 ACCMIP time-slice), the ACCMIP results perform similarly to previously published multi-model assessments. For this time slice, we find a multi-model mean deposition of 50 Tg(N) yr1 from nitrogen oxide emissions, 60 Tg(N) yr1 from ammonia emissions, and 83 Tg(S) yr1 from sulfur emissions. The analysis of changes between 1980 and 2000 indicates significant differences between model and measurements over the United States but less so over Europe. This difference points towards misrepresentation of 1980 NH3 emissions over North America. Based on ice-core records, the 1850 deposition fluxes agree well with Greenland ice cores but the change between 1850 and 2000 seems to be overestimated in the Northern Hemisphere for both nitrogen and sulfur species. Using the Representative Concentration Pathways to define the projected climate and atmospheric chemistry related emissions and concentrations, we find large regional nitrogen deposition increases in 2100 in Latin America, Africa and parts of Asia under some of the scenarios considered. Increases in South Asia are especially large, and are seen in all scenarios, with 2100 values more than double 2000 in some scenarios and reaching 1300 mg(N) m2 yr1 averaged over regional to continental scale regions in RCP 2.6 and 8.5, 3050 larger than the values in any region currently (2000). The new ACCMIP deposition dataset provides novel, consistent and evaluated global gridded deposition fields for use in a wide range of climate and ecological studies.
MouseNet database: digital management of a large-scale mutagenesis project.
Pargent, W; Heffner, S; Schäble, K F; Soewarto, D; Fuchs, H; Hrabé de Angelis, M
2000-07-01
The Munich ENU Mouse Mutagenesis Screen is a large-scale mutant production, phenotyping, and mapping project. It encompasses two animal breeding facilities and a number of screening groups located in the general area of Munich. A central database is required to manage and process the immense amount of data generated by the mutagenesis project. This database, which we named MouseNet(c), runs on a Sybase platform and will finally store and process all data from the entire project. In addition, the system comprises a portfolio of functions needed to support the workflow management of the core facility and the screening groups. MouseNet(c) will make all of the data available to the participating screening groups, and later to the international scientific community. MouseNet(c) will consist of three major software components:* Animal Management System (AMS)* Sample Tracking System (STS)* Result Documentation System (RDS)MouseNet(c) provides the following major advantages:* being accessible from different client platforms via the Internet* being a full-featured multi-user system (including access restriction and data locking mechanisms)* relying on a professional RDBMS (relational database management system) which runs on a UNIX server platform* supplying workflow functions and a variety of plausibility checks.
Kaur, Gaganpreet; Kaur, Maninder; Silakari, Om
2014-01-01
The recent research area endeavors to discover ultimate multi-target ligands, an increasingly feasible and attractive alternative to existing mono-targeted drugs for treatment of complex, multi-factorial inflammation process which underlays plethora of debilitated health conditions. In order to improvise this option, exploration of relevant chemical core scaffold will be an utmost need. Privileged benzimidazole scaffold being historically versatile structural motif could offer a viable starting point in the search for novel multi-target ligands against multi-factorial inflammation process since, when appropriately substituted, it can selectively modulate diverse receptors, pathways and enzymes associated with the pathogenesis of inflammation. Despite this remarkable capability, the multi-target capacity of the benzimidazole scaffold remains largely unexploited. With this in focus, the present review article attempts to provide synopsis of published research to exemplify the valuable use of benzimidazole nucleus and focus on their suitability as starting scaffold to develop multi-targeted anti-inflammatory ligands.
Simeonov, Plamen L
2017-12-01
The goal of this paper is to advance an extensible theory of living systems using an approach to biomathematics and biocomputation that suitably addresses self-organized, self-referential and anticipatory systems with multi-temporal multi-agents. Our first step is to provide foundations for modelling of emergent and evolving dynamic multi-level organic complexes and their sustentative processes in artificial and natural life systems. Main applications are in life sciences, medicine, ecology and astrobiology, as well as robotics, industrial automation, man-machine interface and creative design. Since 2011 over 100 scientists from a number of disciplines have been exploring a substantial set of theoretical frameworks for a comprehensive theory of life known as Integral Biomathics. That effort identified the need for a robust core model of organisms as dynamic wholes, using advanced and adequately computable mathematics. The work described here for that core combines the advantages of a situation and context aware multivalent computational logic for active self-organizing networks, Wandering Logic Intelligence (WLI), and a multi-scale dynamic category theory, Memory Evolutive Systems (MES), hence WLIMES. This is presented to the modeller via a formal augmented reality language as a first step towards practical modelling and simulation of multi-level living systems. Initial work focuses on the design and implementation of this visual language and calculus (VLC) and its graphical user interface. The results will be integrated within the current methodology and practices of theoretical biology and (personalized) medicine to deepen and to enhance the holistic understanding of life. Copyright © 2017 Elsevier B.V. All rights reserved.
ERIC Educational Resources Information Center
Prager, Carolyn; And Others
The education and reeducation of health care professionals remain essential, if somewhat neglected, elements in reforming the nation's health care system. The Pew Health Professions Commission (PHPC) has made the reform of health care contingent upon the reform of education, urging educational institutions to design core curricula with…
Small Launch Vehicle Design Approaches: Clustered Cores Compared with Multi-Stage Inline Concepts
NASA Technical Reports Server (NTRS)
Waters, Eric D.; Beers, Benjamin; Esther, Elizabeth; Philips, Alan; Threet, Grady E., Jr.
2013-01-01
In an effort to better define small launch vehicle design options two approaches were investigated from the small launch vehicle trade space. The primary focus was to evaluate a clustered common core design against a purpose built inline vehicle. Both designs focused on liquid oxygen (LOX) and rocket propellant grade kerosene (RP-1) stages with the terminal stage later evaluated as a LOX/methane (CH4) stage. A series of performance optimization runs were done in order to minimize gross liftoff weight (GLOW) including alternative thrust levels, delivery altitude for payload, vehicle length to diameter ratio, alternative engine feed systems, re-evaluation of mass growth allowances, passive versus active guidance systems, and rail and tower launch methods. Additionally manufacturability, cost, and operations also play a large role in the benefits and detriments for each design. Presented here is the Advanced Concepts Office's Earth to Orbit Launch Team methodology and high level discussion of the performance trades and trends of both small launch vehicle solutions along with design philosophies that shaped both concepts. Without putting forth a decree stating one approach is better than the other; this discussion is meant to educate the community at large and let the reader determine which architecture is truly the most economical; since each path has such a unique set of limitations and potential payoffs.
Hunter, Adam; Dayalan, Saravanan; De Souza, David; Power, Brad; Lorrimar, Rodney; Szabo, Tamas; Nguyen, Thu; O'Callaghan, Sean; Hack, Jeremy; Pyke, James; Nahid, Amsha; Barrero, Roberto; Roessner, Ute; Likic, Vladimir; Tull, Dedreia; Bacic, Antony; McConville, Malcolm; Bellgard, Matthew
2017-01-01
An increasing number of research laboratories and core analytical facilities around the world are developing high throughput metabolomic analytical and data processing pipelines that are capable of handling hundreds to thousands of individual samples per year, often over multiple projects, collaborations and sample types. At present, there are no Laboratory Information Management Systems (LIMS) that are specifically tailored for metabolomics laboratories that are capable of tracking samples and associated metadata from the beginning to the end of an experiment, including data processing and archiving, and which are also suitable for use in large institutional core facilities or multi-laboratory consortia as well as single laboratory environments. Here we present MASTR-MS, a downloadable and installable LIMS solution that can be deployed either within a single laboratory or used to link workflows across a multisite network. It comprises a Node Management System that can be used to link and manage projects across one or multiple collaborating laboratories; a User Management System which defines different user groups and privileges of users; a Quote Management System where client quotes are managed; a Project Management System in which metadata is stored and all aspects of project management, including experimental setup, sample tracking and instrument analysis, are defined, and a Data Management System that allows the automatic capture and storage of raw and processed data from the analytical instruments to the LIMS. MASTR-MS is a comprehensive LIMS solution specifically designed for metabolomics. It captures the entire lifecycle of a sample starting from project and experiment design to sample analysis, data capture and storage. It acts as an electronic notebook, facilitating project management within a single laboratory or a multi-node collaborative environment. This software is being developed in close consultation with members of the metabolomics research community. It is freely available under the GNU GPL v3 licence and can be accessed from, https://muccg.github.io/mastr-ms/.
NASA Astrophysics Data System (ADS)
Wang, Yazhou; Zhang, Yiqiong; Wang, Bochu; Cao, Yang; Yu, Qingsong; Yin, Tieying
2013-06-01
The study aimed at constructing a novel drug delivery system for programmable multiple drug release controlled with core-shell structure. The core-shell structure consisted of chitosan nanoparticles as core and polyvinylpyrrolidone micro/nanocoating as shell to form core-shell micro/nanoparticles, which was fabricated by ionic gelation and emulsion electrospray methods. As model drug agents, Naproxen and rhodamine B were encapsulated in the core and shell regions, respectively. The core-shell micro/nanoparticles thus fabricated were characterized and confirmed by scanning electron microscope, transmission electron microscope, and fluorescence optical microscope. The core-shell micro/nanoparticles showed good release controllability through drug release experiment in vitro. It was noted that a programmable release pattern for dual drug agents was also achieved by adjusting their loading regions in the core-shell structures. The results indicate that emulsion electrospraying technology is a promising approach in fabrication of core-shell micro/nanoparticles for programmable dual drug release. Such a novel multi-drug delivery system has a potential application for the clinical treatment of cancer, tuberculosis, and tissue engineering.
Teodoro, George; Kurc, Tahsin; Andrade, Guilherme; Kong, Jun; Ferreira, Renato; Saltz, Joel
2015-01-01
We carry out a comparative performance study of multi-core CPUs, GPUs and Intel Xeon Phi (Many Integrated Core-MIC) with a microscopy image analysis application. We experimentally evaluate the performance of computing devices on core operations of the application. We correlate the observed performance with the characteristics of computing devices and data access patterns, computation complexities, and parallelization forms of the operations. The results show a significant variability in the performance of operations with respect to the device used. The performances of operations with regular data access are comparable or sometimes better on a MIC than that on a GPU. GPUs are more efficient than MICs for operations that access data irregularly, because of the lower bandwidth of the MIC for random data accesses. We propose new performance-aware scheduling strategies that consider variabilities in operation speedups. Our scheduling strategies significantly improve application performance compared to classic strategies in hybrid configurations. PMID:28239253
Topical perspective on massive threading and parallelism.
Farber, Robert M
2011-09-01
Unquestionably computer architectures have undergone a recent and noteworthy paradigm shift that now delivers multi- and many-core systems with tens to many thousands of concurrent hardware processing elements per workstation or supercomputer node. GPGPU (General Purpose Graphics Processor Unit) technology in particular has attracted significant attention as new software development capabilities, namely CUDA (Compute Unified Device Architecture) and OpenCL™, have made it possible for students as well as small and large research organizations to achieve excellent speedup for many applications over more conventional computing architectures. The current scientific literature reflects this shift with numerous examples of GPGPU applications that have achieved one, two, and in some special cases, three-orders of magnitude increased computational performance through the use of massive threading to exploit parallelism. Multi-core architectures are also evolving quickly to exploit both massive-threading and massive-parallelism such as the 1.3 million threads Blue Waters supercomputer. The challenge confronting scientists in planning future experimental and theoretical research efforts--be they individual efforts with one computer or collaborative efforts proposing to use the largest supercomputers in the world is how to capitalize on these new massively threaded computational architectures--especially as not all computational problems will scale to massive parallelism. In particular, the costs associated with restructuring software (and potentially redesigning algorithms) to exploit the parallelism of these multi- and many-threaded machines must be considered along with application scalability and lifespan. This perspective is an overview of the current state of threading and parallelize with some insight into the future. Published by Elsevier Inc.
Traditional Tracking with Kalman Filter on Parallel Architectures
NASA Astrophysics Data System (ADS)
Cerati, Giuseppe; Elmer, Peter; Lantz, Steven; MacNeill, Ian; McDermott, Kevin; Riley, Dan; Tadel, Matevž; Wittich, Peter; Würthwein, Frank; Yagil, Avi
2015-05-01
Power density constraints are limiting the performance improvements of modern CPUs. To address this, we have seen the introduction of lower-power, multi-core processors, but the future will be even more exciting. In order to stay within the power density limits but still obtain Moore's Law performance/price gains, it will be necessary to parallelize algorithms to exploit larger numbers of lightweight cores and specialized functions like large vector units. Example technologies today include Intel's Xeon Phi and GPGPUs. Track finding and fitting is one of the most computationally challenging problems for event reconstruction in particle physics. At the High Luminosity LHC, for example, this will be by far the dominant problem. The most common track finding techniques in use today are however those based on the Kalman Filter. Significant experience has been accumulated with these techniques on real tracking detector systems, both in the trigger and offline. We report the results of our investigations into the potential and limitations of these algorithms on the new parallel hardware.
NASA Astrophysics Data System (ADS)
Okamoto, Taro; Takenaka, Hiroshi; Nakamura, Takeshi; Aoki, Takayuki
2010-12-01
We adopted the GPU (graphics processing unit) to accelerate the large-scale finite-difference simulation of seismic wave propagation. The simulation can benefit from the high-memory bandwidth of GPU because it is a "memory intensive" problem. In a single-GPU case we achieved a performance of about 56 GFlops, which was about 45-fold faster than that achieved by a single core of the host central processing unit (CPU). We confirmed that the optimized use of fast shared memory and registers were essential for performance. In the multi-GPU case with three-dimensional domain decomposition, the non-contiguous memory alignment in the ghost zones was found to impose quite long time in data transfer between GPU and the host node. This problem was solved by using contiguous memory buffers for ghost zones. We achieved a performance of about 2.2 TFlops by using 120 GPUs and 330 GB of total memory: nearly (or more than) 2200 cores of host CPUs would be required to achieve the same performance. The weak scaling was nearly proportional to the number of GPUs. We therefore conclude that GPU computing for large-scale simulation of seismic wave propagation is a promising approach as a faster simulation is possible with reduced computational resources compared to CPUs.
NASA Astrophysics Data System (ADS)
Carrasco, E.; Sánchez-Blanco, E.; García-Vargas, M. L.; Gil de Paz, A.; Páez, G.; Gallego, J.; Sánchez, F. M.; Vílchez, J. M.
2012-09-01
MEGARA is the next optical Integral-Field Unit (IFU) and Multi-Object Spectrograph (MOS) for Gran Telescopio Canarias. The instrument offers two IFUs plus a Multi-Object Spectroscopy (MOS) mode: a large compact bundle covering 12.5 arcsec x 11.3 arcsec on sky with 100 μm fiber-core; a small compact bundle, of 8.5 arcsec x 6.7 arcsec with 70 μm fiber-core and a fiber MOS positioner that allows to place up to 100 mini-bundles, 7 fibers each, with 100 μm fiber-core, within a 3.5 arcmin x 3.5 arcmin field of view, around the two IFUs. The fibers, organized in bundles, end in the pseudo-slit plate, which will be placed at the entrance focal plane of the MEGARA spectrograph. The large IFU and MOS modes will provide intermediate to high spectral resolutions, R=6800-17000. The small IFU mode will provide R=8000-20000. All these resolutions are possible thanks to a spectrograph design based in the used of volume phase holographic gratings in combination with prisms to keep fixed the collimator and camera angle. The MEGARA optics is composed by a total of 53 large optical elements per spectrograph: the field lens, the collimator and the camera lenses plus the complete set of pupil elements including holograms, windows and prisms. INAOE, a partner of the GTC and a partner of MEGARA consortium, is responsible of the optics manufacturing and tests. INAOE will carry out this project working in an alliance with CIO. This paper summarizes the status of MEGARA spectrograph optics at the Preliminary Design Review, held on March 2012.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jason Heath; Brian McPherson; Thomas Dewers
The assessment of caprocks for geologic CO{sub 2} storage is a multi-scale endeavor. Investigation of a regional caprock - the Kirtland Formation, San Juan Basin, USA - at the pore-network scale indicates high capillary sealing capacity and low permeabilities. Core and wellscale data, however, indicate a potential seal bypass system as evidenced by multiple mineralized fractures and methane gas saturations within the caprock. Our interpretation of {sup 4}He concentrations, measured at the top and bottom of the caprock, suggests low fluid fluxes through the caprock: (1) Of the total {sup 4}He produced in situ (i.e., at the locations of sampling)more » by uranium and thorium decay since deposition of the Kirtland Formation, a large portion still resides in the pore fluids. (2) Simple advection-only and advection-diffusion models, using the measured {sup 4}He concentrations, indicate low permeability ({approx}10-20 m{sup 2} or lower) for the thickness of the Kirtland Formation. These findings, however, do not guarantee the lack of a large-scale bypass system. The measured data, located near the boundary conditions of the models (i.e., the overlying and underlying aquifers), limit our testing of conceptual models and the sensitivity of model parameterization. Thus, we suggest approaches for future studies to better assess the presence or lack of a seal bypass system at this particular site and for other sites in general.« less
Adaptive multi-GPU Exchange Monte Carlo for the 3D Random Field Ising Model
NASA Astrophysics Data System (ADS)
Navarro, Cristóbal A.; Huang, Wei; Deng, Youjin
2016-08-01
This work presents an adaptive multi-GPU Exchange Monte Carlo approach for the simulation of the 3D Random Field Ising Model (RFIM). The design is based on a two-level parallelization. The first level, spin-level parallelism, maps the parallel computation as optimal 3D thread-blocks that simulate blocks of spins in shared memory with minimal halo surface, assuming a constant block volume. The second level, replica-level parallelism, uses multi-GPU computation to handle the simulation of an ensemble of replicas. CUDA's concurrent kernel execution feature is used in order to fill the occupancy of each GPU with many replicas, providing a performance boost that is more notorious at the smallest values of L. In addition to the two-level parallel design, the work proposes an adaptive multi-GPU approach that dynamically builds a proper temperature set free of exchange bottlenecks. The strategy is based on mid-point insertions at the temperature gaps where the exchange rate is most compromised. The extra work generated by the insertions is balanced across the GPUs independently of where the mid-point insertions were performed. Performance results show that spin-level performance is approximately two orders of magnitude faster than a single-core CPU version and one order of magnitude faster than a parallel multi-core CPU version running on 16-cores. Multi-GPU performance is highly convenient under a weak scaling setting, reaching up to 99 % efficiency as long as the number of GPUs and L increase together. The combination of the adaptive approach with the parallel multi-GPU design has extended our possibilities of simulation to sizes of L = 32 , 64 for a workstation with two GPUs. Sizes beyond L = 64 can eventually be studied using larger multi-GPU systems.
ERIC Educational Resources Information Center
Johnstone, D. Bruce
This essay examines the administration of public multi-campus universities, focusing in particular on central administration and the State University of New York (SUNY) system and on responses to the need for fiscal austerity. Before treatment of the main topic begins, the essay offers a look at five reasons for the suspicion and resentment that…
Fire behavior simulation in Mediterranean forests using the minimum travel time algorithm
Kostas Kalabokidis; Palaiologos Palaiologou; Mark A. Finney
2014-01-01
Recent large wildfires in Greece exemplify the need for pre-fire burn probability assessment and possible landscape fire flow estimation to enhance fire planning and resource allocation. The Minimum Travel Time (MTT) algorithm, incorporated as FlamMap's version five module, provide valuable fire behavior functions, while enabling multi-core utilization for the...
Federal Register 2010, 2011, 2012, 2013, 2014
2010-04-02
... Manufacturing, Multi-Plastics, Inc., Division, Sipco, Inc., Division, Including Leased Workers of M-Ploy... Manufacturing, Multi-Plastics, Inc., Division and Sipco, Inc., Division, including leased workers of M-Ploy... applicable to TA-W-70,457 is hereby issued as follows: ``All workers of Core Manufacturing, Multi-Plastics...
DOT National Transportation Integrated Search
2006-12-01
Over the last several years, researchers at the University of Arizonas ATLAS Center have developed an adaptive ramp : metering system referred to as MILOS (Multi-Objective, Integrated, Large-Scale, Optimized System). The goal of this project : is ...
Bosse, Stefan
2015-01-01
Multi-agent systems (MAS) can be used for decentralized and self-organizing data processing in a distributed system, like a resource-constrained sensor network, enabling distributed information extraction, for example, based on pattern recognition and self-organization, by decomposing complex tasks in simpler cooperative agents. Reliable MAS-based data processing approaches can aid the material-integration of structural-monitoring applications, with agent processing platforms scaled to the microchip level. The agent behavior, based on a dynamic activity-transition graph (ATG) model, is implemented with program code storing the control and the data state of an agent, which is novel. The program code can be modified by the agent itself using code morphing techniques and is capable of migrating in the network between nodes. The program code is a self-contained unit (a container) and embeds the agent data, the initialization instructions and the ATG behavior implementation. The microchip agent processing platform used for the execution of the agent code is a standalone multi-core stack machine with a zero-operand instruction format, leading to a small-sized agent program code, low system complexity and high system performance. The agent processing is token-queue-based, similar to Petri-nets. The agent platform can be implemented in software, too, offering compatibility at the operational and code level, supporting agent processing in strong heterogeneous networks. In this work, the agent platform embedded in a large-scale distributed sensor network is simulated at the architectural level by using agent-based simulation techniques. PMID:25690550
Bosse, Stefan
2015-02-16
Multi-agent systems (MAS) can be used for decentralized and self-organizing data processing in a distributed system, like a resource-constrained sensor network, enabling distributed information extraction, for example, based on pattern recognition and self-organization, by decomposing complex tasks in simpler cooperative agents. Reliable MAS-based data processing approaches can aid the material-integration of structural-monitoring applications, with agent processing platforms scaled to the microchip level. The agent behavior, based on a dynamic activity-transition graph (ATG) model, is implemented with program code storing the control and the data state of an agent, which is novel. The program code can be modified by the agent itself using code morphing techniques and is capable of migrating in the network between nodes. The program code is a self-contained unit (a container) and embeds the agent data, the initialization instructions and the ATG behavior implementation. The microchip agent processing platform used for the execution of the agent code is a standalone multi-core stack machine with a zero-operand instruction format, leading to a small-sized agent program code, low system complexity and high system performance. The agent processing is token-queue-based, similar to Petri-nets. The agent platform can be implemented in software, too, offering compatibility at the operational and code level, supporting agent processing in strong heterogeneous networks. In this work, the agent platform embedded in a large-scale distributed sensor network is simulated at the architectural level by using agent-based simulation techniques.
Data acquisition system issues for large experiments
NASA Astrophysics Data System (ADS)
Siskind, E. J.
2007-09-01
This talk consists of personal observations on two classes of data acquisition ("DAQ") systems for Silicon trackers in large experiments with which the author has been concerned over the last three or more years. The first half is a classic "lessons learned" recital based on experience with the high-level debug and configuration of the DAQ system for the GLAST LAT detector. The second half is concerned with a discussion of the promises and pitfalls of using modern (and future) generations of "system-on-a-chip" ("SOC") or "platform" field-programmable gate arrays ("FPGAs") in future large DAQ systems. The DAQ system pipeline for the 864k channels of Si tracker in the GLAST LAT consists of five tiers of hardware buffers which ultimately feed into the main memory of the (two-active-node) level-3 trigger processor farm. The data formats and buffer volumes of these tiers are briefly described, as well as the flow control employed between successive tiers. Lessons learned regarding data formats, buffer volumes, and flow control/data discard policy are discussed. The continued development of platform FPGAs containing large amounts of configurable logic fabric, embedded PowerPC hard processor cores, digital signal processing components, large volumes of on-chip buffer memory, and multi-gigabit serial I/O capability permits DAQ system designers to vastly increase the amount of data preprocessing that can be performed in parallel within the DAQ pipeline for detector systems in large experiments. The capabilities of some currently available FPGA families are reviewed, along with the prospects for next-generation families of announced, but not yet available, platform FPGAs. Some experience with an actual implementation is presented, and reconciliation between advertised and achievable specifications is attempted. The prospects for applying these components to space-borne Si tracker detectors are briefly discussed.
Burnett, Ryan D.; Roberts, L. Jay
2015-01-01
Whether by design or default, single species management often serves as an umbrella for species with similar habitat requirements. In recent decades the focus of National Forest management in the Sierra Nevada of California has shifted towards increasing closed canopy mature forest conditions through the protection of areas occupied by the California Spotted Owl (Strix occidentalis occidentalis). To evaluate the implications of these habitat changes and the potential umbrella resulting from a system of owl reserves on the broader avian community, we estimated occupancy of birds inside and outside of Spotted Owl Home Range Core Areas in northeastern California. We used point count data in a multi-species hierarchical Bayesian model incorporating the detection history of 81 species over a two-year time period (2005-2006). A small set of vegetation cover and topography covariates were included in the model to account for broad differences in habitat conditions, as well as a term identifying whether or not a site was within a Core Area. Seventeen species had a negative Core Area effect, seven had a positive effect, and the rest were not significant. Estimated species richness was significantly different with 23.1 species per 100 m radius circle outside Core Areas and 21.7 inside Core Areas. The majority of the species negatively associated with Core Areas are tied to early successional and other disturbance-dependent habitats. Conservation and climate vulnerability rankings were mixed. On average we found higher scores (greater risk) for the species positively associated with Core Areas, but a larger number of species with the highest scores were negatively associated with Core Areas. We discuss the implications for managing the Sierra Nevada ecosystem and illustrate the role of monitoring broader suites of species in guiding management of large complex ecosystems. PMID:25905920
Multi-Quadrant Biopsy Technique Improves Diagnostic Ability in Large Heterogeneous Renal Masses.
Abel, E Jason; Heckman, Jennifer E; Hinshaw, Louis; Best, Sara; Lubner, Meghan; Jarrard, David F; Downs, Tracy M; Nakada, Stephen Y; Lee, Fred T; Huang, Wei; Ziemlewicz, Timothy
2015-10-01
Percutaneous biopsy obtained from a single location is prone to sampling error in large heterogeneous renal masses, leading to nondiagnostic results or failure to detect poor prognostic features. We evaluated the accuracy of percutaneous biopsy for large renal masses using a modified multi-quadrant technique vs a standard biopsy technique. Clinical and pathological data for all patients with cT2 or greater renal masses who underwent percutaneous biopsy from 2009 to 2014 were reviewed. The multi-quadrant technique was defined as multiple core biopsies from at least 4 separate solid enhancing areas in the tumor. The incidence of nondiagnostic findings, sarcomatoid features and procedural complications was recorded, and concordance between biopsy specimens and nephrectomy pathology was compared. A total of 122 biopsies were performed for 117 tumors in 116 patients (46 using the standard biopsy technique and 76 using the multi-quadrant technique). Median tumor size was 10 cm (IQR 8-12). Biopsy was nondiagnostic in 5 of 46 (10.9%) standard and 0 of 76 (0%) multi-quadrant biopsies (p=0.007). Renal cell carcinoma was identified in 96 of 115 (82.0%) tumors and nonrenal cell carcinoma tumors were identified in 21 (18.0%). One complication occurred using the standard biopsy technique and no complications were reported using the multi-quadrant technique. Sarcomatoid features were present in 23 of 96 (23.9%) large renal cell carcinomas studied. Sensitivity for identifying sarcomatoid features was higher using the multi-quadrant technique compared to the standard biopsy technique at 13 of 15 (86.7%) vs 2 of 8 (25.0%) (p=0.0062). The multi-quadrant percutaneous biopsy technique increases the ability to identify aggressive pathological features in large renal tumors and decreases nondiagnostic biopsy rates. Copyright © 2015 American Urological Association Education and Research, Inc. Published by Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Limmer, D. R.; Clift, P. D.; Koehler, C.; Giosan, L.; Ponton, C.; Henstock, T.; Tabrez, A.
2010-12-01
Source to sink processes in large fluvial systems are complicated by large transport distances and the potential to store and rework material on route to the submarine fan. We target the Indus river system and assess how climate change since the Last Glacial Maximum (LGM) may have affected the storage and deposition of sediment in the nearshore shelf setting. While sediment reworking within the floodplain appears to have been strong during the Holocene, it is unclear whether this can be observed in the deep sea or in the submarine delta. We present a multi-proxy record of mineralogical and geochemical change from two cores obtained from the Indus Shelf during Winter 2008/9, one located close to the modern river and one located in the north-west shelf. Results show a strong contrast in the geochemistry, reflectance spectroscopy and clay mineralogy between Holocene sediments from the two cores. We propose that these differences are caused by both local variations in sediment source and transport mechanisms. Trends common in both cores could be related to climatic processes, such as low values in the chemical alteration index (CIA) and low 87Sr/86Sr that rise between 11 and 8ka suggests more intense chemical weathering at that time. This period coincides with presumed warmer, wet conditions and a stronger summer monsoon. A small decline in chemical weathering after 8ka could be caused by an apparent weakening of the monsoon since that time. These data suggest that sediment weathered in the floodplains is transported quickly to the submarine delta during the Holocene, but that this material has not yet been re-deposited into the deep water via the Indus Canyon.
Design and analysis of a nuclear reactor core for innovative small light water reactors
NASA Astrophysics Data System (ADS)
Soldatov, Alexey I.
In order to address the energy needs of developing countries and remote communities, Oregon State University has proposed the Multi-Application Small Light Water Reactor (MASLWR) design. In order to achieve five years of operation without refueling, use of 8% enriched fuel is necessary. This dissertation is focused on core design issues related with increased fuel enrichment (8.0%) and specific MASLWR operational conditions (such as lower operational pressure and temperature, and increased leakage due to small core). Neutron physics calculations are performed with the commercial nuclear industry tools CASMO-4 and SIMULATE-3, developed by Studsvik Scandpower Inc. The first set of results are generated from infinite lattice level calculations with CASMO-4, and focus on evaluation of the principal differences between standard PWR fuel and MASLWR fuel. Chapter 4-1 covers aspects of fuel isotopic composition changes with burnup, evaluation of kinetic parameters and reactivity coefficients. Chapter 4-2 discusses gadolinium self-shielding and shadowing effects, and subsequent impacts on power generation peaking and Reactor Control System shadowing. The second aspect of the research is dedicated to core design issues, such as reflector design (chapter 4-3), burnable absorber distribution and programmed fuel burnup and fuel use strategy (chapter 4-4). This section also includes discussion of the parameters important for safety and evaluation of Reactor Control System options for the proposed core design. An evaluation of the sensitivity of the proposed design to uncertainty in calculated parameters is presented in chapter 4-5. The results presented in this dissertation cover a new area of reactor design and operational parameters, and may be applicable to other small and large pressurized water reactor designs.
VizieR Online Data Catalog: Massive stars in 30 Dor (Schneider+, 2018)
NASA Astrophysics Data System (ADS)
Schneider, F. R. N.; Sana, H.; Evans, C. J.; Bestenlehner, J. M.; Castro, N.; Fossati, L.; Grafener, G.; Langer, N.; Ramirez-Agudelo, O. H.; Sabin-Sanjulian, C.; Simon-Diaz, S.; Tramper, F.; Crowther, P. A.; de Koter, A.; de Mink, S. E.; Dufton, P. L.; Garcia, M.; Gieles, M.; Henault-Brunet, V.; Herrero, A.; Izzard, R. G.; Kalari, V.; Lennon, D. J.; Apellaniz, J. M.; Markova, N.; Najarro, F.; Podsiadlowski, P.; Puls, J.; Taylor, W. D.; van Loon, J. T.; Vink, J. S.; Norman, C.
2018-02-01
Through the use of the Fibre Large Array Multi Element Spectrograph (FLAMES) on the Very Large Telescope (VLT), the VLT-FLAMES Tarantula Survey (VFTS) has obtained optical spectra of ~800 massive stars in 30 Dor, avoiding the core region of the dense star cluster R136 because of difficulties with crowding. Repeated observations at multiple epochs allow determination of the orbital motion of potentially binary objects. For a sample of 452 apparently single stars, robust stellar parameters-such as effective temperatures, luminosities, surface gravities, and projected rotational velocities-are determined by modeling the observed spectra. Composite spectra of visual multiple systems and spectroscopic binaries are not considered here because their parameters cannot be reliably inferred from the VFTS data. To match the derived atmospheric parameters of the apparently single VFTS stars to stellar evolutionary models, we use the Bayesian code Bonnsai. (2 data files).
Granovsky, Alexander A
2015-12-21
We present a new, very efficient semi-numerical approach for the computation of state-specific nuclear gradients of a generic state-averaged multi-configuration self consistent field wavefunction. Our approach eliminates the costly coupled-perturbed multi-configuration Hartree-Fock step as well as the associated integral transformation stage. The details of the implementation within the Firefly quantum chemistry package are discussed and several sample applications are given. The new approach is routinely applicable to geometry optimization of molecular systems with 1000+ basis functions using a standalone multi-core workstation.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Granovsky, Alexander A., E-mail: alex.granovsky@gmail.com
We present a new, very efficient semi-numerical approach for the computation of state-specific nuclear gradients of a generic state-averaged multi-configuration self consistent field wavefunction. Our approach eliminates the costly coupled-perturbed multi-configuration Hartree-Fock step as well as the associated integral transformation stage. The details of the implementation within the Firefly quantum chemistry package are discussed and several sample applications are given. The new approach is routinely applicable to geometry optimization of molecular systems with 1000+ basis functions using a standalone multi-core workstation.
Turkbey, Baris; Xu, Sheng; Kruecker, Jochen; Locklin, Julia; Pang, Yuxi; Shah, Vijay; Bernardo, Marcelino; Baccala, Angelo; Rastinehad, Ardeshir; Benjamin, Compton; Merino, Maria J; Wood, Bradford J; Choyke, Peter L; Pinto, Peter A
2011-03-29
During transrectal ultrasound (TRUS)-guided prostate biopsies, the actual location of the biopsy site is rarely documented. Here, we demonstrate the capability of TRUS-magnetic resonance imaging (MRI) image fusion to document the biopsy site and correlate biopsy results with multi-parametric MRI findings. Fifty consecutive patients (median age 61 years) with a median prostate-specific antigen (PSA) level of 5.8 ng/ml underwent 12-core TRUS-guided biopsy of the prostate. Pre-procedural T2-weighted magnetic resonance images were fused to TRUS. A disposable needle guide with miniature tracking sensors was attached to the TRUS probe to enable fusion with MRI. Real-time TRUS images during biopsy and the corresponding tracking information were recorded. Each biopsy site was superimposed onto the MRI. Each biopsy site was classified as positive or negative for cancer based on the results of each MRI sequence. Sensitivity, specificity, and receiver operating curve (ROC) area under the curve (AUC) values were calculated for multi-parametric MRI. Gleason scores for each multi-parametric MRI pattern were also evaluated. Six hundred and 5 systemic biopsy cores were analyzed in 50 patients, of whom 20 patients had 56 positive cores. MRI identified 34 of 56 positive cores. Overall, sensitivity, specificity, and ROC area values for multi-parametric MRI were 0.607, 0.727, 0.667, respectively. TRUS-MRI fusion after biopsy can be used to document the location of each biopsy site, which can then be correlated with MRI findings. Based on correlation with tracked biopsies, T2-weighted MRI and apparent diffusion coefficient maps derived from diffusion-weighted MRI are the most sensitive sequences, whereas the addition of delayed contrast enhancement MRI and three-dimensional magnetic resonance spectroscopy demonstrated higher specificity consistent with results obtained using radical prostatectomy specimens.
In-Situ Three-Dimensional Shape Rendering from Strain Values Obtained Through Optical Fiber Sensors
NASA Technical Reports Server (NTRS)
Chan, Hon Man (Inventor); Parker, Jr., Allen R. (Inventor)
2015-01-01
A method and system for rendering the shape of a multi-core optical fiber or multi-fiber bundle in three-dimensional space in real time based on measured fiber strain data. Three optical fiber cores arc arranged in parallel at 120.degree. intervals about a central axis. A series of longitudinally co-located strain sensor triplets, typically fiber Bragg gratings, are positioned along the length of each fiber at known intervals. A tunable laser interrogates the sensors to detect strain on the fiber cores. Software determines the strain magnitude (.DELTA.L/L) for each fiber at a given triplet, but then applies beam theory to calculate curvature, beading angle and torsion of the fiber bundle, and from there it determines the shape of the fiber in s Cartesian coordinate system by solving a series of ordinary differential equations expanded from the Frenet-Serrat equations. This approach eliminates the need for computationally time-intensive curve-tilting and allows the three-dimensional shape of the optical fiber assembly to be displayed in real-time.
Bogren, Sara; Fornara, Andrea; Ludwig, Frank; del Puerto Morales, Maria; Steinhoff, Uwe; Fougt Hansen, Mikkel; Kazakova, Olga; Johansson, Christer
2015-01-01
This study presents classification of different magnetic single- and multi-core particle systems using their measured dynamic magnetic properties together with their nanocrystal and particle sizes. The dynamic magnetic properties are measured with AC (dynamical) susceptometry and magnetorelaxometry and the size parameters are determined from electron microscopy and dynamic light scattering. Using these methods, we also show that the nanocrystal size and particle morphology determines the dynamic magnetic properties for both single- and multi-core particles. The presented results are obtained from the four year EU NMP FP7 project, NanoMag, which is focused on standardization of analysis methods for magnetic nanoparticles. PMID:26343639
Real time display Fourier-domain OCT using multi-thread parallel computing with data vectorization
NASA Astrophysics Data System (ADS)
Eom, Tae Joong; Kim, Hoon Seop; Kim, Chul Min; Lee, Yeung Lak; Choi, Eun-Seo
2011-03-01
We demonstrate a real-time display of processed OCT images using multi-thread parallel computing with a quad-core CPU of a personal computer. The data of each A-line are treated as one vector to maximize the data translation rate between the cores of the CPU and RAM stored image data. A display rate of 29.9 frames/sec for processed OCT data (4096 FFT-size x 500 A-scans) is achieved in our system using a wavelength swept source with 52-kHz swept frequency. The data processing times of the OCT image and a Doppler OCT image with a 4-time average are 23.8 msec and 91.4 msec.
Onboard autonomous mission re-planning for multi-satellite system
NASA Astrophysics Data System (ADS)
Zheng, Zixuan; Guo, Jian; Gill, Eberhard
2018-04-01
This paper presents an onboard autonomous mission re-planning system for Multi-Satellites System (MSS) to perform onboard re-planing in disruptive situations. The proposed re-planning system can deal with different potential emergency situations. This paper uses Multi-Objective Hybrid Dynamic Mutation Genetic Algorithm (MO-HDM GA) combined with re-planning techniques as the core algorithm. The Cyclically Re-planning Method (CRM) and the Near Real-time Re-planning Method (NRRM) are developed to meet different mission requirements. Simulations results show that both methods can provide feasible re-planning sequences under unforeseen situations. The comparisons illustrate that using the CRM is average 20% faster than the NRRM on computation time. However, by using the NRRM more raw data can be observed and transmitted than using the CRM within the same period. The usability of this onboard re-planning system is not limited to multi-satellite system. Other mission planning and re-planning problems related to autonomous multiple vehicles with similar demands are also applicable.
Nan, Zhufen; Chi, Xuefen
2016-12-20
The IEEE 802.15.7 protocol suggests that it could coordinate the channel access process based on the competitive method of carrier sensing. However, the directionality of light and randomness of diffuse reflection would give rise to a serious imperfect carrier sense (ICS) problem [e.g., hidden node (HN) problem and exposed node (EN) problem], which brings great challenges in realizing the optical carrier sense multiple access (CSMA) mechanism. In this paper, the carrier sense process implemented by diffuse reflection light is modeled as the choice of independent sets. We establish an ICS model with the presence of ENs and HNs for the multi-point to multi-point visible light communication (VLC) uplink communications system. Considering the severe optical ICS problem, an optical hard core point process (OHCPP) is developed, which characterizes the optical CSMA for the indoor VLC uplink communications system. Due to the limited coverage of the transmitted optical signal, in our OHCPP, the ENs within the transmitters' carrier sense region could be retained provided that they could not corrupt the ongoing communications. Moreover, because of the directionality of both light emitting diode (LED) transmitters and receivers, theoretical analysis of the HN problem becomes difficult. In this paper, we derive the closed-form expression for approximating the outage probability and transmission capacity of VLC networks with the presence of HNs and ENs. Simulation results validate the analysis and also show the existence of an optimal physical carrier-sensing threshold that maximizes the transmission capacity for a given emission angle of LED.
Multicore fiber beamforming network for broadband satellite communications
NASA Astrophysics Data System (ADS)
Zainullin, Airat; Vidal, Borja; Macho, Andres; Llorente, Roberto
2017-02-01
Multi-core fiber (MCF) has been one of the main innovations in fiber optics in the last decade. Reported work on MCF has been focused on increasing the transmission capacity of optical communication links by exploiting space-division multiplexing. Additionally, MCF presents a strong potential in optical beamforming networks. The use of MCF can increase the compactness of the broadband antenna array controller. This is of utmost importance in platforms where size and weight are critical parameters such as communications satellites and airplanes. Here, an optical beamforming architecture that exploits the space-division capacity of MCF to implement compact optical beamforming networks is proposed, being a new application field for MCF. The experimental demonstration of this system using a 4-core MCF that controls a four-element antenna array is reported. An analysis of the impact of MCF on the performance of antenna arrays is presented. The analysis indicates that the main limitation comes from the relatively high insertion loss in the MCF fan-in and fan-out devices, which leads to angle dependent losses which can be mitigated by using fixed optical attenuators or a photonic lantern to reduce MCF insertion loss. The crosstalk requirements are also experimentally evaluated for the proposed MCF-based architecture. The potential signal impairment in the beamforming network is analytically evaluated, being of special importance when MCF with a large number of cores is considered. Finally, the optimization of the proposed MCF-based beamforming network is addressed targeting the scalability to large arrays.
Development of SPR temperature sensor using Au/TiO2 on hetero-core optical fiber
NASA Astrophysics Data System (ADS)
Kitagawa, Sho; Yamazaki, Hiroshi; Hosoki, Ai; Nishiyama, Michiko; Watanabe, Kazuhiro
2016-03-01
This paper describes a novel temperature sensor based on a hetero-core structured fiber optic surface plasmon resonance (SPR) sensor with multi-layer thin film of gold (Au) and titanium dioxide (TiO2). Temperature condition is an essential parameter in chemical plants for avoiding fire accident and controlling qualities of chemical substances. Several fiber optic temperature sensors have been developed for some advantages such as immunity to electromagnetic interference, corrosion resistance and no electrical leakage. The proposed hetero-core fiber optic SPR sensor detects temperature condition by measuring slight refractive index changes of TiO2 which has a large thermo-optic coefficient. We experimentally confirmed that the SPR resonant wavelength in the hetero-core SPR sensor with coating an Au film which slightly depended on temperature changes in the range from 20 °C to 80 °C. In addition, it was experimentally shown that the proposed SPR temperature sensor with multi-layer film of Au and TiO2 had the SPR resonant wavelength shift of 1.6 nm due to temperature change from -10 °C to 50 °C. As a result, a series of experiments successfully demonstrated that the proposed sensor was able to detect temperature directly depending on the thermo-optic effect of TiO2.
Outcomes of Grazing Impacts between Sub-Neptunes in Kepler Multis
NASA Astrophysics Data System (ADS)
Hwang, Jason; Chatterjee, Sourav; Lombardi, James, Jr.; Steffen, Jason H.; Rasio, Frederic
2018-01-01
Studies of high-multiplicity, tightly packed planetary systems suggest that dynamical instabilities are common and affect both the orbits and planet structures, where the compact orbits and typically low densities make physical collisions likely outcomes. Since the structure of many of these planets is such that the mass is dominated by a rocky core, but the volume is dominated by a tenuous gas envelope, the sticky-sphere approximation, used in dynamical integrators, may be a poor model for these collisions. We perform five sets of collision calculations, including detailed hydrodynamics, sampling mass ratios, and core mass fractions typical in Kepler Multis. In our primary set of calculations, we use Kepler-36 as a nominal remnant system, as the two planets have a small dynamical separation and an extreme density ratio. We use an N-body code, Mercury 6.2, to integrate initially unstable systems and study the resultant collisions in detail. We use these collisions, focusing on grazing collisions, in combination with realistic planet models created using gas profiles from Modules for Experiments in Stellar Astrophysics and core profiles using equations of state from Seager et al. to perform hydrodynamic calculations, finding scatterings, mergers, and even a potential planet–planet binary. We dynamically integrate the remnant systems, examine the stability, and estimate the final densities, finding that the remnant densities are sensitive to the core masses, and collisions result in generally more stable systems. We provide prescriptions for predicting the outcomes and modeling the changes in mass and orbits following collisions for general use in dynamical integrators.
Innovative energy absorbing devices based on composite tubes
NASA Astrophysics Data System (ADS)
Tiwari, Chandrashekhar
Analytical and experimental study of innovative load limiting and energy absorbing devices are presented here. The devices are based on composite tubes and can be categorized in to two groups based upon the energy absorbing mechanisms exhibited by them, namely: foam crushing and foam fracturing. The device based on foam crushing as the energy absorbing mechanism is composed of light weight elastic-plastic foam filling inside an angle ply composite tube. The tube is tailored to have a high Poisson’s ratio (>20). Upon being loaded the device experiences large transverse contraction resulting in rapid decrease in diameter. At a certain axial load the foam core begins to crush and energy is dissipated. This device is termed as crush tube device. The device based upon foam shear fracture as the energy absorbing mechanism involves an elastic-plastic core foam in annulus of two concentric extension-twist coupled composite tubes with opposite angles of fibers. The core foam is bonded to the inner and outer tube walls. Upon being loaded axially, the tubes twist in opposite directions and fracture the core foam in out of plane shear and thus dissipate the energy stored. The device is termed as sandwich core device (SCD). The devices exhibit variations in force-displacement characteristics with changes in design and material parameters, resulting in wide range of energy absorption capabilities. A flexible matrix composite system was selected, which was composed of high stiffness carbon fibers as reinforcements in relatively low stiffness polyurethane matrix, based upon large strain to failure capabilities and large beneficial elastic couplings. Linear and non-linear analytical models were developed encapsulating large deformation theory of the laminated composite shells (using non-linear strain energy formulation) to the fracture mechanics of core foam and elastic-plastic deformation theory of the foam filling. The non-linear model is capable of including material and geometric nonlinearities that arise from large deformation and fiber reorientation. Developed non-linear analysis predicts the behavior of extension-twist coupled and angle ply flexible matrix composite tubes under multi-axial loadings. The predicted results show close correlation with experimental findings. It was also found that these devices exhibit variations with respect to rate of loading. It was found that the novel energy absorbing devices are capable of providing 4-5 times higher specific energy absorption (SEA) than currently used devices for similar purposes (such as wire bender which has SEA of 3.6 J/g).
NASA Astrophysics Data System (ADS)
Ratnesh, R. K.; Mehata, Mohan Singh
2017-02-01
We report two port synthesis of CdSe/CdS/ZnS core-multi-shell quantum dots (Q-dots) and their structural properties. The multi-shell structures of Q-dots were developed by using successive ionic layer adsorption and reaction (SILAR) technique. The obtained Q-dots show high crystallinity with the step-wise adjustment of lattice parameters in the radial direction. The size of the core and core-shell Q-dots estimated by transmission electron microscopy images and absorption spectra is about 3.4 and 5.3 nm, respectively. The water soluble Q-dots (scheme-1) were prepared by using ligand exchange method, and the effect of pH was discussed regarding the variation of quantum yield (QY). The decrease of a lifetime of core-multi-shell Q-dots with respect to core CdSe indicates that the shell growth may be tuned by the lifetimes. Thus, the study clearly demonstrates that the core-shell approach can be used to substantially improve the optical properties of Q-dots desired for various applications.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yoginath, Srikanth B; Perumalla, Kalyan S; Henz, Brian J
2012-01-01
In prior work (Yoginath and Perumalla, 2011; Yoginath, Perumalla and Henz, 2012), the motivation, challenges and issues were articulated in favor of virtual time ordering of Virtual Machines (VMs) in network simulations hosted on multi-core machines. Two major components in the overall virtualization challenge are (1) virtual timeline establishment and scheduling of VMs, and (2) virtualization of inter-VM communication. Here, we extend prior work by presenting scaling results for the first component, with experiment results on up to 128 VMs scheduled in virtual time order on a single 12-core host. We also explore the solution space of design alternatives formore » the second component, and present performance results from a multi-threaded, multi-queue implementation of inter-VM network control for synchronized execution with VM scheduling, incorporated in our NetWarp simulation system.« less
Cheung, Kit; Schultz, Simon R; Luk, Wayne
2015-01-01
NeuroFlow is a scalable spiking neural network simulation platform for off-the-shelf high performance computing systems using customizable hardware processors such as Field-Programmable Gate Arrays (FPGAs). Unlike multi-core processors and application-specific integrated circuits, the processor architecture of NeuroFlow can be redesigned and reconfigured to suit a particular simulation to deliver optimized performance, such as the degree of parallelism to employ. The compilation process supports using PyNN, a simulator-independent neural network description language, to configure the processor. NeuroFlow supports a number of commonly used current or conductance based neuronal models such as integrate-and-fire and Izhikevich models, and the spike-timing-dependent plasticity (STDP) rule for learning. A 6-FPGA system can simulate a network of up to ~600,000 neurons and can achieve a real-time performance of 400,000 neurons. Using one FPGA, NeuroFlow delivers a speedup of up to 33.6 times the speed of an 8-core processor, or 2.83 times the speed of GPU-based platforms. With high flexibility and throughput, NeuroFlow provides a viable environment for large-scale neural network simulation.
Diagnostic and Hardware Upgrades for the US-PRC PMI Collaboration on EAST
NASA Astrophysics Data System (ADS)
Tritz, Kevin; Maingi, R.; Andruczyk, D.; Canik, J.; Wang, Z.; Wirth, B.; Zinkle, S.; Woller, K.; Hu, J. S.; Luo, G. N.; Gong, X. Z.; EAST Team
2017-10-01
Several collaborative diagnostic and hardware upgrades are planned to improve understanding and control of Plasma-Material Interactions on EAST, as part of the US-PRC PMI collaboration. Dual-band thermography adapters, designed by UT-K and ORNL, are being designed for existing IR cameras to improve the accuracy of the divertor heat flux measurements by reducing sensitivity to surface emissivity. These measurements should improve power accounting for EAST discharges, which can show a large gap between input power and divertor exhaust power. MIT is preparing tungsten tiles with fluorine depth markers to measure net erosion of PFC tiles. JHU plans to improve the electronics of the Multi-Energy Soft X-ray diagnostic as well as expand the present edge system to a full core-edge measurement; this will enhance the assessment of the effect of Li injection on tungsten accumulation and transport. In addition to PPPL-developed upgrades to the lithium granule and pellet delivery systems, LANL is assessing core-shell micropellets for pellet ablation analysis. Finally, UIUC and PPPL are developing flowing liquid lithium limiters, both with and without LiMIT tile features, for deployment on EAST. Work supported by DoE award DE-SC0016553.
Cheung, Kit; Schultz, Simon R.; Luk, Wayne
2016-01-01
NeuroFlow is a scalable spiking neural network simulation platform for off-the-shelf high performance computing systems using customizable hardware processors such as Field-Programmable Gate Arrays (FPGAs). Unlike multi-core processors and application-specific integrated circuits, the processor architecture of NeuroFlow can be redesigned and reconfigured to suit a particular simulation to deliver optimized performance, such as the degree of parallelism to employ. The compilation process supports using PyNN, a simulator-independent neural network description language, to configure the processor. NeuroFlow supports a number of commonly used current or conductance based neuronal models such as integrate-and-fire and Izhikevich models, and the spike-timing-dependent plasticity (STDP) rule for learning. A 6-FPGA system can simulate a network of up to ~600,000 neurons and can achieve a real-time performance of 400,000 neurons. Using one FPGA, NeuroFlow delivers a speedup of up to 33.6 times the speed of an 8-core processor, or 2.83 times the speed of GPU-based platforms. With high flexibility and throughput, NeuroFlow provides a viable environment for large-scale neural network simulation. PMID:26834542
NASA Astrophysics Data System (ADS)
Lin, Y.; O'Malley, D.; Vesselinov, V. V.
2015-12-01
Inverse modeling seeks model parameters given a set of observed state variables. However, for many practical problems due to the facts that the observed data sets are often large and model parameters are often numerous, conventional methods for solving the inverse modeling can be computationally expensive. We have developed a new, computationally-efficient Levenberg-Marquardt method for solving large-scale inverse modeling. Levenberg-Marquardt methods require the solution of a dense linear system of equations which can be prohibitively expensive to compute for large-scale inverse problems. Our novel method projects the original large-scale linear problem down to a Krylov subspace, such that the dimensionality of the measurements can be significantly reduced. Furthermore, instead of solving the linear system for every Levenberg-Marquardt damping parameter, we store the Krylov subspace computed when solving the first damping parameter and recycle it for all the following damping parameters. The efficiency of our new inverse modeling algorithm is significantly improved by using these computational techniques. We apply this new inverse modeling method to invert for a random transitivity field. Our algorithm is fast enough to solve for the distributed model parameters (transitivity) at each computational node in the model domain. The inversion is also aided by the use regularization techniques. The algorithm is coded in Julia and implemented in the MADS computational framework (http://mads.lanl.gov). Julia is an advanced high-level scientific programing language that allows for efficient memory management and utilization of high-performance computational resources. By comparing with a Levenberg-Marquardt method using standard linear inversion techniques, our Levenberg-Marquardt method yields speed-up ratio of 15 in a multi-core computational environment and a speed-up ratio of 45 in a single-core computational environment. Therefore, our new inverse modeling method is a powerful tool for large-scale applications.
Energy Efficient Real-Time Scheduling Using DPM on Mobile Sensors with a Uniform Multi-Cores
Kim, Youngmin; Lee, Chan-Gun
2017-01-01
In wireless sensor networks (WSNs), sensor nodes are deployed for collecting and analyzing data. These nodes use limited energy batteries for easy deployment and low cost. The use of limited energy batteries is closely related to the lifetime of the sensor nodes when using wireless sensor networks. Efficient-energy management is important to extending the lifetime of the sensor nodes. Most effort for improving power efficiency in tiny sensor nodes has focused mainly on reducing the power consumed during data transmission. However, recent emergence of sensor nodes equipped with multi-cores strongly requires attention to be given to the problem of reducing power consumption in multi-cores. In this paper, we propose an energy efficient scheduling method for sensor nodes supporting a uniform multi-cores. We extend the proposed T-Ler plane based scheduling for global optimal scheduling of a uniform multi-cores and multi-processors to enable power management using dynamic power management. In the proposed approach, processor selection for a scheduling and mapping method between the tasks and processors is proposed to efficiently utilize dynamic power management. Experiments show the effectiveness of the proposed approach compared to other existing methods. PMID:29240695
LOD-based clustering techniques for efficient large-scale terrain storage and visualization
NASA Astrophysics Data System (ADS)
Bao, Xiaohong; Pajarola, Renato
2003-05-01
Large multi-resolution terrain data sets are usually stored out-of-core. To visualize terrain data at interactive frame rates, the data needs to be organized on disk, loaded into main memory part by part, then rendered efficiently. Many main-memory algorithms have been proposed for efficient vertex selection and mesh construction. Organization of terrain data on disk is quite difficult because the error, the triangulation dependency and the spatial location of each vertex all need to be considered. Previous terrain clustering algorithms did not consider the per-vertex approximation error of individual terrain data sets. Therefore, the vertex sequences on disk are exactly the same for any terrain. In this paper, we propose a novel clustering algorithm which introduces the level-of-detail (LOD) information to terrain data organization to map multi-resolution terrain data to external memory. In our approach the LOD parameters of the terrain elevation points are reflected during clustering. The experiments show that dynamic loading and paging of terrain data at varying LOD is very efficient and minimizes page faults. Additionally, the preprocessing of this algorithm is very fast and works from out-of-core.
Multi-emitter laser multiplexer using a two-mirror beam shaper
NASA Astrophysics Data System (ADS)
Cobb, Joshua M.; Brennan, John; Bhatia, Vikram
2014-12-01
A system was designed and built to spatially multiplex four broad area laser diodes (BALD) and condense the light into a multi-mode fiber with a core diameter of 105 um and an NA of 0.15. The lasers were efficiently combined with an étendue aspect ratio scaler (EARS) optic. The EARS works under the principle of a two mirror beam shaper. We were able to successfully couple more than 87% of the optical energy into the fiber. The design of the optical system and the results of several built systems are discussed.
'Micro-hole' optical dating of quartz from HOTRAX-05 Arctic Ocean cores
NASA Astrophysics Data System (ADS)
Berger, G. W.; Polyak, L. V.
2011-12-01
For Quaternary Arctic Ocean cores, numeric dating methods are needed spanning and exceeding the age range of the widely used radiocarbon (C-14) method. Previously, luminescence sediment dating of 4-11 μm diameter quartz and feldspar grains from core tops has often produced large burial-age overestimates (e.g., by >7 kyr) due to failure to resolve mixed-age histories. However, application of micro-focused-laser ('micro-hole') photon-stimulated-luminescence (PSL) applied to quartz grains of 11-90 μm diameters from the tops (upper 2 cm) of high-sedimentation- rate HOTRAX-05 multi-cores at the Alaska margin provides expected near zero ages (0-200 a), thus overcoming the earlier problem of large PSL age over-estimation. This micro-hole PSL dating approach has also been applied to >11 μm quartz grains from multi-cores at two sites on the central Lomonosov Ridge. For a core top within a perched basin, a burial-age estimate of ~2 ka for 11-62 μm quartz was obtained, in accord with published C-14 age estimates from foraminifera, demonstrating the efficacy of the micro-hole approach to this ridge area. At a nearby 'erosive' ridge-top site, the micro-hole PSL approach paradoxically produces two different burial-age estimates from the same core-top horizon. The >90 μm quartz grains yield a burial age of ~25 ka, in accord with a C-14 age estimate of ~26 ka from >250 μm foraminifers from the same horizon. However, the 11-90 μm quartz produces a burial-age estimate of ~9 ka, indicating a differently preserved burial history for the medium silt grains than for the sand grains within a single horizon. This unexpected result provides a unique insight into past, complicated, depositional processes on this ridge top over a time range spanning the LGM. These results from the micro-hole PSL approach thus indicate a clear potential for dating times of detrital quartz deposition at other ridge tops in the Arctic Ocean, and for providing perhaps new insights into local preservation of burial ages. These PSL procedures are being applied also to sediment above and below a diamicton in a HOTRAX-05 core from the Northwind Ridge, with the aim of dating indirectly the diamicton. Preliminary results from this core will be presented.
A Combined Eulerian-Lagrangian Data Representation for Large-Scale Applications.
Sauer, Franz; Xie, Jinrong; Ma, Kwan-Liu
2017-10-01
The Eulerian and Lagrangian reference frames each provide a unique perspective when studying and visualizing results from scientific systems. As a result, many large-scale simulations produce data in both formats, and analysis tasks that simultaneously utilize information from both representations are becoming increasingly popular. However, due to their fundamentally different nature, drawing correlations between these data formats is a computationally difficult task, especially in a large-scale setting. In this work, we present a new data representation which combines both reference frames into a joint Eulerian-Lagrangian format. By reorganizing Lagrangian information according to the Eulerian simulation grid into a "unit cell" based approach, we can provide an efficient out-of-core means of sampling, querying, and operating with both representations simultaneously. We also extend this design to generate multi-resolution subsets of the full data to suit the viewer's needs and provide a fast flow-aware trajectory construction scheme. We demonstrate the effectiveness of our method using three large-scale real world scientific datasets and provide insight into the types of performance gains that can be achieved.
NASA Astrophysics Data System (ADS)
Balaji, V.; Benson, Rusty; Wyman, Bruce; Held, Isaac
2016-10-01
Climate models represent a large variety of processes on a variety of timescales and space scales, a canonical example of multi-physics multi-scale modeling. Current hardware trends, such as Graphical Processing Units (GPUs) and Many Integrated Core (MIC) chips, are based on, at best, marginal increases in clock speed, coupled with vast increases in concurrency, particularly at the fine grain. Multi-physics codes face particular challenges in achieving fine-grained concurrency, as different physics and dynamics components have different computational profiles, and universal solutions are hard to come by. We propose here one approach for multi-physics codes. These codes are typically structured as components interacting via software frameworks. The component structure of a typical Earth system model consists of a hierarchical and recursive tree of components, each representing a different climate process or dynamical system. This recursive structure generally encompasses a modest level of concurrency at the highest level (e.g., atmosphere and ocean on different processor sets) with serial organization underneath. We propose to extend concurrency much further by running more and more lower- and higher-level components in parallel with each other. Each component can further be parallelized on the fine grain, potentially offering a major increase in the scalability of Earth system models. We present here first results from this approach, called coarse-grained component concurrency, or CCC. Within the Geophysical Fluid Dynamics Laboratory (GFDL) Flexible Modeling System (FMS), the atmospheric radiative transfer component has been configured to run in parallel with a composite component consisting of every other atmospheric component, including the atmospheric dynamics and all other atmospheric physics components. We will explore the algorithmic challenges involved in such an approach, and present results from such simulations. Plans to achieve even greater levels of coarse-grained concurrency by extending this approach within other components, such as the ocean, will be discussed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhang, Hongbin; Szilard, Ronaldo; Epiney, Aaron
Under the auspices of the DOE LWRS Program RISMC Industry Application ECCS/LOCA, INL has engaged staff from both South Texas Project (STP) and the Texas A&M University (TAMU) to produce a generic pressurized water reactor (PWR) model including reactor core, clad/fuel design and systems thermal hydraulics based on the South Texas Project (STP) nuclear power plant, a 4-Loop Westinghouse PWR. A RISMC toolkit, named LOCA Toolkit for the U.S. (LOTUS), has been developed for use in this generic PWR plant model to assess safety margins for the proposed NRC 10 CFR 50.46c rule, Emergency Core Cooling System (ECCS) performance duringmore » LOCA. This demonstration includes coupled analysis of core design, fuel design, thermalhydraulics and systems analysis, using advanced risk analysis tools and methods to investigate a wide range of results. Within this context, a multi-physics best estimate plus uncertainty (MPBEPU) methodology framework is proposed.« less
Carbonaceous aerosol tracers in ice-cores record multi-decadal climate oscillations
Seki, Osamu; Kawamura, Kimitaka; Bendle, James A. P.; Izawa, Yusuke; Suzuki, Ikuko; Shiraiwa, Takayuki; Fujii, Yoshiyuki
2015-01-01
Carbonaceous aerosols influence the climate via direct and indirect effects on radiative balance. However, the factors controlling the emissions, transport and role of carbonaceous aerosols in the climate system are highly uncertain. Here we investigate organic tracers in ice cores from Greenland and Kamchatka and find that, throughout the period covered by the records (1550 to 2000 CE), the concentrations and composition of biomass burning-, soil bacterial- and plant wax- tracers correspond to Arctic and regional temperatures as well as the warm season Arctic Oscillation (AO) over multi-decadal time-scales. Specifically, order of magnitude decreases (increases) in abundances of ice-core organic tracers, likely representing significant decreases (increases) in the atmospheric loading of carbonaceous aerosols, occur during colder (warmer) phases in the high latitudinal Northern Hemisphere. This raises questions about causality and possible carbonaceous aerosol feedback mechanisms. Our work opens new avenues for ice core research. Translating concentrations of organic tracers (μg/kg-ice or TOC) from ice-cores, into estimates of the atmospheric loading of carbonaceous aerosols (μg/m3) combined with new model constraints on the strength and sign of climate forcing by carbonaceous aerosols should be a priority for future research. PMID:26411576
NASA Astrophysics Data System (ADS)
Shi, X.
2015-12-01
As NSF indicated - "Theory and experimentation have for centuries been regarded as two fundamental pillars of science. It is now widely recognized that computational and data-enabled science forms a critical third pillar." Geocomputation is the third pillar of GIScience and geosciences. With the exponential growth of geodata, the challenge of scalable and high performance computing for big data analytics become urgent because many research activities are constrained by the inability of software or tool that even could not complete the computation process. Heterogeneous geodata integration and analytics obviously magnify the complexity and operational time frame. Many large-scale geospatial problems may be not processable at all if the computer system does not have sufficient memory or computational power. Emerging computer architectures, such as Intel's Many Integrated Core (MIC) Architecture and Graphics Processing Unit (GPU), and advanced computing technologies provide promising solutions to employ massive parallelism and hardware resources to achieve scalability and high performance for data intensive computing over large spatiotemporal and social media data. Exploring novel algorithms and deploying the solutions in massively parallel computing environment to achieve the capability for scalable data processing and analytics over large-scale, complex, and heterogeneous geodata with consistent quality and high-performance has been the central theme of our research team in the Department of Geosciences at the University of Arkansas (UARK). New multi-core architectures combined with application accelerators hold the promise to achieve scalability and high performance by exploiting task and data levels of parallelism that are not supported by the conventional computing systems. Such a parallel or distributed computing environment is particularly suitable for large-scale geocomputation over big data as proved by our prior works, while the potential of such advanced infrastructure remains unexplored in this domain. Within this presentation, our prior and on-going initiatives will be summarized to exemplify how we exploit multicore CPUs, GPUs, and MICs, and clusters of CPUs, GPUs and MICs, to accelerate geocomputation in different applications.
High-efficiency wavefunction updates for large scale Quantum Monte Carlo
NASA Astrophysics Data System (ADS)
Kent, Paul; McDaniel, Tyler; Li, Ying Wai; D'Azevedo, Ed
Within ab intio Quantum Monte Carlo (QMC) simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunctions. The evaluation of each Monte Carlo move requires finding the determinant of a dense matrix, which is traditionally iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. For calculations with thousands of electrons, this operation dominates the execution profile. We propose a novel rank- k delayed update scheme. This strategy enables probability evaluation for multiple successive Monte Carlo moves, with application of accepted moves to the matrices delayed until after a predetermined number of moves, k. Accepted events grouped in this manner are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency. This procedure does not change the underlying Monte Carlo sampling or the sampling efficiency. For large systems and algorithms such as diffusion Monte Carlo where the acceptance ratio is high, order of magnitude speedups can be obtained on both multi-core CPU and on GPUs, making this algorithm highly advantageous for current petascale and future exascale computations.
Progress Towards a Rad-Hydro Code for Modern Computing Architectures LA-UR-10-02825
NASA Astrophysics Data System (ADS)
Wohlbier, J. G.; Lowrie, R. B.; Bergen, B.; Calef, M.
2010-11-01
We are entering an era of high performance computing where data movement is the overwhelming bottleneck to scalable performance, as opposed to the speed of floating-point operations per processor. All multi-core hardware paradigms, whether heterogeneous or homogeneous, be it the Cell processor, GPGPU, or multi-core x86, share this common trait. In multi-physics applications such as inertial confinement fusion or astrophysics, one may be solving multi-material hydrodynamics with tabular equation of state data lookups, radiation transport, nuclear reactions, and charged particle transport in a single time cycle. The algorithms are intensely data dependent, e.g., EOS, opacity, nuclear data, and multi-core hardware memory restrictions are forcing code developers to rethink code and algorithm design. For the past two years LANL has been funding a small effort referred to as Multi-Physics on Multi-Core to explore ideas for code design as pertaining to inertial confinement fusion and astrophysics applications. The near term goals of this project are to have a multi-material radiation hydrodynamics capability, with tabular equation of state lookups, on cartesian and curvilinear block structured meshes. In the longer term we plan to add fully implicit multi-group radiation diffusion and material heat conduction, and block structured AMR. We will report on our progress to date.
Precise masses for the transiting planetary system HD 106315 with HARPS
NASA Astrophysics Data System (ADS)
Barros, S. C. C.; Gosselin, H.; Lillo-Box, J.; Bayliss, D.; Delgado Mena, E.; Brugger, B.; Santerne, A.; Armstrong, D. J.; Adibekyan, V.; Armstrong, J. D.; Barrado, D.; Bento, J.; Boisse, I.; Bonomo, A. S.; Bouchy, F.; Brown, D. J. A.; Cochran, W. D.; Collier Cameron, A.; Deleuil, M.; Demangeon, O.; Díaz, R. F.; Doyle, A.; Dumusque, X.; Ehrenreich, D.; Espinoza, N.; Faedi, F.; Faria, J. P.; Figueira, P.; Foxell, E.; Hébrard, G.; Hojjatpanah, S.; Jackman, J.; Lendl, M.; Ligi, R.; Lovis, C.; Melo, C.; Mousis, O.; Neal, J. J.; Osborn, H. P.; Pollacco, D.; Santos, N. C.; Sefako, R.; Shporer, A.; Sousa, S. G.; Triaud, A. H. M. J.; Udry, S.; Vigan, A.; Wyttenbach, A.
2017-12-01
Context. The multi-planetary system HD 106315 was recently found in K2 data. The planets have periods of Pb 9.55 and Pc 21.06 days, and radii of rb = 2.44 ± 0.17 R⊕ and rc = 4.35 ± 0.23 R⊕ . The brightness of the host star (V = 9.0 mag) makes it an excellent target for transmission spectroscopy. However, to interpret transmission spectra it is crucial to measure the planetary masses. Aims: We obtained high precision radial velocities for HD 106315 to determine the mass of the two transiting planets discovered with Kepler K2. Our successful observation strategy was carefully tailored to mitigate the effect of stellar variability. Methods: We modelled the new radial velocity data together with the K2 transit photometry and a new ground-based partial transit of HD 106315c to derive system parameters. Results: We estimate the mass of HD 106315b to be 12.6 ± 3.2 M⊕ and the density to be 4.7 ± 1.7 g cm-3, while for HD 106315c we estimate a mass of 15.2 ± 3.7 M⊕ and a density of 1.01 ± 0.29 g cm-3. Hence, despite planet c having a radius almost twice as large as planet b, their masses are consistent with one another. Conclusions: We conclude that HD 106315c has a thick hydrogen-helium gaseous envelope. A detailed investigation of HD 106315b using a planetary interior model constrains the core mass fraction to be 5-29%, and the water mass fraction to be 10-50%. An alternative, not considered by our model, is that HD 106315b is composed of a large rocky core with a thick H-He envelope. Transmission spectroscopy of these planets will give insight into their atmospheric compositions and also help constrain their core compositions. Based on observations collected at the European Organisation for Astronomical Research in the Southern Hemisphere under ESO programme 198.C-0168.
2 × 2 MIMO OFDM/OQAM radio signals over an elliptical core few-mode fiber.
Mo, Qi; He, Jiale; Yu, Dawei; Deng, Lei; Fu, Songnian; Tang, Ming; Liu, Deming
2016-10-01
We experimentally demonstrate a 4.46 Gb/s2×2 multi-input multi-output (MIMO) orthogonal frequency division multiplexing (OFDM)/OQAM radio signal over a 2 km elliptical core 3-mode fiber, together with 0.4 m wireless transmission. Meanwhile, to cope with differential channel delay (DCD) among involved MIMO channels, we propose a time-offset crosstalk cancellation algorithm to extend the DCD tolerance from 10 to 60 ns without using a circle prefix (CP), leading to an 18.7% improvement of spectral efficiency. For the purpose of comparison, we also examine the transmission performance of CP-OFDM signals with different lengths of CPs, under the same system configuration. The proposed algorithm is also effective for the DCD compensation of a radio signal over a 2 km 7-core fiber. These results not only demonstrate the feasibility of space division multiplexing for RoF application but also validate that the elliptical core few-mode fiber can provide the same independent channels as the multicore fiber.
Kalman Filter Tracking on Parallel Architectures
NASA Astrophysics Data System (ADS)
Cerati, Giuseppe; Elmer, Peter; Lantz, Steven; McDermott, Kevin; Riley, Dan; Tadel, Matevž; Wittich, Peter; Würthwein, Frank; Yagil, Avi
2015-12-01
Power density constraints are limiting the performance improvements of modern CPUs. To address this we have seen the introduction of lower-power, multi-core processors, but the future will be even more exciting. In order to stay within the power density limits but still obtain Moore's Law performance/price gains, it will be necessary to parallelize algorithms to exploit larger numbers of lightweight cores and specialized functions like large vector units. Example technologies today include Intel's Xeon Phi and GPGPUs. Track finding and fitting is one of the most computationally challenging problems for event reconstruction in particle physics. At the High Luminosity LHC, for example, this will be by far the dominant problem. The need for greater parallelism has driven investigations of very different track finding techniques including Cellular Automata or returning to Hough Transform. The most common track finding techniques in use today are however those based on the Kalman Filter [2]. Significant experience has been accumulated with these techniques on real tracking detector systems, both in the trigger and offline. They are known to provide high physics performance, are robust and are exactly those being used today for the design of the tracking system for HL-LHC. Our previous investigations showed that, using optimized data structures, track fitting with Kalman Filter can achieve large speedup both with Intel Xeon and Xeon Phi. We report here our further progress towards an end-to-end track reconstruction algorithm fully exploiting vectorization and parallelization techniques in a realistic simulation setup.
Matrix theory for baryons: an overview of holographic QCD for nuclear physics.
Aoki, Sinya; Hashimoto, Koji; Iizuka, Norihiro
2013-10-01
We provide, for non-experts, a brief overview of holographic QCD (quantum chromodynamics) and a review of the recent proposal (Hashimoto et al 2010 (arXiv:1003.4988[hep-th])) of a matrix-like description of multi-baryon systems in holographic QCD. Based on the matrix model, we derive the baryon interaction at short distances in multi-flavor holographic QCD. We show that there is a very universal repulsive core of inter-baryon forces for a generic number of flavors. This is consistent with a recent lattice QCD analysis for Nf = 2, 3 where the repulsive core looks universal. We also provide a comparison of our results with the lattice QCD and the operator product expansion analysis.
"Photonic lantern" spectral filters in multi-core Fiber.
Birks, T A; Mangan, B J; Díez, A; Cruz, J L; Murphy, D F
2012-06-18
Fiber Bragg gratings are written across all 120 single-mode cores of a multi-core optical Fiber. The Fiber is interfaced to multimode ports by tapering it within a depressed-index glass jacket. The result is a compact multimode "photonic lantern" filter with astrophotonic applications. The tapered structure is also an effective mode scrambler.
The last Deglaciation in the Mediterranean region: a multi-archives synthesis
NASA Astrophysics Data System (ADS)
Bazin, Lucie; Siani, Giuseppe; Landais, Amaelle; Bassinot, Frank; Genty, Dominique; Govin, Aline; Michel, Elisabeth; Nomade, Sebastien; Waelbroeck, Claire
2016-04-01
Multiple proxies record past climatic changes in different climate archives. These proxies are influenced by different component of the climate system and bring complementary information on past climate variability. The major limitation when combining proxies from different archives comes from the coherency of their chronologies. Indeed, each climate archives possess their own dating methods, not necessarily coherent with each other's. Consequently, when we want to assess the latitudinal changes and mechanisms behind a climate event, we often have to rely on assumptions of synchronisation between the different archives, such as synchronous temperature changes during warming events (Austin and Hibbert 2010). Recently, a dating method originally developed to produce coherent chronologies for ice cores (Datice,Lemieux-Dudon et al., 2010) has been adapted in order to integrate different climate archives (ice cores, sediment cores and speleothems (Lemieux-Dudon et al., 2015, Bazin et al., in prep)). In this presentation we present the validation of this multi-archives dating tool with a first application covering the last Deglaciation in the Mediterranean region. For this experiment, we consider the records from Monticchio, the MD90-917, Tenaghi Philippon and Lake Orhid sediment cores as well as continuous speleothems from Sofular, Soreq and La Mine caves. Using the Datice dating tool, and with the identification of common tephra layers between the cores considered, we are able to produce a multi-archives coherent chronology for this region, independently of any climatic assumption. Using this common chronological framework, we show that the usual climatic synchronisation assumptions are not valid over this region for the last glacial-interglacial transition. Finally, we compare our coherent Mediterranean chronology with Greenland ice core records in order to discuss the sequence of events of the last Deglaciation between these two regions.
Assessment User Guide for Colleges and Universities
ERIC Educational Resources Information Center
American Association of Collegiate Registrars and Admissions Officers (AACRAO), 2015
2015-01-01
The Smarter Balanced Assessment Consortium is one of two multi-state consortia that have built new assessment systems aligned to the Common Core State Standards. The Smarter Balanced Assessment Consortium is composed of 18 states and the U.S. Virgin Islands that have worked together to develop a comprehensive assessment system aligned to the…
NASA Astrophysics Data System (ADS)
Liu, Jiping; Kang, Xiaochen; Dong, Chun; Xu, Shenghua
2017-12-01
Surface area estimation is a widely used tool for resource evaluation in the physical world. When processing large scale spatial data, the input/output (I/O) can easily become the bottleneck in parallelizing the algorithm due to the limited physical memory resources and the very slow disk transfer rate. In this paper, we proposed a stream tilling approach to surface area estimation that first decomposed a spatial data set into tiles with topological expansions. With these tiles, the one-to-one mapping relationship between the input and the computing process was broken. Then, we realized a streaming framework towards the scheduling of the I/O processes and computing units. Herein, each computing unit encapsulated a same copy of the estimation algorithm, and multiple asynchronous computing units could work individually in parallel. Finally, the performed experiment demonstrated that our stream tilling estimation can efficiently alleviate the heavy pressures from the I/O-bound work, and the measured speedup after being optimized have greatly outperformed the directly parallel versions in shared memory systems with multi-core processors.
Owen, Sheldon F.; Berl, Jacob L.; Edwards, John W.; Ford, W. Mark; Wood, Petra Bohall
2015-01-01
We studied a raccoon (Procyon lotor) population within a managed central Appalachian hardwood forest in West Virginia to investigate the effects of intensive forest management on raccoon spatial requirements and habitat selection. Raccoon home-range (95% utilization distribution) and core-area (50% utilization distribution) size differed between sexes with males maintaining larger (2×) home ranges and core areas than females. Home-range and core-area size did not differ between seasons for either sex. We used compositional analysis to quantify raccoon selection of six different habitat types at multiple spatial scales. Raccoons selected riparian corridors (riparian management zones [RMZ]) and intact forests (> 70 y old) at the core-area spatial scale. RMZs likely were used by raccoons because they provided abundant denning resources (i.e., large-diameter trees) as well as access to water. Habitat composition associated with raccoon foraging locations indicated selection for intact forests, riparian areas, and regenerating harvest (stands <10 y old). Although raccoons were able to utilize multiple habitat types for foraging resources, a selection of intact forest and RMZs at multiple spatial scales indicates the need of mature forest (with large-diameter trees) for this species in managed forests in the central Appalachians.
Chemical Convention in the Lunar Core from Melting Experiments on the Ironsulfur System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, J.; Liu, J.; Chen, B.
2012-03-26
By reanalyzing Apollo lunar seismograms using array-processing methods, a recent study suggests that the Moon has a solid inner core and a fluid outer core, much like the Earth. The volume fraction of the lunar inner core is 38%, compared with 4% for the Earth. The pressure at the Moon's core-mantle boundary is 4.8 GPa, and that at the ICB is 5.2 GPa. The partially molten state of the lunar core provides constraints on the thermal and chemical states of the Moon: The temperature at the inner core boundary (ICB) corresponds to the liquidus of the outer core composition, andmore » the mass fraction of the solid core allows us to infer the bulk composition of the core from an estimated thermal profile. Moreover, knowledge on the extent of core solidification can be used to evaluate the role of chemical convection in the origin of early lunar core dynamo. Sulfur is considered an antifreeze component in the lunar core. Here we investigate the melting behavior of the Fe-S system at the pressure conditions of the lunar core, using the multi-anvil apparatus and synchrotron and laboratory-based analytical methods. Our goal is to understand compositionally driven convection in the lunar core and assess its role in generating an internal magnetic field in the early history of the Moon.« less
Final Report for Project DE-FC02-06ER25755 [Pmodels2
DOE Office of Scientific and Technical Information (OSTI.GOV)
Panda, Dhabaleswar; Sadayappan, P.
2014-03-12
In this report, we describe the research accomplished by the OSU team under the Pmodels2 project. The team has worked on various angles: designing high performance MPI implementations on modern networking technologies (Mellanox InfiniBand (including the new ConnectX2 architecture and Quad Data Rate), QLogic InfiniPath, the emerging 10GigE/iWARP and RDMA over Converged Enhanced Ethernet (RoCE) and Obsidian IB-WAN), studying MPI scalability issues for multi-thousand node clusters using XRC transport, scalable job start-up, dynamic process management support, efficient one-sided communication, protocol offloading and designing scalable collective communication libraries for emerging multi-core architectures. New designs conforming to the Argonne’s Nemesis interface havemore » also been carried out. All of these above solutions have been integrated into the open-source MVAPICH/MVAPICH2 software. This software is currently being used by more than 2,100 organizations worldwide (in 71 countries). As of January ’14, more than 200,000 downloads have taken place from the OSU Web site. In addition, many InfiniBand vendors, server vendors, system integrators and Linux distributors have been incorporating MVAPICH/MVAPICH2 into their software stacks and distributing it. Several InfiniBand systems using MVAPICH/MVAPICH2 have obtained positions in the TOP500 ranking of supercomputers in the world. The latest November ’13 ranking include the following systems: 7th ranked Stampede system at TACC with 462,462 cores; 11th ranked Tsubame 2.5 system at Tokyo Institute of Technology with 74,358 cores; 16th ranked Pleiades system at NASA with 81,920 cores; Work on PGAS models has proceeded on multiple directions. The Scioto framework, which supports task-parallelism in one-sided and global-view parallel programming, has been extended to allow multi-processor tasks that are executed by processor groups. A quantum Monte Carlo application is being ported onto the extended Scioto framework. A public release of Global Trees (GT) has been made, along with the Global Chunks (GC) framework on which GT is built. The Global Chunks (GC) layer is also being used as the basis for the development of a higher level Global Graphs (GG) layer. The Global Graphs (GG) system will provide a global address space view of distributed graph data structures on distributed memory systems.« less
Machine Learning Toolkit for Extreme Scale
DOE Office of Scientific and Technical Information (OSTI.GOV)
2014-03-31
Support Vector Machines (SVM) is a popular machine learning technique, which has been applied to a wide range of domains such as science, finance, and social networks for supervised learning. MaTEx undertakes the challenge of designing a scalable parallel SVM training algorithm for large scale systems, which includes commodity multi-core machines, tightly connected supercomputers and cloud computing systems. Several techniques are proposed for improved speed and memory space usage including adaptive and aggressive elimination of samples for faster convergence , and sparse format representation of data samples. Several heuristics for earliest possible to lazy elimination of non-contributing samples are consideredmore » in MaTEx. In many cases, where an early sample elimination might result in a false positive, low overhead mechanisms for reconstruction of key data structures are proposed. The proposed algorithm and heuristics are implemented and evaluated on various publicly available datasets« less
2017-03-20
computation, Prime Implicates, Boolean Abstraction, real- time embedded software, software synthesis, correct by construction software design , model...types for time -dependent data-flow networks". J.-P. Talpin, P. Jouvelot, S. Shukla. ACM-IEEE Conference on Methods and Models for System Design ...information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and
Symbolic Analysis of Concurrent Programs with Polymorphism
NASA Technical Reports Server (NTRS)
Rungta, Neha Shyam
2010-01-01
The current trend of multi-core and multi-processor computing is causing a paradigm shift from inherently sequential to highly concurrent and parallel applications. Certain thread interleavings, data input values, or combinations of both often cause errors in the system. Systematic verification techniques such as explicit state model checking and symbolic execution are extensively used to detect errors in such systems [7, 9]. Explicit state model checking enumerates possible thread schedules and input data values of a program in order to check for errors [3, 9]. To partially mitigate the state space explosion from data input values, symbolic execution techniques substitute data input values with symbolic values [5, 7, 6]. Explicit state model checking and symbolic execution techniques used in conjunction with exhaustive search techniques such as depth-first search are unable to detect errors in medium to large-sized concurrent programs because the number of behaviors caused by data and thread non-determinism is extremely large. We present an overview of abstraction-guided symbolic execution for concurrent programs that detects errors manifested by a combination of thread schedules and data values [8]. The technique generates a set of key program locations relevant in testing the reachability of the target locations. The symbolic execution is then guided along these locations in an attempt to generate a feasible execution path to the error state. This allows the execution to focus in parts of the behavior space more likely to contain an error.
NASA Astrophysics Data System (ADS)
Jones, Jerry; Rhoades, Valerie; Arner, Radford; Clem, Timothy; Cuneo, Adam
2007-04-01
NDE measurements, monitoring, and control of smart and adaptive composite structures requires that the central knowledge system have an awareness of the entire structure. Achieving this goal necessitates the implementation of an integrated network of significant numbers of sensors. Additionally, in order to temporally coordinate the data from specially distributed sensors, the data must be time relevant. Early adoption precludes development of sensor technology specifically for this application, instead it will depend on the ability to utilize legacy systems. Partially supported by the U.S. Department of Commerce, National Institute of Standards and Technology, Advanced Technology Development Program (NIST-ATP), a scalable integrated system has been developed to implement monitoring of structural integrity and the control of adaptive/intelligent structures. The project, called SHIELD (Structural Health Identification and Electronic Life Determination), was jointly undertaken by: Caterpillar, N.A. Tech., Motorola, and Microstrain. SHIELD is capable of operation with composite structures, metallic structures, or hybrid structures. SHIELD consists of a real-time processing core on a Motorola MPC5200 using a C language based real-time operating system (RTOS). The RTOS kernel was customized to include a virtual backplane which makes the system completely scalable. This architecture provides for multiple processes to be operating simultaneously. They may be embedded as multiple threads on the core hardware or as separate independent processors connected to the core using a software driver called a NAT-Network Integrator (NATNI). NATNI's can be created for any communications application. In it's current embodiment, NATNI's have been created for CAN bus, TCP/IP (Ethernet) - both wired and 802.11 b and g, and serial communications using RS485 and RS232. Since SHIELD uses standard C language, it is easy to port any monitoring or control algorithm, thus providing for legacy technology which may use other hardware processors and various communications means. For example, two demonstrations of SHIELD have been completed, in January and May 2005 respectively. One demonstration used algorithms in C running in multiple threads in the SHIELD core and utilizing two different sensor networks, one CAN bus and one wireless. The second had algorithms operating in C on the SHIELD core and other algorithms running on multiple Texas Instruments DSP processors using a NATNI that communicated via wired TCP/IP. A key feature of SHIELD is the implementation of a wireless ZIGBEE (802.15.4) network for implementing large numbers of small, low cost, low power sensors communication via a meshstar wireless network. While SHIELD was designed to integrate with a wide variety of existing communications protocols, a ZIGBEE network capability was implemented specifically for SHIELD. This will facilitate the monitoring of medium to very large structures including marine applications, utility scale multi-megawatt wind energy systems, and aircraft/spacecraft. The SHIELD wireless network will facilitate large numbers of sensors (up to 32000), accommodate sensors embedded into the composite material, can communicate to both sensors and actuators, and prevents obsolescence by providing for re-programming of the nodes via remote RF communications. The wireless network provides for ultra-low energy use, spatial location, and accurate timestamping, utilizing the beaconing feature of ZIGBEE.
Testing of the Multi-Fluid Evaporator Engineering Development Unit
NASA Technical Reports Server (NTRS)
Quinn, Gregory; O'Connor, Ed; Riga, Ken; Anderson, Molly; Westheimer, David
2007-01-01
Hamilton Sundstrand is under contract with the NASA Johnson Space Center to develop a scalable, evaporative heat rejection system called the Multi-Fluid Evaporator (MFE). It is being designed to support the Orion Crew Module and to support future Constellation missions. The MFE would be used from Earth sea level conditions to the vacuum of space. The current Shuttle configuration utilizes an ammonia boiler and flash evaporator system to achieve cooling at all altitudes. The MFE system combines both functions into a single compact package with significant weight reduction and improved freeze-up protection. The heat exchanger core is designed so that radial flow of the evaporant provides increasing surface area to keep the back pressure low. The multiple layer construction of the core allows for efficient scale up to the desired heat rejection rate. The full scale MFE prototype will be constructed with four core sections that, combined with a novel control scheme, manage the risk of freezing the heat exchanger cores. A sub-scale MFE engineering development unit (EDU) has been built, and is identical to one of the four sections of a full scale prototype. The EDU has completed testing at Hamilton Sundstrand. The overall test objective was to determine the thermal performance of the EDU. The first set of tests simulated how each of the four sections of the prototype would perform by varying the chamber pressure, evaporant flow rate, coolant flow rate and coolant temperature. A second set of tests was conducted with an outlet steam header in place to verify that the outlet steam orifices prevent freeze-up in the core while also allowing the desired thermal turn-down ratio. This paper discusses the EDU tests and results.
Large-Scale, Parallel, Multi-Sensor Atmospheric Data Fusion Using Cloud Computing
NASA Astrophysics Data System (ADS)
Wilson, B. D.; Manipon, G.; Hua, H.; Fetzer, E.
2013-05-01
NASA's Earth Observing System (EOS) is an ambitious facility for studying global climate change. The mandate now is to combine measurements from the instruments on the "A-Train" platforms (AIRS, AMSR-E, MODIS, MISR, MLS, and CloudSat) and other Earth probes to enable large-scale studies of climate change over decades. Moving to multi-sensor, long-duration analyses of important climate variables presents serious challenges for large-scale data mining and fusion. For example, one might want to compare temperature and water vapor retrievals from one instrument (AIRS) to another (MODIS), and to a model (ECMWF), stratify the comparisons using a classification of the "cloud scenes" from CloudSat, and repeat the entire analysis over 10 years of data. To efficiently assemble such datasets, we are utilizing Elastic Computing in the Cloud and parallel map/reduce-based algorithms. However, these problems are Data Intensive computing so the data transfer times and storage costs (for caching) are key issues. SciReduce is a Hadoop-like parallel analysis system, programmed in parallel python, that is designed from the ground up for Earth science. SciReduce executes inside VMWare images and scales to any number of nodes in the Cloud. Unlike Hadoop, SciReduce operates on bundles of named numeric arrays, which can be passed in memory or serialized to disk in netCDF4 or HDF5. Figure 1 shows the architecture of the full computational system, with SciReduce at the core. Multi-year datasets are automatically "sharded" by time and space across a cluster of nodes so that years of data (millions of files) can be processed in a massively parallel way. Input variables (arrays) are pulled on-demand into the Cloud using OPeNDAP URLs or other subsetting services, thereby minimizing the size of the cached input and intermediate datasets. We are using SciReduce to automate the production of multiple versions of a ten-year A-Train water vapor climatology under a NASA MEASURES grant. We will present the architecture of SciReduce, describe the achieved "clock time" speedups in fusing datasets on our own nodes and in the Cloud, and discuss the Cloud cost tradeoffs for storage, compute, and data transfer. We will also present a concept/prototype for staging NASA's A-Train Atmospheric datasets (Levels 2 & 3) in the Amazon Cloud so that any number of compute jobs can be executed "near" the multi-sensor data. Given such a system, multi-sensor climate studies over 10-20 years of data could be performed in an efficient way, with the researcher paying only his own Cloud compute bill.; Figure 1 -- Architecture.
Large-Scale, Parallel, Multi-Sensor Atmospheric Data Fusion Using Cloud Computing
NASA Astrophysics Data System (ADS)
Wilson, B. D.; Manipon, G.; Hua, H.; Fetzer, E. J.
2013-12-01
NASA's Earth Observing System (EOS) is an ambitious facility for studying global climate change. The mandate now is to combine measurements from the instruments on the 'A-Train' platforms (AIRS, AMSR-E, MODIS, MISR, MLS, and CloudSat) and other Earth probes to enable large-scale studies of climate change over decades. Moving to multi-sensor, long-duration analyses of important climate variables presents serious challenges for large-scale data mining and fusion. For example, one might want to compare temperature and water vapor retrievals from one instrument (AIRS) to another (MODIS), and to a model (MERRA), stratify the comparisons using a classification of the 'cloud scenes' from CloudSat, and repeat the entire analysis over 10 years of data. To efficiently assemble such datasets, we are utilizing Elastic Computing in the Cloud and parallel map/reduce-based algorithms. However, these problems are Data Intensive computing so the data transfer times and storage costs (for caching) are key issues. SciReduce is a Hadoop-like parallel analysis system, programmed in parallel python, that is designed from the ground up for Earth science. SciReduce executes inside VMWare images and scales to any number of nodes in the Cloud. Unlike Hadoop, SciReduce operates on bundles of named numeric arrays, which can be passed in memory or serialized to disk in netCDF4 or HDF5. Figure 1 shows the architecture of the full computational system, with SciReduce at the core. Multi-year datasets are automatically 'sharded' by time and space across a cluster of nodes so that years of data (millions of files) can be processed in a massively parallel way. Input variables (arrays) are pulled on-demand into the Cloud using OPeNDAP URLs or other subsetting services, thereby minimizing the size of the cached input and intermediate datasets. We are using SciReduce to automate the production of multiple versions of a ten-year A-Train water vapor climatology under a NASA MEASURES grant. We will present the architecture of SciReduce, describe the achieved 'clock time' speedups in fusing datasets on our own compute nodes and in the public Cloud, and discuss the Cloud cost tradeoffs for storage, compute, and data transfer. We will also present a concept/prototype for staging NASA's A-Train Atmospheric datasets (Levels 2 & 3) in the Amazon Cloud so that any number of compute jobs can be executed 'near' the multi-sensor data. Given such a system, multi-sensor climate studies over 10-20 years of data could be performed in an efficient way, with the researcher paying only his own Cloud compute bill. SciReduce Architecture
NASA Astrophysics Data System (ADS)
Unterberg, Ea; Donovan, D.; Barton, J.; Wampler, Wr; Abrams, T.; Thomas, Dm; Petrie, T.; Guo, Hy; Stangeby, Pg; Elder, Jd; Rudakov, D.; Grierson, B.; Victor, B.
2017-10-01
Experiments using metal inserts with novel isotopically-enriched tungsten coatings at the outer divertor strike point (OSP) have provided unique insight into the ELM-induced sourcing, main-SOL transport, and core accumulation control mechanisms of W for a range of operating conditions. This experimental approach has used a multi-head, dual-facing collector probe (CP) at the outboard midplane, as well as W-I and core W spectroscopy. Using the CP system, the total amount of W deposited relative to source measurements shows a clear dependence on ELM size, ELM frequency, and strike point location, with large ELMs depositing significantly more W on the CP from the far-SOL source. Additionally, high spatial ( 1mm) and ELM resolved spectroscopic measurements of W sourcing indicate shifts in the peak erosion rate. Furthermore, high performance discharges with rapid ELMs show core W concentrations of few 10-5, and the CP deposition profile indicates W is predominantly transported to the midplane from the OSP rather than from the far-SOL region. The low central W concentration is shown to be due to flattening of the main plasma density profile, presumably by on-axis electron cyclotron heating. Work supported under USDOE Cooperative Agreement DE-FC02-04ER54698.
Sparse PDF Volumes for Consistent Multi-Resolution Volume Rendering.
Sicat, Ronell; Krüger, Jens; Möller, Torsten; Hadwiger, Markus
2014-12-01
This paper presents a new multi-resolution volume representation called sparse pdf volumes, which enables consistent multi-resolution volume rendering based on probability density functions (pdfs) of voxel neighborhoods. These pdfs are defined in the 4D domain jointly comprising the 3D volume and its 1D intensity range. Crucially, the computation of sparse pdf volumes exploits data coherence in 4D, resulting in a sparse representation with surprisingly low storage requirements. At run time, we dynamically apply transfer functions to the pdfs using simple and fast convolutions. Whereas standard low-pass filtering and down-sampling incur visible differences between resolution levels, the use of pdfs facilitates consistent results independent of the resolution level used. We describe the efficient out-of-core computation of large-scale sparse pdf volumes, using a novel iterative simplification procedure of a mixture of 4D Gaussians. Finally, our data structure is optimized to facilitate interactive multi-resolution volume rendering on GPUs.
Precise strong lensing mass profile of the CLASH galaxy cluster MACS 2129
NASA Astrophysics Data System (ADS)
Monna, A.; Seitz, S.; Balestra, I.; Rosati, P.; Grillo, C.; Halkola, A.; Suyu, S. H.; Coe, D.; Caminha, G. B.; Frye, B.; Koekemoer, A.; Mercurio, A.; Nonino, M.; Postman, M.; Zitrin, A.
2017-04-01
We present a detailed strong lensing (SL) mass reconstruction of the core of the galaxy cluster MACS J2129.4-0741 (zcl = 0.589) obtained by combining high-resolution Hubble Space Telescope photometry from the CLASH (Cluster Lensing And Supernovae survey with Hubble) survey with new spectroscopic observations from the CLASH-VLT (Very Large Telescope) survey. A background bright red passive galaxy at zsp = 1.36, sextuply lensed in the cluster core, has four radial lensed images located over the three central cluster members. Further 19 background lensed galaxies are spectroscopically confirmed by our VLT survey, including 3 additional multiple systems. A total of 31 multiple images are used in the lensing analysis. This allows us to trace with high precision the total mass profile of the cluster in its very inner region (R < 100 kpc). Our final lensing mass model reproduces the multiple images systems identified in the cluster core with high accuracy of 0.4 arcsec. This translates to a high-precision mass reconstruction of MACS 2129, which is constrained at a level of 2 per cent. The cluster has Einstein parameter ΘE = (29 ± 4) arcsec and a projected total mass of Mtot(<ΘE) = (1.35 ± 0.03) × 1014 M⊙ within such radius. Together with the cluster mass profile, we provide here also the complete spectroscopic data set for the cluster members and lensed images measured with VLT/Visible Multi-Object Spectrograph within the CLASH-VLT survey.
NASA Technical Reports Server (NTRS)
Cleveland, Paul; Parrish, Keith; Thomson, Shaun; Marsh, James; Comber, Brian
2016-01-01
The James Webb Space Telescope (JWST), successor to the Hubble Space Telescope, will be the largest astronomical telescope ever sent into space. To observe the very first light of the early universe, JWST requires a large deployed 6.5-meter primary mirror cryogenically cooled to less than 50 Kelvin. Three scientific instruments are further cooled via a large radiator system to less than 40 Kelvin. A fourth scientific instrument is cooled to less than 7 Kelvin using a combination pulse-tube Joule-Thomson mechanical cooler. Passive cryogenic cooling enables the large scale of the telescope which must be highly folded for launch on an Ariane 5 launch vehicle and deployed once on orbit during its journey to the second Earth-Sun Lagrange point. Passive cooling of the observatory is enabled by the deployment of a large tennis court sized five layer Sunshield combined with the use of a network of high efficiency radiators. A high purity aluminum heat strap system connects the three instrument's detector systems to the radiator systems to dissipate less than a single watt of parasitic and instrument dissipated heat. JWST's large scale features, while enabling passive cooling, also prevent the typical flight configuration fully-deployed thermal balance test that is the keystone of most space missions' thermal verification plans. This paper describes the JWST Core 2 Test, which is a cryogenic thermal balance test of a full size, high fidelity engineering model of the Observatory's 'Core' area thermal control hardware. The 'Core' area is the key mechanical and cryogenic interface area between all Observatory elements. The 'Core' area thermal control hardware allows for temperature transition of 300K to approximately 50 K by attenuating heat from the room temperature IEC (instrument electronics) and the Spacecraft Bus. Since the flight hardware is not available for test, the Core 2 test uses high fidelity and flight-like reproductions.
STAMPS: Software Tool for Automated MRI Post-processing on a supercomputer.
Bigler, Don C; Aksu, Yaman; Miller, David J; Yang, Qing X
2009-08-01
This paper describes a Software Tool for Automated MRI Post-processing (STAMP) of multiple types of brain MRIs on a workstation and for parallel processing on a supercomputer (STAMPS). This software tool enables the automation of nonlinear registration for a large image set and for multiple MR image types. The tool uses standard brain MRI post-processing tools (such as SPM, FSL, and HAMMER) for multiple MR image types in a pipeline fashion. It also contains novel MRI post-processing features. The STAMP image outputs can be used to perform brain analysis using Statistical Parametric Mapping (SPM) or single-/multi-image modality brain analysis using Support Vector Machines (SVMs). Since STAMPS is PBS-based, the supercomputer may be a multi-node computer cluster or one of the latest multi-core computers.
Timing and climate forcing of volcanic eruptions for the past 2,500 years.
Sigl, M; Winstrup, M; McConnell, J R; Welten, K C; Plunkett, G; Ludlow, F; Büntgen, U; Caffee, M; Chellman, N; Dahl-Jensen, D; Fischer, H; Kipfstuhl, S; Kostick, C; Maselli, O J; Mekhaldi, F; Mulvaney, R; Muscheler, R; Pasteris, D R; Pilcher, J R; Salzer, M; Schüpbach, S; Steffensen, J P; Vinther, B M; Woodruff, T E
2015-07-30
Volcanic eruptions contribute to climate variability, but quantifying these contributions has been limited by inconsistencies in the timing of atmospheric volcanic aerosol loading determined from ice cores and subsequent cooling from climate proxies such as tree rings. Here we resolve these inconsistencies and show that large eruptions in the tropics and high latitudes were primary drivers of interannual-to-decadal temperature variability in the Northern Hemisphere during the past 2,500 years. Our results are based on new records of atmospheric aerosol loading developed from high-resolution, multi-parameter measurements from an array of Greenland and Antarctic ice cores as well as distinctive age markers to constrain chronologies. Overall, cooling was proportional to the magnitude of volcanic forcing and persisted for up to ten years after some of the largest eruptive episodes. Our revised timescale more firmly implicates volcanic eruptions as catalysts in the major sixth-century pandemics, famines, and socioeconomic disruptions in Eurasia and Mesoamerica while allowing multi-millennium quantification of climate response to volcanic forcing.
Timing and climate forcing of volcanic eruptions for the past 2,500 years
NASA Astrophysics Data System (ADS)
Sigl, M.; Winstrup, M.; McConnell, J. R.; Welten, K. C.; Plunkett, G.; Ludlow, F.; Büntgen, U.; Caffee, M.; Chellman, N.; Dahl-Jensen, D.; Fischer, H.; Kipfstuhl, S.; Kostick, C.; Maselli, O. J.; Mekhaldi, F.; Mulvaney, R.; Muscheler, R.; Pasteris, D. R.; Pilcher, J. R.; Salzer, M.; Schüpbach, S.; Steffensen, J. P.; Vinther, B. M.; Woodruff, T. E.
2015-07-01
Volcanic eruptions contribute to climate variability, but quantifying these contributions has been limited by inconsistencies in the timing of atmospheric volcanic aerosol loading determined from ice cores and subsequent cooling from climate proxies such as tree rings. Here we resolve these inconsistencies and show that large eruptions in the tropics and high latitudes were primary drivers of interannual-to-decadal temperature variability in the Northern Hemisphere during the past 2,500 years. Our results are based on new records of atmospheric aerosol loading developed from high-resolution, multi-parameter measurements from an array of Greenland and Antarctic ice cores as well as distinctive age markers to constrain chronologies. Overall, cooling was proportional to the magnitude of volcanic forcing and persisted for up to ten years after some of the largest eruptive episodes. Our revised timescale more firmly implicates volcanic eruptions as catalysts in the major sixth-century pandemics, famines, and socioeconomic disruptions in Eurasia and Mesoamerica while allowing multi-millennium quantification of climate response to volcanic forcing.
Structural Considerations of a 20MW Multi-Rotor Wind Energy System
NASA Astrophysics Data System (ADS)
Jamieson, P.; Branney, M.
2014-12-01
The drive to upscale offshore wind turbines relates especially to possiblereductions in O&M and electrical interconnection costs per MW of installed capacity.Even with best current technologies, designs with rated capacity above about 3 MW are less cost effective exfactory per rated MW(turbine system costs) than smaller machines.Very large offshore wind turbines are thereforejustifiedprimarily by overall offshore project economics. Furthermore, continuing progress in materials and structures has been essential to avoid severe penalties in the power/mass ratio of large multi-MW machines.The multi-rotor concept employs many small rotors to maximise energy capture area withminimum systemvolume. Previous work has indicated that this can enablea very large reduction in the total weight and cost of rotors and drive trains compared to an equivalent large single rotor system.Thus the multi rotor concept may enable rated capacities of 20 MW or more at a single maintenancesite. Establishing the cost benefit of a multi rotor system requires examination of solutions for the support structure and yawing, ensuring aerodynamic losses from rotor interaction are not significant and that overall logistics, with much increased part count (more reliable components) and less consequence of single failuresare favourable. This paper addresses the viability of a support structure in respect of structural concept and likely weight as one necessary step in exploring the potential of the multi rotor concept.
Exploring Manycore Multinode Systems for Irregular Applications with FPGA Prototyping
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ceriani, Marco; Palermo, Gianluca; Secchi, Simone
We present a prototype of a multi-core architecture implemented on FPGA, designed to enable efficient execution of irregular applications on distributed shared memory machines, while maintaining high performance on regular workloads. The architecture is composed of off-the-shelf soft-core cores, local interconnection and memory interface, integrated with custom components that optimize it for irregular applications. It relies on three key elements: a global address space, multithreading, and fine-grained synchronization. Global addresses are scrambled to reduce the formation of network hot-spots, while the latency of the transactions is covered by integrating an hardware scheduler within the custom load/store buffers to take advantagemore » from the availability of multiple executions threads, increasing the efficiency in a transparent way to the application. We evaluated a dual node system irregular kernels showing scalability in the number of cores and threads.« less
Jay, Raman; Heckman, J E; Hinshaw, L; Best, S; Lubner, M; Jarrard, D F; Downs, T M; Nakada, S Y; Lee, F T; Huang, W; Ziemlewicz, T
2017-03-01
Percutaneous biopsy obtained from a single location is prone to sampling error in large heterogeneous renal masses, leading to nondiagnostic results or failure to detect poor prognostic features. We evaluated the accuracy of percutaneous biopsy for large renal masses using a modified multi-quadrant technique vs. a standard biopsy technique. Clinical and pathological data for all patients with cT2 or greater renal masses who underwent percutaneous biopsy from 2009 to 2014 were reviewed. The multi-quadrant technique was defined as multiple core biopsies from at least 4 separate solid enhancing areas in the tumor. The incidence of nondiagnostic findings, sarcomatoid features and procedural complications was recorded, and concordance between biopsy specimens and nephrectomy pathology was compared. A total of 122 biopsies were performed for 117 tumors in 116 patients (46 using the standard biopsy technique and 76 using the multi-quadrant technique). Median tumor size was 10cm (IQR: 8-12). Biopsy was nondiagnostic in 5 of 46 (10.9%) standard and 0 of 76 (0%) multi-quadrant biopsies (P = 0.007). Renal cell carcinoma was identified in 96 of 115 (82.0%) tumors and nonrenal cell carcinoma tumors were identified in 21 (18.0%). One complication occurred using the standard biopsy technique and no complications were reported using the multi-quadrant technique. Sarcomatoid features were present in 23 of 96 (23.9%) large renal cell carcinomas studied. Sensitivity for identifying sarcomatoid features was higher using the multi-quadrant technique compared to the standard biopsy technique at 13 of 15 (86.7%) vs. 2 of 8 (25.0%) (P = 0.0062). The multi-quadrant percutaneous biopsy technique increases the ability to identify aggressive pathological features in large renal tumors and decreases nondiagnostic biopsy rates. Copyright © 2017. Published by Elsevier Inc.
Fast and Accurate Support Vector Machines on Large Scale Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vishnu, Abhinav; Narasimhan, Jayenthi; Holder, Larry
Support Vector Machines (SVM) is a supervised Machine Learning and Data Mining (MLDM) algorithm, which has become ubiquitous largely due to its high accuracy and obliviousness to dimensionality. The objective of SVM is to find an optimal boundary --- also known as hyperplane --- which separates the samples (examples in a dataset) of different classes by a maximum margin. Usually, very few samples contribute to the definition of the boundary. However, existing parallel algorithms use the entire dataset for finding the boundary, which is sub-optimal for performance reasons. In this paper, we propose a novel distributed memory algorithm to eliminatemore » the samples which do not contribute to the boundary definition in SVM. We propose several heuristics, which range from early (aggressive) to late (conservative) elimination of the samples, such that the overall time for generating the boundary is reduced considerably. In a few cases, a sample may be eliminated (shrunk) pre-emptively --- potentially resulting in an incorrect boundary. We propose a scalable approach to synchronize the necessary data structures such that the proposed algorithm maintains its accuracy. We consider the necessary trade-offs of single/multiple synchronization using in-depth time-space complexity analysis. We implement the proposed algorithm using MPI and compare it with libsvm--- de facto sequential SVM software --- which we enhance with OpenMP for multi-core/many-core parallelism. Our proposed approach shows excellent efficiency using up to 4096 processes on several large datasets such as UCI HIGGS Boson dataset and Offending URL dataset.« less
Optimizing the Performance of Reactive Molecular Dynamics Simulations for Multi-core Architectures
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aktulga, Hasan Metin; Coffman, Paul; Shan, Tzu-Ray
2015-12-01
Hybrid parallelism allows high performance computing applications to better leverage the increasing on-node parallelism of modern supercomputers. In this paper, we present a hybrid parallel implementation of the widely used LAMMPS/ReaxC package, where the construction of bonded and nonbonded lists and evaluation of complex ReaxFF interactions are implemented efficiently using OpenMP parallelism. Additionally, the performance of the QEq charge equilibration scheme is examined and a dual-solver is implemented. We present the performance of the resulting ReaxC-OMP package on a state-of-the-art multi-core architecture Mira, an IBM BlueGene/Q supercomputer. For system sizes ranging from 32 thousand to 16.6 million particles, speedups inmore » the range of 1.5-4.5x are observed using the new ReaxC-OMP software. Sustained performance improvements have been observed for up to 262,144 cores (1,048,576 processes) of Mira with a weak scaling efficiency of 91.5% in larger simulations containing 16.6 million particles.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lamarque, Jean-Francois; Dentener, Frank; McConnell, J.R.
2013-08-20
We present multi-model global datasets of nitrogen and sulfate deposition covering time periods from 1850 to 2100, calculated within the Atmospheric Chemistry and Climate Model Intercomparison Project (ACCMIP). The computed deposition fluxes are compared to surface wet deposition and ice-core measurements. We use a new dataset of wet deposition for 2000-2002 based on critical assessment of the quality of existing regional network data. We show that for present-day (year 2000 ACCMIP time-slice), the ACCMIP results perform similarly to previously published multi-model assessments. The analysis of changes between 1980 and 2000 indicates significant differences between model and measurements over the Unitedmore » States, but less so over Europe. This difference points towards misrepresentation of 1980 NH3 emissions over North America. Based on ice-core records, the 1850 deposition fluxes agree well with Greenland ice cores but the change between 1850 and 2000 seems to be overestimated in the Northern Hemisphere for both nitrogen and sulfur species. Using the Representative Concentration Pathways to define the projected climate and atmospheric chemistry related emissions and concentrations, we find large regional nitrogen deposition increases in 2100 in Latin America, Africa and parts of Asia under some of the scenarios considered. Increases in South Asia are especially large, and are seen in all scenarios, with 2100 values more than double 2000 in some scenarios and reaching >1300 mgN/m2/yr averaged over regional to continental scale regions in RCP 2.6 and 8.5, ~30-50% larger than the values in any region currently (2000). Despite known issues, the new ACCMIP deposition dataset provides novel, consistent and evaluated global gridded deposition fields for use in a wide range of climate and ecological studies.« less
DOT National Transportation Integrated Search
2012-01-01
Recent research indicates that multi-destination transit systems are far more effective in attracting passengers than central business district (CBD)-focused systems. However, the same research suggests that multi-destination systems appeal largely t...
NASA Astrophysics Data System (ADS)
Khuwaileh, Bassam
High fidelity simulation of nuclear reactors entails large scale applications characterized with high dimensionality and tremendous complexity where various physics models are integrated in the form of coupled models (e.g. neutronic with thermal-hydraulic feedback). Each of the coupled modules represents a high fidelity formulation of the first principles governing the physics of interest. Therefore, new developments in high fidelity multi-physics simulation and the corresponding sensitivity/uncertainty quantification analysis are paramount to the development and competitiveness of reactors achieved through enhanced understanding of the design and safety margins. Accordingly, this dissertation introduces efficient and scalable algorithms for performing efficient Uncertainty Quantification (UQ), Data Assimilation (DA) and Target Accuracy Assessment (TAA) for large scale, multi-physics reactor design and safety problems. This dissertation builds upon previous efforts for adaptive core simulation and reduced order modeling algorithms and extends these efforts towards coupled multi-physics models with feedback. The core idea is to recast the reactor physics analysis in terms of reduced order models. This can be achieved via identifying the important/influential degrees of freedom (DoF) via the subspace analysis, such that the required analysis can be recast by considering the important DoF only. In this dissertation, efficient algorithms for lower dimensional subspace construction have been developed for single physics and multi-physics applications with feedback. Then the reduced subspace is used to solve realistic, large scale forward (UQ) and inverse problems (DA and TAA). Once the elite set of DoF is determined, the uncertainty/sensitivity/target accuracy assessment and data assimilation analysis can be performed accurately and efficiently for large scale, high dimensional multi-physics nuclear engineering applications. Hence, in this work a Karhunen-Loeve (KL) based algorithm previously developed to quantify the uncertainty for single physics models is extended for large scale multi-physics coupled problems with feedback effect. Moreover, a non-linear surrogate based UQ approach is developed, used and compared to performance of the KL approach and brute force Monte Carlo (MC) approach. On the other hand, an efficient Data Assimilation (DA) algorithm is developed to assess information about model's parameters: nuclear data cross-sections and thermal-hydraulics parameters. Two improvements are introduced in order to perform DA on the high dimensional problems. First, a goal-oriented surrogate model can be used to replace the original models in the depletion sequence (MPACT -- COBRA-TF - ORIGEN). Second, approximating the complex and high dimensional solution space with a lower dimensional subspace makes the sampling process necessary for DA possible for high dimensional problems. Moreover, safety analysis and design optimization depend on the accurate prediction of various reactor attributes. Predictions can be enhanced by reducing the uncertainty associated with the attributes of interest. Accordingly, an inverse problem can be defined and solved to assess the contributions from sources of uncertainty; and experimental effort can be subsequently directed to further improve the uncertainty associated with these sources. In this dissertation a subspace-based gradient-free and nonlinear algorithm for inverse uncertainty quantification namely the Target Accuracy Assessment (TAA) has been developed and tested. The ideas proposed in this dissertation were first validated using lattice physics applications simulated using SCALE6.1 package (Pressurized Water Reactor (PWR) and Boiling Water Reactor (BWR) lattice models). Ultimately, the algorithms proposed her were applied to perform UQ and DA for assembly level (CASL progression problem number 6) and core wide problems representing Watts Bar Nuclear 1 (WBN1) for cycle 1 of depletion (CASL Progression Problem Number 9) modeled via simulated using VERA-CS which consists of several multi-physics coupled models. The analysis and algorithms developed in this dissertation were encoded and implemented in a newly developed tool kit algorithms for Reduced Order Modeling based Uncertainty/Sensitivity Estimator (ROMUSE).
The development and test of ultra-large-format multi-anode microchannel array detector systems
NASA Technical Reports Server (NTRS)
Timothy, J. G.
1984-01-01
The specific tasks that were accomplished with each of the key elements of the multi-anode microchannel array detector system are described. The modes of operation of position-sensitive electronic readout systems for use with high-gain microchannel plates are described and their performance characteristics compared and contrasted. Multi-anode microchannel array detector systems with formats as large as 256 x 1024 pixels are currently under evaluation. Preliminary performance data for sealed ultraviolet and visible-light detector tubes show that the detector systems have unique characteristics which make them complementary to photoconductive array detectors, such as CCDs, and superior to alternative pulse-counting detector systems employing high-gain MCPs.
Multi Modal Anticipation in Fuzzy Space
NASA Astrophysics Data System (ADS)
Asproth, Viveca; Holmberg, Stig C.; Hâkansson, Anita
2006-06-01
We are all stakeholders in the geographical space, which makes up our common living and activity space. This means that a careful, creative, and anticipatory planning, design, and management of that space will be of paramount importance for our sustained life on earth. Here it is shown that the quality of such planning could be significantly increased with help of a computer based modelling and simulation tool. Further, the design and implementation of such a tool ought to be guided by the conceptual integration of some core concepts like anticipation and retardation, multi modal system modelling, fuzzy space modelling, and multi actor interaction.
Sahoo, Yu Vin; Nakai, Shun'ichi; Ali, Arshad
2006-03-01
Tungsten isotope composition of a sample of deep-seated rock can record the influence of core-mantle interaction of the parent magma. Samples of kimberlite, which is known as a carrier of diamond, from the deep mantle might exhibit effects of core-mantle interaction. Although tungsten isotope anomaly was reported for kimberlites from South Africa, a subsequent investigation did not verify the anomaly. The magnesium-rich and calcium-rich chemical composition of kimberlite might engender difficulty during chemical separation of tungsten for isotope analyses. This paper presents a simple, one-step anion exchange technique for precise and accurate determination of tungsten isotopes in kimberlites using multi-collector inductively coupled plasma mass spectrometry (MC-ICP-MS). Large quantities of Ca and Mg in kimberlite samples were precipitated and removed with aqueous H(2)SO(4). Highly pure fractions of tungsten for isotopic measurements were obtained following an anion exchange chromatographic procedure involving mixed acids. That procedure enabled efficient removal of high field strength elements (HFSE), such as Hf, Zr and Ti, which are small ions that carry strong charges and develop intense electrostatic fields. The tungsten yields were 85%-95%. Advantages of this system include less time and less use of reagents. Precise and accurate isotopic measurements are possible using fractions of tungsten that are obtained using this method. The accuracy and precision of these measurements were confirmed using various silicate standard rock samples, JB-2, JB-3 and AGV-1.
T-L Plane Abstraction-Based Energy-Efficient Real-Time Scheduling for Multi-Core Wireless Sensors.
Kim, Youngmin; Lee, Ki-Seong; Pham, Ngoc-Son; Lee, Sun-Ro; Lee, Chan-Gun
2016-07-08
Energy efficiency is considered as a critical requirement for wireless sensor networks. As more wireless sensor nodes are equipped with multi-cores, there are emerging needs for energy-efficient real-time scheduling algorithms. The T-L plane-based scheme is known to be an optimal global scheduling technique for periodic real-time tasks on multi-cores. Unfortunately, there has been a scarcity of studies on extending T-L plane-based scheduling algorithms to exploit energy-saving techniques. In this paper, we propose a new T-L plane-based algorithm enabling energy-efficient real-time scheduling on multi-core sensor nodes with dynamic power management (DPM). Our approach addresses the overhead of processor mode transitions and reduces fragmentations of the idle time, which are inherent in T-L plane-based algorithms. Our experimental results show the effectiveness of the proposed algorithm compared to other energy-aware scheduling methods on T-L plane abstraction.
Energy Efficient Engine core design and performance report
NASA Technical Reports Server (NTRS)
Stearns, E. Marshall
1982-01-01
The Energy Efficient Engine (E3) is a NASA program to develop fuel saving technology for future large transport aircraft engines. Testing of the General Electric E3 core showed that the core component performance and core system performance necessary to meet the program goals can be achieved. The E3 core design and test results are described.
Core and Off-Core Processes in Systems Engineering
NASA Technical Reports Server (NTRS)
Breidenthal, Julian; Forsberg, Kevin
2010-01-01
An emerging methodology of organizing systems-engineering plans is based on a concept of core and off-core processes or activities. This concept has emerged as a result of recognition of a risk in the traditional representation of systems-engineering plans by a Vee model alone, according to which a large system is decomposed into levels of smaller subsystems, then integrated through levels of increasing scope until the full system is constructed. Actual systems-engineering activity is more complicated, raising the possibility that the staff will become confused in the absence of plans which explain the nature and ordering of work beyond the traditional Vee model.
An Efficient VLSI Architecture for Multi-Channel Spike Sorting Using a Generalized Hebbian Algorithm
Chen, Ying-Lun; Hwang, Wen-Jyi; Ke, Chi-En
2015-01-01
A novel VLSI architecture for multi-channel online spike sorting is presented in this paper. In the architecture, the spike detection is based on nonlinear energy operator (NEO), and the feature extraction is carried out by the generalized Hebbian algorithm (GHA). To lower the power consumption and area costs of the circuits, all of the channels share the same core for spike detection and feature extraction operations. Each channel has dedicated buffers for storing the detected spikes and the principal components of that channel. The proposed circuit also contains a clock gating system supplying the clock to only the buffers of channels currently using the computation core to further reduce the power consumption. The architecture has been implemented by an application-specific integrated circuit (ASIC) with 90-nm technology. Comparisons to the existing works show that the proposed architecture has lower power consumption and hardware area costs for real-time multi-channel spike detection and feature extraction. PMID:26287193
Chen, Ying-Lun; Hwang, Wen-Jyi; Ke, Chi-En
2015-08-13
A novel VLSI architecture for multi-channel online spike sorting is presented in this paper. In the architecture, the spike detection is based on nonlinear energy operator (NEO), and the feature extraction is carried out by the generalized Hebbian algorithm (GHA). To lower the power consumption and area costs of the circuits, all of the channels share the same core for spike detection and feature extraction operations. Each channel has dedicated buffers for storing the detected spikes and the principal components of that channel. The proposed circuit also contains a clock gating system supplying the clock to only the buffers of channels currently using the computation core to further reduce the power consumption. The architecture has been implemented by an application-specific integrated circuit (ASIC) with 90-nm technology. Comparisons to the existing works show that the proposed architecture has lower power consumption and hardware area costs for real-time multi-channel spike detection and feature extraction.
The research and application of multi-biometric acquisition embedded system
NASA Astrophysics Data System (ADS)
Deng, Shichao; Liu, Tiegen; Guo, Jingjing; Li, Xiuyan
2009-11-01
The identification technology based on multi-biometric can greatly improve the applicability, reliability and antifalsification. This paper presents a multi-biometric system bases on embedded system, which includes: three capture daughter boards are applied to obtain different biometric: one each for fingerprint, iris and vein of the back of hand; FPGA (Field Programmable Gate Array) is designed as coprocessor, which uses to configure three daughter boards on request and provides data path between DSP (digital signal processor) and daughter boards; DSP is the master processor and its functions include: control the biometric information acquisition, extracts feature as required and responsible for compare the results with the local database or data server through network communication. The advantages of this system were it can acquire three different biometric in real time, extracts complexity feature flexibly in different biometrics' raw data according to different purposes and arithmetic and network interface on the core-board will be the solution of big data scale. Because this embedded system has high stability, reliability, flexibility and fit for different data scale, it can satisfy the demand of multi-biometric recognition.
Outlook for grid service technologies within the @neurIST eHealth environment.
Arbona, A; Benkner, S; Fingberg, J; Frangi, A F; Hofmann, M; Hose, D R; Lonsdale, G; Ruefenacht, D; Viceconti, M
2006-01-01
The aim of the @neurIST project is to create an IT infrastructure for the management of all processes linked to research, diagnosis and treatment development for complex and multi-factorial diseases. The IT infrastructure will be developed for one such disease, cerebral aneurysm and subarachnoid haemorrhage, but its core technologies will be transferable to meet the needs of other medical areas. Since the IT infrastructure for @neurIST will need to encompass data repositories, computational analysis services and information systems handling multi-scale, multi-modal information at distributed sites, the natural basis for the IT infrastructure is a Grid Service middleware. The project will adopt a service-oriented architecture because it aims to provide a system addressing the needs of medical researchers, clinicians and health care specialists (and their IT providers/systems) and medical supplier/consulting industries.
A uniform approach for programming distributed heterogeneous computing systems
Grasso, Ivan; Pellegrini, Simone; Cosenza, Biagio; Fahringer, Thomas
2014-01-01
Large-scale compute clusters of heterogeneous nodes equipped with multi-core CPUs and GPUs are getting increasingly popular in the scientific community. However, such systems require a combination of different programming paradigms making application development very challenging. In this article we introduce libWater, a library-based extension of the OpenCL programming model that simplifies the development of heterogeneous distributed applications. libWater consists of a simple interface, which is a transparent abstraction of the underlying distributed architecture, offering advanced features such as inter-context and inter-node device synchronization. It provides a runtime system which tracks dependency information enforced by event synchronization to dynamically build a DAG of commands, on which we automatically apply two optimizations: collective communication pattern detection and device-host-device copy removal. We assess libWater’s performance in three compute clusters available from the Vienna Scientific Cluster, the Barcelona Supercomputing Center and the University of Innsbruck, demonstrating improved performance and scaling with different test applications and configurations. PMID:25844015
A uniform approach for programming distributed heterogeneous computing systems.
Grasso, Ivan; Pellegrini, Simone; Cosenza, Biagio; Fahringer, Thomas
2014-12-01
Large-scale compute clusters of heterogeneous nodes equipped with multi-core CPUs and GPUs are getting increasingly popular in the scientific community. However, such systems require a combination of different programming paradigms making application development very challenging. In this article we introduce libWater, a library-based extension of the OpenCL programming model that simplifies the development of heterogeneous distributed applications. libWater consists of a simple interface, which is a transparent abstraction of the underlying distributed architecture, offering advanced features such as inter-context and inter-node device synchronization. It provides a runtime system which tracks dependency information enforced by event synchronization to dynamically build a DAG of commands, on which we automatically apply two optimizations: collective communication pattern detection and device-host-device copy removal. We assess libWater's performance in three compute clusters available from the Vienna Scientific Cluster, the Barcelona Supercomputing Center and the University of Innsbruck, demonstrating improved performance and scaling with different test applications and configurations.
Large calculation of the flow over a hypersonic vehicle using a GPU
NASA Astrophysics Data System (ADS)
Elsen, Erich; LeGresley, Patrick; Darve, Eric
2008-12-01
Graphics processing units are capable of impressive computing performance up to 518 Gflops peak performance. Various groups have been using these processors for general purpose computing; most efforts have focussed on demonstrating relatively basic calculations, e.g. numerical linear algebra, or physical simulations for visualization purposes with limited accuracy. This paper describes the simulation of a hypersonic vehicle configuration with detailed geometry and accurate boundary conditions using the compressible Euler equations. To the authors' knowledge, this is the most sophisticated calculation of this kind in terms of complexity of the geometry, the physical model, the numerical methods employed, and the accuracy of the solution. The Navier-Stokes Stanford University Solver (NSSUS) was used for this purpose. NSSUS is a multi-block structured code with a provably stable and accurate numerical discretization which uses a vertex-based finite-difference method. A multi-grid scheme is used to accelerate the solution of the system. Based on a comparison of the Intel Core 2 Duo and NVIDIA 8800GTX, speed-ups of over 40× were demonstrated for simple test geometries and 20× for complex geometries.
NASA Astrophysics Data System (ADS)
Pommier, Anne; Laurenz, Vera; Davies, Christopher J.; Frost, Daniel J.
2018-05-01
We report an experimental investigation of phase equilibria in the Fe-S and Fe-S-O systems. Experiments were performed at high temperatures (1400-1850 °C) and high pressures (14 and 20 GPa) using a multi-anvil apparatus. The results of this study are used to understand the effect of sulfur and oxygen on core dynamics in small terrestrial bodies. We observe that the formation of solid FeO grains occurs at the Fe-S liquid - Fe solid interface at high temperature ( > 1400 °C at 20 GPa). Oxygen fugacities calculated for each O-bearing sample show that redox conditions vary from ΔIW = -0.65 to 0. Considering the relative density of each phase and existing evolutionary models of terrestrial cores, we apply our experimental results to the cores of Mars and Ganymede. We suggest that the presence of FeO in small terrestrial bodies tends to contribute to outer-core compositional stratification. Depending on the redox and thermal history of the planet, FeO may also help form a transitional redox zone at the core-mantle boundary.
High-capacity mixed fiber-wireless backhaul networks using MMW radio-over-MCF and MIMO
NASA Astrophysics Data System (ADS)
Pham, Thu A.; Pham, Hien T. T.; Le, Hai-Chau; Dang, Ngoc T.
2017-10-01
In this paper, we have proposed a high-capacity backhaul network, which is based on mixed fiber-wireless systems using millimeter-wave radio-over-multi-core fiber (MMW RoMCF) and multiple-input multiple-output (MIMO) transmission, for next generation mobile access networks. In addition, we also investigate the use of avalanche photodiode (APD) to improve capacity of the proposed backhaul downlink. We then theoretically analyze the system capacity comprehensively while considering various physical impairments including noise, MCF crosstalk, and fading modeled by Rician MIMO channel. The feasibility of the proposed backhaul architecture is verified via the numerical simulation experiments. The research results demonstrate that our developed backhaul solution can significantly enhance the backhaul capacity; the system capacity of 24 bps/Hz can be achieved with 20-km 8-core MCF and 8 × 8 MIMO transmitted over 100-m Rician fading link. It is also shown that the system performance, in term of channel capacity, strongly depend on the MCF inter-core crosstalk, which is governed by the mode coupling coefficient, the core pitch, and the bending radius.
Development INTERDATA 8/32 computer system
NASA Technical Reports Server (NTRS)
Sonett, C. P.
1983-01-01
The capabilities of the Interdata 8/32 minicomputer were examined regarding data and word processing, editing, retrieval, and budgeting as well as data management demands of the user groups in the network. Based on four projected needs: (1) a hands on (open shop) computer for data analysis with large core and disc capability; (2) the expected requirements of the NASA data networks; (3) the need for intermittent large core capacity for theoretical modeling; (4) the ability to access data rapidly either directly from tape or from core onto hard copy, the system proved useful and adequate for the planned requirements.
NASA Astrophysics Data System (ADS)
Greaves, M.; Elderfield, H.; Hodell, D. A.; Skinner, L. C.; Sevilgen, D.; Grauel, A. L.; de la Fuente, M.; Misra, S.
2014-12-01
A significant body of work exists on the chemistry of pore waters from DSDP and ODP drilling cores (e.g. Gieskes 1975; Sayles 1981) showing large gradients in sea salt cations and anions interpreted in terms of diagenetic reactions such as the formation of Mg-rich clays and dolomite formation (Higgins and Schrag, 2010). Another class of diagenetic reactions involves the breakdown of organic matter and trace element behaviour (Froelich et al., 1979). The translation of chemical gradients into fluxes requires estimates of pore water chemistry across the sea water - sediment surface boundary. Additionally, the use of the chemistry of benthic foraminiferal calcite for seawater paleochemistry requires estimation of the chemistry of pore waters which may differ from that of bottom seawater because of diagenetic reactions. In this work we have collected multi core samples from 10 core sites on cruise RRS James Cook JC089 on the southwest Iberian continental margin. Pore waters were extracted from the core surface and at 1 cm depth intervals down core (typically to ~40 cm depth) using Rhizon samplers and analysed for Alkalinity, DIC, ∂13C and Na, K, Mg, Ca, Li, Mn, Fe, Ba, B, Sr by atomic emission spectrophotometry as well as O2 penetration and pH by microelectrodes. This has allowed us to inspect chemical behavior at the bottom water - sediment interface. Some examples of results are a large gradient in ∂13C of DIC, the similarity of zero O2 penetration followed by an increase in Mn concentration and then decrease to zero, the similarity of Li to Mn and, in contrast to much DSDP/ODP work, Ca2+ and Mg2+both decrease with depth in pore waters near the sediment surface. References: Gieskes J.M. Annu. Rev. Earth Planet. Sci. 3, 433 (1975). Sayles F. L. Geochim. Cosmochim. Acta45, 1061 (1981). Higgins J.A. and D.P. Schrag. Geochim. Cosmochim. Acta.74, 5039 (2010). Froelich, P.N., et al., Geochim. Cosmochim. Acta. 43, 1075 (1979).
NASA Technical Reports Server (NTRS)
Adams, Marc A.; Zwissler, James G.; Hayes, Charles; Fabensky, Beth; Cornelison, Charles; Alexander, Lesley; Bishop, Karen
2005-01-01
A new technology is being developed that can protect spacecraft and satellite components against damage from meteoroid strikes and control the thermal environment of the protected components. This technology, called Foam Core Shield (FCS) systems, has the potential to replace the multi-layer insulation blankets (MLI) that have been used on spacecraft for decades. In order to be an attractive candidate for replacing MLI, FCS systems should not only provide superior protection against meteoroid strikes but also provide an equal or superior ability to control the temperature of the protected component. Properly designed FCS systems can provide these principal functions, meteoroid strike protection and thermal control, with lower system mass and a smaller system envelope than ML.
Multi-Modal Traveler Information System - Gateway Functional Requirements
DOT National Transportation Integrated Search
1997-11-17
The Multi-Modal Traveler Information System (MMTIS) project involves a large number of Intelligent Transportation System (ITS) related tasks. It involves research of all ITS initiatives in the Gary-Chicago-Milwaukee (GCM) Corridor which are currently...
Multi-Modal Traveler Information System - Gateway Interface Control Requirements
DOT National Transportation Integrated Search
1997-10-30
The Multi-Modal Traveler Information System (MMTIS) project involves a large number of Intelligent Transportation System (ITS) related tasks. It involves research of all ITS initiatives in the Gary-Chicago-Milwaukee (GCM) Corridor which are currently...
Rommel, Simon; Mendinueta, José Manuel Delgado; Klaus, Werner; Sakaguchi, Jun; Olmos, Juan José Vegas; Awaji, Yoshinari; Monroy, Idelfonso Tafur; Wada, Naoya
2017-09-18
This paper discusses spatially diverse optical vector network analysis for space division multiplexing (SDM) component and system characterization, which is becoming essential as SDM is widely considered to increase the capacity of optical communication systems. Characterization of a 108-channel photonic lantern spatial multiplexer, coupled to a 36-core 3-mode fiber, is experimentally demonstrated, extracting the full impulse response and complex transfer function matrices as well as insertion loss (IL) and mode-dependent loss (MDL) data. Moreover, the mode-mixing behavior of fiber splices in the few-mode multi-core fiber and their impact on system IL and MDL are analyzed, finding splices to cause significant mode-mixing and to be non-negligible in system capacity analysis.
Center for Technology for Advanced Scientific Componet Software (TASCS)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Govindaraju, Madhusudhan
Advanced Scientific Computing Research Computer Science FY 2010Report Center for Technology for Advanced Scientific Component Software: Distributed CCA State University of New York, Binghamton, NY, 13902 Summary The overall objective of Binghamton's involvement is to work on enhancements of the CCA environment, motivated by the applications and research initiatives discussed in the proposal. This year we are working on re-focusing our design and development efforts to develop proof-of-concept implementations that have the potential to significantly impact scientific components. We worked on developing parallel implementations for non-hydrostatic code and worked on a model coupling interface for biogeochemical computations coded in MATLAB.more » We also worked on the design and implementation modules that will be required for the emerging MapReduce model to be effective for scientific applications. Finally, we focused on optimizing the processing of scientific datasets on multi-core processors. Research Details We worked on the following research projects that we are working on applying to CCA-based scientific applications. 1. Non-Hydrostatic Hydrodynamics: Non-static hydrodynamics are significantly more accurate at modeling internal waves that may be important in lake ecosystems. Non-hydrostatic codes, however, are significantly more computationally expensive, often prohibitively so. We have worked with Chin Wu at the University of Wisconsin to parallelize non-hydrostatic code. We have obtained a speed up of about 26 times maximum. Although this is significant progress, we hope to improve the performance further, such that it becomes a practical alternative to hydrostatic codes. 2. Model-coupling for water-based ecosystems: To answer pressing questions about water resources requires that physical models (hydrodynamics) be coupled with biological and chemical models. Most hydrodynamics codes are written in Fortran, however, while most ecologists work in MATLAB. This disconnect creates a great barrier. To address this, we are working on a model coupling interface that will allow biogeochemical computations written in MATLAB to couple with Fortran codes. This will greatly improve the productivity of ecosystem scientists. 2. Low overhead and Elastic MapReduce Implementation Optimized for Memory and CPU-Intensive Applications: Since its inception, MapReduce has frequently been associated with Hadoop and large-scale datasets. Its deployment at Amazon in the cloud, and its applications at Yahoo! for large-scale distributed document indexing and database building, among other tasks, have thrust MapReduce to the forefront of the data processing application domain. The applicability of the paradigm however extends far beyond its use with data intensive applications and diskbased systems, and can also be brought to bear in processing small but CPU intensive distributed applications. MapReduce however carries its own burdens. Through experiments using Hadoop in the context of diverse applications, we uncovered latencies and delay conditions potentially inhibiting the expected performance of a parallel execution in CPU-intensive applications. Furthermore, as it currently stands, MapReduce is favored for data-centric applications, and as such tends to be solely applied to disk-based applications. The paradigm, falls short in bringing its novelty to diskless systems dedicated to in-memory applications, and compute intensive programs processing much smaller data, but requiring intensive computations. In this project, we focused both on the performance of processing large-scale hierarchical data in distributed scientific applications, as well as the processing of smaller but demanding input sizes primarily used in diskless, and memory resident I/O systems. We designed LEMO-MR [1], a Low overhead, elastic, configurable for in- memory applications, and on-demand fault tolerance, an optimized implementation of MapReduce, for both on disk and in memory applications. We conducted experiments to identify not only the necessary components of this model, but also trade offs and factors to be considered. We have initial results to show the efficacy of our implementation in terms of potential speedup that can be achieved for representative data sets used by cloud applications. We have quantified the performance gains exhibited by our MapReduce implementation over Apache Hadoop in a compute intensive environment. 3. Cache Performance Optimization for Processing XML and HDF-based Application Data on Multi-core Processors: It is important to design and develop scientific middleware libraries to harness the opportunities presented by emerging multi-core processors. Implementations of scientific middleware and applications that do not adapt to the programming paradigm when executing on emerging processors can severely impact the overall performance. In this project, we focused on the utilization of the L2 cache, which is a critical shared resource on chip multiprocessors (CMP). The access pattern of the shared L2 cache, which is dependent on how the application schedules and assigns processing work to each thread, can either enhance or hurt the ability to hide memory latency on a multi-core processor. Therefore, while processing scientific datasets such as HDF5, it is essential to conduct fine-grained analysis of cache utilization, to inform scheduling decisions in multi-threaded programming. In this project, using the TAU toolkit for performance feedback from dual- and quad-core machines, we conducted performance analysis and recommendations on how processing threads can be scheduled on multi-core nodes to enhance the performance of a class of scientific applications that requires processing of HDF5 data. In particular, we quantified the gains associated with the use of the adaptations we have made to the Cache-Affinity and Balanced-Set scheduling algorithms to improve L2 cache performance, and hence the overall application execution time [2]. References: 1. Zacharia Fadika, Madhusudhan Govindaraju, ``MapReduce Implementation for Memory-Based and Processing Intensive Applications'', accepted in 2nd IEEE International Conference on Cloud Computing Technology and Science, Indianapolis, USA, Nov 30 - Dec 3, 2010. 2. Rajdeep Bhowmik, Madhusudhan Govindaraju, ``Cache Performance Optimization for Processing XML-based Application Data on Multi-core Processors'', in proceedings of The 10th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing, May 17-20, 2010, Melbourne, Victoria, Australia. Contact Information: Madhusudhan Govindaraju Binghamton University State University of New York (SUNY) mgovinda@cs.binghamton.edu Phone: 607-777-4904« less
Plate & tube bridge deck evaluation in the deck test bed of the Troutville, Virginia, weigh station.
DOT National Transportation Integrated Search
2004-01-01
This report addresses the laboratory and field performance of multi-cellular fiber-reinforced polymer (FRP) composite bridge deck systems. We focus specifically on FRP decks produced from adhesively bonded pultrusions where the core of the deck posse...
NASA Astrophysics Data System (ADS)
Leggett, C.; Binet, S.; Jackson, K.; Levinthal, D.; Tatarkhanov, M.; Yao, Y.
2011-12-01
Thermal limitations have forced CPU manufacturers to shift from simply increasing clock speeds to improve processor performance, to producing chip designs with multi- and many-core architectures. Further the cores themselves can run multiple threads as a zero overhead context switch allowing low level resource sharing (Intel Hyperthreading). To maximize bandwidth and minimize memory latency, memory access has become non uniform (NUMA). As manufacturers add more cores to each chip, a careful understanding of the underlying architecture is required in order to fully utilize the available resources. We present AthenaMP and the Atlas event loop manager, the driver of the simulation and reconstruction engines, which have been rewritten to make use of multiple cores, by means of event based parallelism, and final stage I/O synchronization. However, initial studies on 8 andl6 core Intel architectures have shown marked non-linearities as parallel process counts increase, with as much as 30% reductions in event throughput in some scenarios. Since the Intel Nehalem architecture (both Gainestown and Westmere) will be the most common choice for the next round of hardware procurements, an understanding of these scaling issues is essential. Using hardware based event counters and Intel's Performance Tuning Utility, we have studied the performance bottlenecks at the hardware level, and discovered optimization schemes to maximize processor throughput. We have also produced optimization mechanisms, common to all large experiments, that address the extreme nature of today's HEP code, which due to it's size, places huge burdens on the memory infrastructure of today's processors.
Parallel Agent-Based Simulations on Clusters of GPUs and Multi-Core Processors
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aaby, Brandon G; Perumalla, Kalyan S; Seal, Sudip K
2010-01-01
An effective latency-hiding mechanism is presented in the parallelization of agent-based model simulations (ABMS) with millions of agents. The mechanism is designed to accommodate the hierarchical organization as well as heterogeneity of current state-of-the-art parallel computing platforms. We use it to explore the computation vs. communication trade-off continuum available with the deep computational and memory hierarchies of extant platforms and present a novel analytical model of the tradeoff. We describe our implementation and report preliminary performance results on two distinct parallel platforms suitable for ABMS: CUDA threads on multiple, networked graphical processing units (GPUs), and pthreads on multi-core processors. Messagemore » Passing Interface (MPI) is used for inter-GPU as well as inter-socket communication on a cluster of multiple GPUs and multi-core processors. Results indicate the benefits of our latency-hiding scheme, delivering as much as over 100-fold improvement in runtime for certain benchmark ABMS application scenarios with several million agents. This speed improvement is obtained on our system that is already two to three orders of magnitude faster on one GPU than an equivalent CPU-based execution in a popular simulator in Java. Thus, the overall execution of our current work is over four orders of magnitude faster when executed on multiple GPUs.« less
Neural simulations on multi-core architectures.
Eichner, Hubert; Klug, Tobias; Borst, Alexander
2009-01-01
Neuroscience is witnessing increasing knowledge about the anatomy and electrophysiological properties of neurons and their connectivity, leading to an ever increasing computational complexity of neural simulations. At the same time, a rather radical change in personal computer technology emerges with the establishment of multi-cores: high-density, explicitly parallel processor architectures for both high performance as well as standard desktop computers. This work introduces strategies for the parallelization of biophysically realistic neural simulations based on the compartmental modeling technique and results of such an implementation, with a strong focus on multi-core architectures and automation, i.e. user-transparent load balancing.
Neural Simulations on Multi-Core Architectures
Eichner, Hubert; Klug, Tobias; Borst, Alexander
2009-01-01
Neuroscience is witnessing increasing knowledge about the anatomy and electrophysiological properties of neurons and their connectivity, leading to an ever increasing computational complexity of neural simulations. At the same time, a rather radical change in personal computer technology emerges with the establishment of multi-cores: high-density, explicitly parallel processor architectures for both high performance as well as standard desktop computers. This work introduces strategies for the parallelization of biophysically realistic neural simulations based on the compartmental modeling technique and results of such an implementation, with a strong focus on multi-core architectures and automation, i.e. user-transparent load balancing. PMID:19636393
A Systematic Study of Explosions in Core Collapse Supernovae
NASA Technical Reports Server (NTRS)
Swesty, F. Douglas; Mihalas, Dimitri; Norman, Michael
1997-01-01
This report covers the research conducted from September 1996 to August 1997 (eighteen months into the three year grant). We have obtained a number of significant findings based on the on the work that we have conducted under this grant during the past year. As we stated in our original proposal the work has focused on multi-dimensional models of the convective epoch in core collapse supernovae. During the past year we have developed a large number of models of the convective epoch in 2-D under two levels of neutrino transport approximation and we are currently working on 3-D models. In the following pages will endeavor to give brief descriptions of our results.
Multi-Modal Traveler Information System - GCM Corridor Architecture Interface Control Requirements
DOT National Transportation Integrated Search
1997-10-31
The Multi-Modal Traveler Information System (MMTIS) project involves a large number of Intelligent Transportation System (ITS) related tasks. It involves research of all ITS initiatives in the Gary-Chicago-Milwaukee (GCM) Corridor which are currently...
Multi-Modal Traveler Information System - GCM Corridor Architecture Functional Requirements
DOT National Transportation Integrated Search
1997-11-17
The Multi-Modal Traveler Information System (MMTIS) project involves a large number of Intelligent Transportation System (ITS) related tasks. It involves research of all ITS initiatives in the Gary-Chicago-Milwaukee (GCM) Corridor which are currently...
NASA Astrophysics Data System (ADS)
Li, Peng; Wu, Di
2018-01-01
Two competing approaches have been developed over the years for multi-echelon inventory system optimization, stochastic-service approach (SSA) and guaranteed-service approach (GSA). Although they solve the same inventory policy optimization problem in their core, they make different assumptions with regard to the role of safety stock. This paper provides a detailed comparison of the two approaches by considering operating flexibility costs in the optimization of (R, Q) policies for a continuous review serial inventory system. The results indicate the GSA model is more efficiency in solving the complicated inventory problem in terms of the computation time, and the cost difference of the two approaches is quite small.
Recent Developments in Smart Adaptive Structures for Solar Sailcraft
NASA Technical Reports Server (NTRS)
Whorton, M. S.; Kim, Y. K.; Oakley, J.; Adetona, O.; Keel, L. H.
2007-01-01
The "Smart Adaptive Structures for Solar Sailcraft" development activity at MSFC has investigated issues associated with understanding how to model and scale the subsystem and multi-body system dynamics of a gossamer solar sailcraft with the objective of designing sailcraft attitude control systems. This research and development activity addressed three key tasks that leveraged existing facilities and core competencies of MSFC to investigate dynamics and control issues of solar sails. Key aspects of this effort included modeling and testing of a 30 m deployable boom; modeling of the multi-body system dynamics of a gossamer sailcraft; investigation of control-structures interaction for gossamer sailcraft; and development and experimental demonstration of adaptive control technologies to mitigate control-structures interaction.
Electronic Structure Calculations and Adaptation Scheme in Multi-core Computing Environments
DOE Office of Scientific and Technical Information (OSTI.GOV)
Seshagiri, Lakshminarasimhan; Sosonkina, Masha; Zhang, Zhao
2009-05-20
Multi-core processing environments have become the norm in the generic computing environment and are being considered for adding an extra dimension to the execution of any application. The T2 Niagara processor is a very unique environment where it consists of eight cores having a capability of running eight threads simultaneously in each of the cores. Applications like General Atomic and Molecular Electronic Structure (GAMESS), used for ab-initio molecular quantum chemistry calculations, can be good indicators of the performance of such machines and would be a guideline for both hardware designers and application programmers. In this paper we try to benchmarkmore » the GAMESS performance on a T2 Niagara processor for a couple of molecules. We also show the suitability of using a middleware based adaptation algorithm on GAMESS on such a multi-core environment.« less
NASA Astrophysics Data System (ADS)
Pearce, R.; Mitchell, T. M.; Moorkamp, M.; Araya, J.; Cembrano, J. M.; Yanez, G. A.; Hammond, J. O. S.
2017-12-01
At convergent plate boundaries, volcanic orogeny is largely controlled by major thrust fault systems that act as magmatic and hydrothermal fluid conduits through the crust. In the south-central Andes, the volcanically and seismically active Tinguiririca and Planchon-Peteroa volcanoes are considered to be tectonically related to the major El Fierro thrust fault system. These large scale reverse faults are characterized by 500 - 1000m wide hydrothermally altered fault cores, which possess a distinct conductive signature relative to surrounding lithology. In order to establish the subsurface architecture of these fault systems, such conductivity contrasts can be detected using the magnetotelluric method. In this study, LEMI fluxgate-magnetometer long-period and Metronix broadband MT data were collected at 21 sites in a 40km2 survey grid that surrounds this fault system and associated volcanic complexes. Multi-remote referencing techniques is used together with robust processing to obtain reliable impedance estimates between 100 Hz and 1,000s. Our preliminary inversion results provide evidence of structures within the 10 - 20 km depth range that are attributed to this fault system. Further inversions will be conducted to determine the approximate depth extent of these features, and ultimately provide constraints for future geophysical studies aimed to deduce the role of these faults in volcanic orogeny and hydrothermal fluid migration processes in this region of the Andes.
Einstein@Home Discovery of 24 Pulsars in the Parkes Multi-beam Pulsar Survey
NASA Astrophysics Data System (ADS)
Knispel, B.; Eatough, R. P.; Kim, H.; Keane, E. F.; Allen, B.; Anderson, D.; Aulbert, C.; Bock, O.; Crawford, F.; Eggenstein, H.-B.; Fehrmann, H.; Hammer, D.; Kramer, M.; Lyne, A. G.; Machenschalk, B.; Miller, R. B.; Papa, M. A.; Rastawicki, D.; Sarkissian, J.; Siemens, X.; Stappers, B. W.
2013-09-01
We have conducted a new search for radio pulsars in compact binary systems in the Parkes multi-beam pulsar survey (PMPS) data, employing novel methods to remove the Doppler modulation from binary motion. This has yielded unparalleled sensitivity to pulsars in compact binaries. The required computation time of ≈17, 000 CPU core years was provided by the distributed volunteer computing project Einstein@Home, which has a sustained computing power of about 1 PFlop s-1. We discovered 24 new pulsars in our search, 18 of which were isolated pulsars, and 6 were members of binary systems. Despite the wide filterbank channels and relatively slow sampling time of the PMPS data, we found pulsars with very large ratios of dispersion measure (DM) to spin period. Among those is PSR J1748-3009, the millisecond pulsar with the highest known DM (≈420 pc cm-3). We also discovered PSR J1840-0643, which is in a binary system with an orbital period of 937 days, the fourth largest known. The new pulsar J1750-2536 likely belongs to the rare class of intermediate-mass binary pulsars. Three of the isolated pulsars show long-term nulling or intermittency in their emission, further increasing this growing family. Our discoveries demonstrate the value of distributed volunteer computing for data-driven astronomy and the importance of applying new analysis methods to extensively searched data.
Compact Multimedia Systems in Multi-chip Module Technology
NASA Technical Reports Server (NTRS)
Fang, Wai-Chi; Alkalaj, Leon
1995-01-01
This tutorial paper shows advanced multimedia system designs based on multi-chip module (MCM) technologies that provide essential computing, compression, communication, and storage capabilities for various large scale information highway applications.!.
The Evolvable Advanced Multi-Mission Operations System (AMMOS): Making Systems Interoperable
NASA Technical Reports Server (NTRS)
Ko, Adans Y.; Maldague, Pierre F.; Bui, Tung; Lam, Doris T.; McKinney, John C.
2010-01-01
The Advanced Multi-Mission Operations System (AMMOS) provides a common Mission Operation System (MOS) infrastructure to NASA deep space missions. The evolution of AMMOS has been driven by two factors: increasingly challenging requirements from space missions, and the emergence of new IT technology. The work described in this paper focuses on three key tasks related to IT technology requirements: first, to eliminate duplicate functionality; second, to promote the use of loosely coupled application programming interfaces, text based file interfaces, web-based frameworks and integrated Graphical User Interfaces (GUI) to connect users, data, and core functionality; and third, to build, develop, and deploy AMMOS services that are reusable, agile, adaptive to project MOS configurations, and responsive to industrially endorsed information technology standards.
Equation of state and critical point behavior of hard-core double-Yukawa fluids.
Montes, J; Robles, M; López de Haro, M
2016-02-28
A theoretical study on the equation of state and the critical point behavior of hard-core double-Yukawa fluids is presented. Thermodynamic perturbation theory, restricted to first order in the inverse temperature and having the hard-sphere fluid as the reference system, is used to derive a relatively simple analytical equation of state of hard-core multi-Yukawa fluids. Using such an equation of state, the compressibility factor and phase behavior of six representative hard-core double-Yukawa fluids are examined and compared with available simulation results. The effect of varying the parameters of the hard-core double-Yukawa intermolecular potential on the location of the critical point is also analyzed using different perspectives. The relevance of this analysis for fluids whose molecules interact with realistic potentials is also pointed out.
HACC: Simulating sky surveys on state-of-the-art supercomputing architectures
NASA Astrophysics Data System (ADS)
Habib, Salman; Pope, Adrian; Finkel, Hal; Frontiere, Nicholas; Heitmann, Katrin; Daniel, David; Fasel, Patricia; Morozov, Vitali; Zagaris, George; Peterka, Tom; Vishwanath, Venkatram; Lukić, Zarija; Sehrish, Saba; Liao, Wei-keng
2016-01-01
Current and future surveys of large-scale cosmic structure are associated with a massive and complex datastream to study, characterize, and ultimately understand the physics behind the two major components of the 'Dark Universe', dark energy and dark matter. In addition, the surveys also probe primordial perturbations and carry out fundamental measurements, such as determining the sum of neutrino masses. Large-scale simulations of structure formation in the Universe play a critical role in the interpretation of the data and extraction of the physics of interest. Just as survey instruments continue to grow in size and complexity, so do the supercomputers that enable these simulations. Here we report on HACC (Hardware/Hybrid Accelerated Cosmology Code), a recently developed and evolving cosmology N-body code framework, designed to run efficiently on diverse computing architectures and to scale to millions of cores and beyond. HACC can run on all current supercomputer architectures and supports a variety of programming models and algorithms. It has been demonstrated at scale on Cell- and GPU-accelerated systems, standard multi-core node clusters, and Blue Gene systems. HACC's design allows for ease of portability, and at the same time, high levels of sustained performance on the fastest supercomputers available. We present a description of the design philosophy of HACC, the underlying algorithms and code structure, and outline implementation details for several specific architectures. We show selected accuracy and performance results from some of the largest high resolution cosmological simulations so far performed, including benchmarks evolving more than 3.6 trillion particles.
HACC: Simulating sky surveys on state-of-the-art supercomputing architectures
DOE Office of Scientific and Technical Information (OSTI.GOV)
Habib, Salman; Pope, Adrian; Finkel, Hal
2016-01-01
Current and future surveys of large-scale cosmic structure are associated with a massive and complex datastream to study, characterize, and ultimately understand the physics behind the two major components of the ‘Dark Universe’, dark energy and dark matter. In addition, the surveys also probe primordial perturbations and carry out fundamental measurements, such as determining the sum of neutrino masses. Large-scale simulations of structure formation in the Universe play a critical role in the interpretation of the data and extraction of the physics of interest. Just as survey instruments continue to grow in size and complexity, so do the supercomputers thatmore » enable these simulations. Here we report on HACC (Hardware/Hybrid Accelerated Cosmology Code), a recently developed and evolving cosmology N-body code framework, designed to run efficiently on diverse computing architectures and to scale to millions of cores and beyond. HACC can run on all current supercomputer architectures and supports a variety of programming models and algorithms. It has been demonstrated at scale on Cell- and GPU-accelerated systems, standard multi-core node clusters, and Blue Gene systems. HACC’s design allows for ease of portability, and at the same time, high levels of sustained performance on the fastest supercomputers available. We present a description of the design philosophy of HACC, the underlying algorithms and code structure, and outline implementation details for several specific architectures. We show selected accuracy and performance results from some of the largest high resolution cosmological simulations so far performed, including benchmarks evolving more than 3.6 trillion particles.« less
Nonlinear dynamic simulation of single- and multi-spool core engines
NASA Technical Reports Server (NTRS)
Schobeiri, T.; Lippke, C.; Abouelkheir, M.
1993-01-01
In this paper a new computational method for accurate simulation of the nonlinear dynamic behavior of single- and multi-spool core engines, turbofan engines, and power generation gas turbine engines is presented. In order to perform the simulation, a modularly structured computer code has been developed which includes individual mathematical modules representing various engine components. The generic structure of the code enables the dynamic simulation of arbitrary engine configurations ranging from single-spool thrust generation to multi-spool thrust/power generation engines under adverse dynamic operating conditions. For precise simulation of turbine and compressor components, row-by-row calculation procedures were implemented that account for the specific turbine and compressor cascade and blade geometry and characteristics. The dynamic behavior of the subject engine is calculated by solving a number of systems of partial differential equations, which describe the unsteady behavior of the individual components. In order to ensure the capability, accuracy, robustness, and reliability of the code, comprehensive critical performance assessment and validation tests were performed. As representatives, three different transient cases with single- and multi-spool thrust and power generation engines were simulated. The transient cases range from operating with a prescribed fuel schedule, to extreme load changes, to generator and turbine shut down.
Kalman Filter Tracking on Parallel Architectures
NASA Astrophysics Data System (ADS)
Cerati, Giuseppe; Elmer, Peter; Krutelyov, Slava; Lantz, Steven; Lefebvre, Matthieu; McDermott, Kevin; Riley, Daniel; Tadel, Matevž; Wittich, Peter; Würthwein, Frank; Yagil, Avi
2016-11-01
Power density constraints are limiting the performance improvements of modern CPUs. To address this we have seen the introduction of lower-power, multi-core processors such as GPGPU, ARM and Intel MIC. In order to achieve the theoretical performance gains of these processors, it will be necessary to parallelize algorithms to exploit larger numbers of lightweight cores and specialized functions like large vector units. Track finding and fitting is one of the most computationally challenging problems for event reconstruction in particle physics. At the High-Luminosity Large Hadron Collider (HL-LHC), for example, this will be by far the dominant problem. The need for greater parallelism has driven investigations of very different track finding techniques such as Cellular Automata or Hough Transforms. The most common track finding techniques in use today, however, are those based on a Kalman filter approach. Significant experience has been accumulated with these techniques on real tracking detector systems, both in the trigger and offline. They are known to provide high physics performance, are robust, and are in use today at the LHC. Given the utility of the Kalman filter in track finding, we have begun to port these algorithms to parallel architectures, namely Intel Xeon and Xeon Phi. We report here on our progress towards an end-to-end track reconstruction algorithm fully exploiting vectorization and parallelization techniques in a simplified experimental environment.
The United States Environmental Protection Agency (EPA) is developing a comprehensive environmental exposure and risk analysis software system for agency-wide application using the methodology of a Multi-media, Multi-pathway, Multi-receptor Risk Assessment (3MRA) model. This sof...
The U.S. Environmental Protection Agency (EPA) is developing a comprehensive environmental exposure and risk analysis software system for agency-wide application using the methodology of a Multi-media, Multi-pathway, Multi-receptor Risk Assessment (3MRA) model. This software sys...
NASA Astrophysics Data System (ADS)
Chen, Peng; Xue, Zugang; Tian, Youmei; Zhao, Zheming; Wang, Xunsi; Liu, Zijun; Zhang, Peiqing; Dai, Shixun; Nie, Qiuhua; Wang, Rongping
2018-06-01
Two kinds of step-index As-S fibers have been fabricated by an isolated extrusion method with a numerical aperture (NA) of 0.52, but with different core size of 10 or 50 µm. With a femtosecond laser pumping, their supercontinnum (SC) generation spectra were recorded in order to testify the effect of high-order modes on SC generation. The spectra spanning from 1.0 to 6.7 µm and from 1.5 to 8.6 µm can be obtained in a 16-cm-long fiber with 10 µm-core diameter pumping by central wavelength of 2.9 and 4.0 µm, respectively. The results show that high-order modes would deplete the spectra spanning in red-shifting part. The SC generation in small-core fiber is much more efficient than that in large-core fiber. This is the first comparative investigation on the SC generation from the quasi single- and multi-mode ChG fibers under the same conditions.
NASA Astrophysics Data System (ADS)
Rajchl, Martin; Abhari, Kamyar; Stirrat, John; Ukwatta, Eranga; Cantor, Diego; Li, Feng P.; Peters, Terry M.; White, James A.
2014-03-01
Multi-center trials provide the unique ability to investigate novel techniques across a range of geographical sites with sufficient statistical power, the inclusion of multiple operators determining feasibility under a wider array of clinical environments and work-flows. For this purpose, we introduce a new means of distributing pre-procedural cardiac models for image-guided interventions across a large scale multi-center trial. In this method, a single core facility is responsible for image processing, employing a novel web-based interface for model visualization and distribution. The requirements for such an interface, being WebGL-based, are minimal and well within the realms of accessibility for participating centers. We then demonstrate the accuracy of our approach using a single-center pacemaker lead implantation trial with generic planning models.
Optimization on fixed low latency implementation of the GBT core in FPGA
Chen, K.; Chen, H.; Wu, W.; ...
2017-07-11
We present that in the upgrade of ATLAS experiment, the front-end electronics components are subjected to a large radiation background. Meanwhile high speed optical links are required for the data transmission between the on-detector and off-detector electronics. The GBT architecture and the Versatile Link (VL) project are designed by CERN to support the 4.8 Gbps line rate bidirectional high-speed data transmission which is called GBT link. In the ATLAS upgrade, besides the link with on-detector, the GBT link is also used between different off-detector systems. The GBTX ASIC is designed for the on-detector front-end, correspondingly for the off-detector electronics, themore » GBT architecture is implemented in Field Programmable Gate Arrays (FPGA). CERN launches the GBT-FPGA project to provide examples in different types of FPGA. In the ATLAS upgrade framework, the Front-End LInk eXchange (FELIX) system is used to interface the front end electronics of several ATLAS subsystems. The GBT link is used between them, to transfer the detector data and the timing, trigger, control and monitoring information. The trigger signal distributed in the down-link from FELIX to the front-end requires a fixed and low latency. In this paper, several optimizations on the GBT-FPGA IP core are introduced, to achieve a lower fixed latency. For FELIX, a common firmware will be used to interface different front-ends with support of both GBT modes: the forward error correction mode and the wide mode. The modified GBT-FPGA core has the ability to switch between the GBT modes without FPGA reprogramming. Finally, the system clock distribution of the multi-channel FELIX firmware is also discussed in this paper.« less
Optimization on fixed low latency implementation of the GBT core in FPGA
NASA Astrophysics Data System (ADS)
Chen, K.; Chen, H.; Wu, W.; Xu, H.; Yao, L.
2017-07-01
In the upgrade of ATLAS experiment [1], the front-end electronics components are subjected to a large radiation background. Meanwhile high speed optical links are required for the data transmission between the on-detector and off-detector electronics. The GBT architecture and the Versatile Link (VL) project are designed by CERN to support the 4.8 Gbps line rate bidirectional high-speed data transmission which is called GBT link [2]. In the ATLAS upgrade, besides the link with on-detector, the GBT link is also used between different off-detector systems. The GBTX ASIC is designed for the on-detector front-end, correspondingly for the off-detector electronics, the GBT architecture is implemented in Field Programmable Gate Arrays (FPGA). CERN launches the GBT-FPGA project to provide examples in different types of FPGA [3]. In the ATLAS upgrade framework, the Front-End LInk eXchange (FELIX) system [4, 5] is used to interface the front-end electronics of several ATLAS subsystems. The GBT link is used between them, to transfer the detector data and the timing, trigger, control and monitoring information. The trigger signal distributed in the down-link from FELIX to the front-end requires a fixed and low latency. In this paper, several optimizations on the GBT-FPGA IP core are introduced, to achieve a lower fixed latency. For FELIX, a common firmware will be used to interface different front-ends with support of both GBT modes: the forward error correction mode and the wide mode. The modified GBT-FPGA core has the ability to switch between the GBT modes without FPGA reprogramming. The system clock distribution of the multi-channel FELIX firmware is also discussed in this paper.
Sam2bam: High-Performance Framework for NGS Data Preprocessing Tools
Cheng, Yinhe; Tzeng, Tzy-Hwa Kathy
2016-01-01
This paper introduces a high-throughput software tool framework called sam2bam that enables users to significantly speed up pre-processing for next-generation sequencing data. The sam2bam is especially efficient on single-node multi-core large-memory systems. It can reduce the runtime of data pre-processing in marking duplicate reads on a single node system by 156–186x compared with de facto standard tools. The sam2bam consists of parallel software components that can fully utilize multiple processors, available memory, high-bandwidth storage, and hardware compression accelerators, if available. The sam2bam provides file format conversion between well-known genome file formats, from SAM to BAM, as a basic feature. Additional features such as analyzing, filtering, and converting input data are provided by using plug-in tools, e.g., duplicate marking, which can be attached to sam2bam at runtime. We demonstrated that sam2bam could significantly reduce the runtime of next generation sequencing (NGS) data pre-processing from about two hours to about one minute for a whole-exome data set on a 16-core single-node system using up to 130 GB of memory. The sam2bam could reduce the runtime of NGS data pre-processing from about 20 hours to about nine minutes for a whole-genome sequencing data set on the same system using up to 711 GB of memory. PMID:27861637
Planning and Execution: The Spirit of Opportunity for Robust Autonomous Systems
NASA Technical Reports Server (NTRS)
Muscettola, Nicola
2004-01-01
One of the most exciting endeavors pursued by human kind is the search for life in the Solar System and the Universe at large. NASA is leading this effort by designing, deploying and operating robotic systems that will reach planets, planet moons, asteroids and comets searching for water, organic building blocks and signs of past or present microbial life. None of these missions will be achievable without substantial advances in.the design, implementation and validation of autonomous control agents. These agents must be capable of robustly controlling a robotic explorer in a hostile environment with very limited or no communication with Earth. The talk focuses on work pursued at the NASA Ames Research center ranging from basic research on algorithm to deployed mission support systems. We will start by discussing how planning and scheduling technology derived from the Remote Agent experiment is being used daily in the operations of the Spirit and Opportunity rovers. Planning and scheduling is also used as the fundamental paradigm at the core of our research in real-time autonomous agents. In particular, we will describe our efforts in the Intelligent Distributed Execution Architecture (IDEA), a multi-agent real-time architecture that exploits artificial intelligence planning as the core reasoning engine of an autonomous agent. We will also describe how the issue of plan robustness at execution can be addressed by novel constraint propagation algorithms capable of giving the tightest exact bounds on resource consumption or all possible executions of a flexible plan.
NASA Astrophysics Data System (ADS)
Yu, Yanan; Wang, Xiaoxun; He, Chengcheng; Lai, Chenlong; Liu, Yuanchao
2015-11-01
For overcoming the problems such as remote operation and dangerous tasks, multi-terminal remote monitoring and warning system based on STC89C52 Micro Control Unit and wireless communication technique was proposed. The system with MCU as its core adopted multiple sets of sensor device to monitor environment parameters of different locations, such as temperature, humidity, smoke other harmful gas concentration. Data information collected was transmitted remotely by wireless transceiver module, and then multi-channel data parameter was processed and displayed through serial communication protocol between the module and PC. The results of system could be checked in the form of web pages within a local network which plays a wireless monitoring and warning role. In a remote operation, four-rotor micro air vehicle which fixed airborne data acquisition device was utilized as a middleware between collecting terminal and PC to increase monitoring scope. Whole test system has characteristics of simple construction, convenience, real time ability and high reliability, which could meet the requirements of actual use.
High numerical aperture large-core photonic crystal fiber for a broadband infrared transmission
NASA Astrophysics Data System (ADS)
Pniewski, J.; Stepniewski, G.; Kasztelanic, R.; Siwicki, B.; Pierscinska, D.; Pierscinski, K.; Pysz, D.; Borzycki, K.; Stepien, R.; Bugajski, M.; Buczynski, R.
2016-11-01
In this paper we present a large mode area photonic crystal fiber made of the heavy metal oxide glass CS-740, dedicated for a broadband light guidance in the visible, near- and mid-infrared regions of wavelengths from 0.4 to 4.7 μm. The fiber is effectively multi-mode in the considered wavelength range. It is composed of a ring of air-holes surrounding the core, with a high linear filling factor of 0.97. The fiber was made using a standard stack-and-draw technique. Each hole has a size of approx. 2.5 × 3.0 μm and diameter of core is 80 μm. Fiber attenuation is below 3 dB/m in the 0.9-1.7 μm wavelength range, while at 4.4 μm (mid-IR) it is approx. 5 dB/cm. Bending loss at the 1.55 μm wavelength is 0.45 dB per loop of 8 mm radius. Fiber numerical aperture is 0.53 at 1.55 μm. The effective mode area of the fundamental mode is approx. 2400 μm2 in the wavelength range of 0.8-1.7 μm. We present a proof-of-concept demonstration that our large core photonic crystal fiber is able to efficiently collect light directly from a mid-IR quantum cascade laser without use of additional optics and can be used for pigtailing mid-IR sources and detectors.
A Multi-scale Cognitive Approach to Intrusion Detection and Response
2015-12-28
the behavior of the traffic on the network, either by using mathematical formulas or by replaying packet streams. As a result, simulators depend...large scale. Summary of the most important results We obtained a powerful machine, which has 768 cores and 1.25 TB memory . RBG has been...time. Each client is configured with 1GB memory , 10 GB disk space, and one 100M Ethernet interface. The server nodes include web servers
A multimodal image sensor system for identifying water stress in grapevines
NASA Astrophysics Data System (ADS)
Zhao, Yong; Zhang, Qin; Li, Minzan; Shao, Yongni; Zhou, Jianfeng; Sun, Hong
2012-11-01
Water stress is one of the most common limitations of fruit growth. Water is the most limiting resource for crop growth. In grapevines, as well as in other fruit crops, fruit quality benefits from a certain level of water deficit which facilitates to balance vegetative and reproductive growth and the flow of carbohydrates to reproductive structures. A multi-modal sensor system was designed to measure the reflectance signature of grape plant surfaces and identify different water stress levels in this paper. The multi-modal sensor system was equipped with one 3CCD camera (three channels in R, G, and IR). The multi-modal sensor can capture and analyze grape canopy from its reflectance features, and identify the different water stress levels. This research aims at solving the aforementioned problems. The core technology of this multi-modal sensor system could further be used as a decision support system that combines multi-modal sensory data to improve plant stress detection and identify the causes of stress. The images were taken by multi-modal sensor which could output images in spectral bands of near-infrared, green and red channel. Based on the analysis of the acquired images, color features based on color space and reflectance features based on image process method were calculated. The results showed that these parameters had the potential as water stress indicators. More experiments and analysis are needed to validate the conclusion.
NASA Astrophysics Data System (ADS)
Bourdine, Anton V.; Zhukov, Alexander E.
2017-04-01
High bit rate laser-based data transmission over silica optical fibers with enlarged core diameter in comparison with standard singlemode fibers is found variety infocommunication applications. Since IEEE 802.3z standard was ratified on 1998 this technique started to be widely used for short-range in-premises distributed multi-Gigabit networks based on new generation laser optimized multimode fibers 50/125 of Cat. OM2…OM4. Nowadays it becomes to be in demand for on-board cable systems and industrial network applications requiring 1Gps and more bit rates over fibers with extremely enlarged core diameter up to 100 μm. This work presents an alternative method for design the special refractive index profiles of silica few-mode fibers with extremely enlarged core diameter, that provides modal bandwidth enhancing under a few-mode regime of laser-based data optical transmission. Here some results are presented concerning with refractive index profile synthesis for few-mode fibers with reduced differential mode delay for "O"-band central region, as well as computed differential mode delay spectral curves corresponding to profiles for fibers 50/125 and 100/125 for in-premises and on-board/industrial cable systems.
Enhancing Image Processing Performance for PCID in a Heterogeneous Network of Multi-core Processors
2009-09-01
TFLOPS of Playstation 3 (PS3) nodes with IBM Cell Broadband Engine multi-cores and 15 dual-quad Xeon head nodes. The interconnect fabric includes... 4 3. INFORMATION MANAGEMENT FOR PARALLELIZATION AND...STREAMING............................................................. 7 4 . RESULTS
T-L Plane Abstraction-Based Energy-Efficient Real-Time Scheduling for Multi-Core Wireless Sensors
Kim, Youngmin; Lee, Ki-Seong; Pham, Ngoc-Son; Lee, Sun-Ro; Lee, Chan-Gun
2016-01-01
Energy efficiency is considered as a critical requirement for wireless sensor networks. As more wireless sensor nodes are equipped with multi-cores, there are emerging needs for energy-efficient real-time scheduling algorithms. The T-L plane-based scheme is known to be an optimal global scheduling technique for periodic real-time tasks on multi-cores. Unfortunately, there has been a scarcity of studies on extending T-L plane-based scheduling algorithms to exploit energy-saving techniques. In this paper, we propose a new T-L plane-based algorithm enabling energy-efficient real-time scheduling on multi-core sensor nodes with dynamic power management (DPM). Our approach addresses the overhead of processor mode transitions and reduces fragmentations of the idle time, which are inherent in T-L plane-based algorithms. Our experimental results show the effectiveness of the proposed algorithm compared to other energy-aware scheduling methods on T-L plane abstraction. PMID:27399722
A pluggable framework for parallel pairwise sequence search.
Archuleta, Jeremy; Feng, Wu-chun; Tilevich, Eli
2007-01-01
The current and near future of the computing industry is one of multi-core and multi-processor technology. Most existing sequence-search tools have been designed with a focus on single-core, single-processor systems. This discrepancy between software design and hardware architecture substantially hinders sequence-search performance by not allowing full utilization of the hardware. This paper presents a novel framework that will aid the conversion of serial sequence-search tools into a parallel version that can take full advantage of the available hardware. The framework, which is based on a software architecture called mixin layers with refined roles, enables modules to be plugged into the framework with minimal effort. The inherent modular design improves maintenance and extensibility, thus opening up a plethora of opportunities for advanced algorithmic features to be developed and incorporated while routine maintenance of the codebase persists.
Improvement of Speckle Contrast Image Processing by an Efficient Algorithm.
Steimers, A; Farnung, W; Kohl-Bareis, M
2016-01-01
We demonstrate an efficient algorithm for the temporal and spatial based calculation of speckle contrast for the imaging of blood flow by laser speckle contrast analysis (LASCA). It reduces the numerical complexity of necessary calculations, facilitates a multi-core and many-core implementation of the speckle analysis and enables an independence of temporal or spatial resolution and SNR. The new algorithm was evaluated for both spatial and temporal based analysis of speckle patterns with different image sizes and amounts of recruited pixels as sequential, multi-core and many-core code.
Collective Machine Learning: Team Learning and Classification in Multi-Agent Systems
ERIC Educational Resources Information Center
Gifford, Christopher M.
2009-01-01
This dissertation focuses on the collaboration of multiple heterogeneous, intelligent agents (hardware or software) which collaborate to learn a task and are capable of sharing knowledge. The concept of collaborative learning in multi-agent and multi-robot systems is largely under studied, and represents an area where further research is needed to…
Multi-mode optical fibers for connecting space-based spectrometers
NASA Astrophysics Data System (ADS)
Roberts, W. T.; Lindenmisth, C. A.; Bender, S.; Miller, E. A.; Motts, E.; Ott, M.; LaRocca, F.; Thomes, J.
2017-11-01
Laser spectral analysis systems are increasingly being considered for in situ analysis of the atomic and molecular composition of selected rock and soil samples on other planets [1][2][3]. Both Laser Induced Breakdown Spectroscopy (LIBS) and Raman spectroscopy are used to identify the constituents of soil and rock samples in situ. LIBS instruments use a high peak-power laser to ablate a minute area of the surface of a sample. The resulting plasma is observed with an optical head, which collects the emitted light for analysis by one or more spectrometers. By identifying the ion emission lines observed in the plasma, the constituent elements and their abundance can be deduced. In Raman spectroscopy, laser photons incident on the sample surface are scattered and experience a Raman shift, exchanging small amounts of energy with the molecules scattering the light. By observing the spectrum of the scattered light, it is possible to determine the molecular composition of the sample. For both types of instruments, there are advantages to physically separating the light collecting optics from the spectroscopy optics. The light collection system will often have articulating or rotating elements to facilitate the interrogation of multiple samples with minimum expenditure of energy and motion. As such, the optical head is often placed on a boom or an appendage allowing it to be pointed in different directions or easily positioned in different locations. By contrast, the spectrometry portion of the instrument is often well-served by placing it in a more static location. The detectors often operate more consistently in a thermally-controlled environment. Placing them deep within the spacecraft structure also provides some shielding from ionizing radiation, extending the instrument's useful life. Finally, the spectrometry portion of the instrument often contains significant mass, such that keeping it off of the moving portion of the platform, allowing that portion to be significantly smaller, less massive and less robust. Large core multi-mode optical fibers are often used to accommodate the optical connection of the two separated portions of such instrumentation. In some cases, significant throughput efficiency improvement can be realized by judiciously orienting the strands of multi-fiber cable, close-bunching them to accommodate a tight focus of the optical system on the optical side of the connection, and splaying them out linearly along a spectrometer slit on the other end. For such instrumentation to work effectively in identifying elements and molecules, and especially to produce accurate quantitative results, the spectral throughput of the optical fiber connection must be consistent over varying temperatures, over the range of motion of the optical head (and it's implied optical cable stresses), and over angle-aperture invariant of the total system. While the first two of these conditions have been demonstrated[4], spectral observations of the latter present a cause for concern, and may have an impact on future design of fiber-connected LIBS and Raman spectroscopy instruments. In short, we have observed that the shape of the spectral efficiency curve of a large multi-mode core optical fiber changes as a function of input angle.
Study of Solid State Drives performance in PROOF distributed analysis system
NASA Astrophysics Data System (ADS)
Panitkin, S. Y.; Ernst, M.; Petkus, R.; Rind, O.; Wenaus, T.
2010-04-01
Solid State Drives (SSD) is a promising storage technology for High Energy Physics parallel analysis farms. Its combination of low random access time and relatively high read speed is very well suited for situations where multiple jobs concurrently access data located on the same drive. It also has lower energy consumption and higher vibration tolerance than Hard Disk Drive (HDD) which makes it an attractive choice in many applications raging from personal laptops to large analysis farms. The Parallel ROOT Facility - PROOF is a distributed analysis system which allows to exploit inherent event level parallelism of high energy physics data. PROOF is especially efficient together with distributed local storage systems like Xrootd, when data are distributed over computing nodes. In such an architecture the local disk subsystem I/O performance becomes a critical factor, especially when computing nodes use multi-core CPUs. We will discuss our experience with SSDs in PROOF environment. We will compare performance of HDD with SSD in I/O intensive analysis scenarios. In particular we will discuss PROOF system performance scaling with a number of simultaneously running analysis jobs.
Scalability Issues for Remote Sensing Infrastructure: A Case Study.
Liu, Yang; Picard, Sean; Williamson, Carey
2017-04-29
For the past decade, a team of University of Calgary researchers has operated a large "sensor Web" to collect, analyze, and share scientific data from remote measurement instruments across northern Canada. This sensor Web receives real-time data streams from over a thousand Internet-connected sensors, with a particular emphasis on environmental data (e.g., space weather, auroral phenomena, atmospheric imaging). Through research collaborations, we had the opportunity to evaluate the performance and scalability of their remote sensing infrastructure. This article reports the lessons learned from our study, which considered both data collection and data dissemination aspects of their system. On the data collection front, we used benchmarking techniques to identify and fix a performance bottleneck in the system's memory management for TCP data streams, while also improving system efficiency on multi-core architectures. On the data dissemination front, we used passive and active network traffic measurements to identify and reduce excessive network traffic from the Web robots and JavaScript techniques used for data sharing. While our results are from one specific sensor Web system, the lessons learned may apply to other scientific Web sites with remote sensing infrastructure.
Design of tangential multi-energy SXR cameras for tokamak plasmas
NASA Astrophysics Data System (ADS)
Yamazaki, H.; Delgado-Aparicio, L. F.; Pablant, N.; Hill, K.; Bitter, M.; Takase, Y.; Ono, M.; Stratton, B.
2017-10-01
A new synthetic diagnostic capability has been built to study the response of tangential multi-energy soft x-ray pin-hole cameras for arbitrary plasma densities (ne , D), temperature (Te) and ion concentrations (nZ). For tokamaks and future facilities to operate safely in a high-pressure long-pulse discharge, it is imperative to address key issues associated with impurity sources, core transport and high-Z impurity accumulation. Multi-energy soft xray imaging provides a unique opportunity for measuring, simultaneously, a variety of important plasma properties (e.g. Te, nZ and ΔZeff). These systems are designed to sample the continuum- and line-emission from low- to high-Z impurities (e.g. C, O, Al, Si, Ar, Ca, Fe, Ni and Mo) in multiple energy-ranges. These x-ray cameras will be installed in the MST-RFP, as well as NSTX-U and DIII-D tokamaks, measuring the radial structure of the photon emissivity with a radial resolution below 1 cm at a 500 Hz frame rate and a photon-energy resolution of 500 eV. The layout and response expected for the new systems will be shown for different plasma conditions and impurity concentrations. The effect of toroidal rotation driving poloidal asymmetries in the core radiation is also addressed for the case of NSTX-U.
Zhang, Zhen; Ma, Cheng; Zhu, Rong
2017-08-23
Artificial Neural Networks (ANNs), including Deep Neural Networks (DNNs), have become the state-of-the-art methods in machine learning and achieved amazing success in speech recognition, visual object recognition, and many other domains. There are several hardware platforms for developing accelerated implementation of ANN models. Since Field Programmable Gate Array (FPGA) architectures are flexible and can provide high performance per watt of power consumption, they have drawn a number of applications from scientists. In this paper, we propose a FPGA-based, granularity-variable neuromorphic processor (FBGVNP). The traits of FBGVNP can be summarized as granularity variability, scalability, integrated computing, and addressing ability: first, the number of neurons is variable rather than constant in one core; second, the multi-core network scale can be extended in various forms; third, the neuron addressing and computing processes are executed simultaneously. These make the processor more flexible and better suited for different applications. Moreover, a neural network-based controller is mapped to FBGVNP and applied in a multi-input, multi-output, (MIMO) real-time, temperature-sensing and control system. Experiments validate the effectiveness of the neuromorphic processor. The FBGVNP provides a new scheme for building ANNs, which is flexible, highly energy-efficient, and can be applied in many areas.
Zhang, Zhen; Zhu, Rong
2017-01-01
Artificial Neural Networks (ANNs), including Deep Neural Networks (DNNs), have become the state-of-the-art methods in machine learning and achieved amazing success in speech recognition, visual object recognition, and many other domains. There are several hardware platforms for developing accelerated implementation of ANN models. Since Field Programmable Gate Array (FPGA) architectures are flexible and can provide high performance per watt of power consumption, they have drawn a number of applications from scientists. In this paper, we propose a FPGA-based, granularity-variable neuromorphic processor (FBGVNP). The traits of FBGVNP can be summarized as granularity variability, scalability, integrated computing, and addressing ability: first, the number of neurons is variable rather than constant in one core; second, the multi-core network scale can be extended in various forms; third, the neuron addressing and computing processes are executed simultaneously. These make the processor more flexible and better suited for different applications. Moreover, a neural network-based controller is mapped to FBGVNP and applied in a multi-input, multi-output, (MIMO) real-time, temperature-sensing and control system. Experiments validate the effectiveness of the neuromorphic processor. The FBGVNP provides a new scheme for building ANNs, which is flexible, highly energy-efficient, and can be applied in many areas. PMID:28832522
Merging history of three bimodal clusters
NASA Astrophysics Data System (ADS)
Maurogordato, S.; Sauvageot, J. L.; Bourdin, H.; Cappi, A.; Benoist, C.; Ferrari, C.; Mars, G.; Houairi, K.
2011-01-01
We present a combined X-ray and optical analysis of three bimodal galaxy clusters selected as merging candidates at z ~ 0.1. These targets are part of MUSIC (MUlti-Wavelength Sample of Interacting Clusters), which is a general project designed to study the physics of merging clusters by means of multi-wavelength observations. Observations include spectro-imaging with XMM-Newton EPIC camera, multi-object spectroscopy (260 new redshifts), and wide-field imaging at the ESO 3.6 m and 2.2 m telescopes. We build a global picture of these clusters using X-ray luminosity and temperature maps together with galaxy density and velocity distributions. Idealized numerical simulations were used to constrain the merging scenario for each system. We show that A2933 is very likely an equal-mass advanced pre-merger ~200 Myr before the core collapse, while A2440 and A2384 are post-merger systems (~450 Myr and ~1.5 Gyr after core collapse, respectively). In the case of A2384, we detect a spectacular filament of galaxies and gas spreading over more than 1 h-1 Mpc, which we infer to have been stripped during the previous collision. The analysis of the MUSIC sample allows us to outline some general properties of merging clusters: a strong luminosity segregation of galaxies in recent post-mergers; the existence of preferential axes - corresponding to the merging directions - along which the BCGs and structures on various scales are aligned; the concomitance, in most major merger cases, of secondary merging or accretion events, with groups infalling onto the main cluster, and in some cases the evidence of previous merging episodes in one of the main components. These results are in good agreement with the hierarchical scenario of structure formation, in which clusters are expected to form by successive merging events, and matter is accreted along large-scale filaments. Based on data obtained with the European Southern Observatory, Chile (programs 072.A-0595, 075.A-0264, and 079.A-0425).Tables 5-7 are only available in electronic form at the CDS via anonymous ftp to cdsarc.u-strasbg.fr (130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/525/A79
Core Principles for Transforming Remediation within a Comprehensive: Student Success Strategy
ERIC Educational Resources Information Center
Achieving the Dream, 2015
2015-01-01
Colleges and postsecondary systems across the nation are demonstrating remarkable progress in phasing out standalone or multi-course remediation sequences, resulting in tens of thousands of students more quickly enrolling in and completing college-level courses. These organizations have collaborated to describe the principles they see in common…
NASA Astrophysics Data System (ADS)
Greynolds, Alan W.
2013-09-01
Results from the GelOE optical engineering software are presented for the through-focus, monochromatic coherent and polychromatic incoherent imaging of a radial "star" target for equivalent t-number circular and Gaussian pupils. The FFT-based simulations are carried out using OpenMP threading on a multi-core desktop computer, with and without the aid of a many-core NVIDIA GPU accessing its cuFFT library. It is found that a custom FFT optimized for the 12-core host has similar performance to a simply implemented 256-core GPU FFT. A more sophisticated version of the latter but tuned to reduce overhead on a 448-core GPU is 20 to 28 times faster than a basic FFT implementation running on one CPU core.
Using a Java Dynamic Tree to manage the terminology in a suite of medical applications.
Yang, K; Evens, M W; Trace, D A
2008-01-01
Now that the National Library of Medicine has made SNOMED-CT widely available, we are trying to manage the terminology of a whole suite of medical applications and map our terminology into that in SNOMED. This paper describes the design and implementation of the Java Dynamic Tree that provides structure to our medical terminology and explains how it functions as the core of our system. The tree was designed to reflect the stages in a patient interview, so it contains components for identifying the patient and the provider, a large set of chief complaints, review of systems, physical examination, several history modules, medications, laboratory tests, imaging, and special procedures. The tree is mirrored in a commercial DBMS, which also stores multi-encounter patient data, disorder patterns for our Bayesian diagnostic system, and the data and rules for other expert systems. The DBMS facilitates the import and export of large terminology files. Our Java Dynamic Tree allows the health care provider to view the entire terminology along with the structure that supports it, as well as the mechanism for the generation of progress notes and other documents, in terms of a single hierarchical structure. Changes in terminology can be propagated through the system under the control of the expert. The import/ export facility has been a major help by replacing our original terminology by the terminology in SNOMED-CT.
Mobile access to virtual randomization for investigator-initiated trials.
Deserno, Thomas M; Keszei, András P
2017-08-01
Background/aims Randomization is indispensable in clinical trials in order to provide unbiased treatment allocation and a valid statistical inference. Improper handling of allocation lists can be avoided using central systems, for example, human-based services. However, central systems are unaffordable for investigator-initiated trials and might be inaccessible from some places, where study subjects need allocations. We propose mobile access to virtual randomization, where the randomization lists are non-existent and the appropriate allocation is computed on demand. Methods The core of the system architecture is an electronic data capture system or a clinical trial management system, which is extended by an R interface connecting the R server using the Java R Interface. Mobile devices communicate via the representational state transfer web services. Furthermore, a simple web-based setup allows configuring the appropriate statistics by non-statisticians. Our comprehensive R script supports simple randomization, restricted randomization using a random allocation rule, block randomization, and stratified randomization for un-blinded, single-blinded, and double-blinded trials. For each trial, the electronic data capture system or the clinical trial management system stores the randomization parameters and the subject assignments. Results Apps are provided for iOS and Android and subjects are randomized using smartphones. After logging onto the system, the user selects the trial and the subject, and the allocation number and treatment arm are displayed instantaneously and stored in the core system. So far, 156 subjects have been allocated from mobile devices serving five investigator-initiated trials. Conclusion Transforming pre-printed allocation lists into virtual ones ensures the correct conduct of trials and guarantees a strictly sequential processing in all trial sites. Covering 88% of all randomization models that are used in recent trials, virtual randomization becomes available for investigator-initiated trials and potentially for large multi-center trials.
Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo
McDaniel, Tyler; D’Azevedo, Ed F.; Li, Ying Wai; ...
2017-11-07
Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is therefore formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with applicationmore » of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. Here this procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi- core CPUs and GPUs.« less
Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo
DOE Office of Scientific and Technical Information (OSTI.GOV)
McDaniel, Tyler; D’Azevedo, Ed F.; Li, Ying Wai
Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is therefore formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with applicationmore » of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. Here this procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi- core CPUs and GPUs.« less
Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo.
McDaniel, T; D'Azevedo, E F; Li, Y W; Wong, K; Kent, P R C
2017-11-07
Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is, therefore, formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with an application of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. This procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo, where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi-core central processing units and graphical processing units.
Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo
NASA Astrophysics Data System (ADS)
McDaniel, T.; D'Azevedo, E. F.; Li, Y. W.; Wong, K.; Kent, P. R. C.
2017-11-01
Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is, therefore, formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with an application of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. This procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo, where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi-core central processing units and graphical processing units.
NASA Astrophysics Data System (ADS)
Nishiura, Daisuke; Furuichi, Mikito; Sakaguchi, Hide
2015-09-01
The computational performance of a smoothed particle hydrodynamics (SPH) simulation is investigated for three types of current shared-memory parallel computer devices: many integrated core (MIC) processors, graphics processing units (GPUs), and multi-core CPUs. We are especially interested in efficient shared-memory allocation methods for each chipset, because the efficient data access patterns differ between compute unified device architecture (CUDA) programming for GPUs and OpenMP programming for MIC processors and multi-core CPUs. We first introduce several parallel implementation techniques for the SPH code, and then examine these on our target computer architectures to determine the most effective algorithms for each processor unit. In addition, we evaluate the effective computing performance and power efficiency of the SPH simulation on each architecture, as these are critical metrics for overall performance in a multi-device environment. In our benchmark test, the GPU is found to produce the best arithmetic performance as a standalone device unit, and gives the most efficient power consumption. The multi-core CPU obtains the most effective computing performance. The computational speed of the MIC processor on Xeon Phi approached that of two Xeon CPUs. This indicates that using MICs is an attractive choice for existing SPH codes on multi-core CPUs parallelized by OpenMP, as it gains computational acceleration without the need for significant changes to the source code.
Massive Multi-Agent Systems Control
NASA Technical Reports Server (NTRS)
Campagne, Jean-Charles; Gardon, Alain; Collomb, Etienne; Nishida, Toyoaki
2004-01-01
In order to build massive multi-agent systems, considered as complex and dynamic systems, one needs a method to analyze and control the system. We suggest an approach using morphology to represent and control the state of large organizations composed of a great number of light software agents. Morphology is understood as representing the state of the multi-agent system as shapes in an abstract geometrical space, this notion is close to the notion of phase space in physics.
NASA Astrophysics Data System (ADS)
Schmidt, K. B.; Treu, T.; Brammer, G. B.; Bradač, M.; Wang, X.; Dijkstra, M.; Dressler, A.; Fontana, A.; Gavazzi, R.; Henry, A. L.; Hoag, A.; Jones, T. A.; Kelly, P. L.; Malkan, M. A.; Mason, C.; Pentericci, L.; Poggianti, B.; Stiavelli, M.; Trenti, M.; von der Linden, A.; Vulcani, B.
2014-02-01
The Grism Lens-Amplified Survey from Space (GLASS) is a Hubble Space Telescope (HST) Large Program, which will obtain 140 orbits of grism spectroscopy of the core and infall regions of 10 galaxy clusters, selected to be among the very best cosmic telescopes. Extensive HST imaging is available from many sources including the CLASH and Frontier Fields programs. We introduce the survey by analyzing spectra of faint multiply-imaged galaxies and z >~ 6 galaxy candidates obtained from the first 7 orbits out of 14 targeting the core of the Frontier Fields cluster MACSJ0717.5+3745. Using the G102 and G141 grisms to cover the wavelength range 0.8-1.7 μm, we confirm four strongly lensed systems by detecting emission lines in each of the images. For the 9 z >~ 6 galaxy candidates clear from contamination, we do not detect any emission lines down to a 7 orbit 1σ noise level of ~5 × 10-18 erg s-1 cm-2. Taking lensing magnification into account, our flux sensitivity reaches ~0.2-5 × 10-18 erg s-1cm-2. These limits over an uninterrupted wavelength range rule out the possibility that the high-z galaxy candidates are instead strong line emitters at lower redshift. These results show that by means of careful modeling of the background—and with the assistance of lensing magnification—interesting flux limits can be reached for large numbers of objects, avoiding pre-selection and the wavelength restrictions inherent to ground-based multi-slit spectroscopy. These observations confirm the power of slitless HST spectroscopy even in fields as crowded as a cluster core.
Sparse PDF Volumes for Consistent Multi-Resolution Volume Rendering
Sicat, Ronell; Krüger, Jens; Möller, Torsten; Hadwiger, Markus
2015-01-01
This paper presents a new multi-resolution volume representation called sparse pdf volumes, which enables consistent multi-resolution volume rendering based on probability density functions (pdfs) of voxel neighborhoods. These pdfs are defined in the 4D domain jointly comprising the 3D volume and its 1D intensity range. Crucially, the computation of sparse pdf volumes exploits data coherence in 4D, resulting in a sparse representation with surprisingly low storage requirements. At run time, we dynamically apply transfer functions to the pdfs using simple and fast convolutions. Whereas standard low-pass filtering and down-sampling incur visible differences between resolution levels, the use of pdfs facilitates consistent results independent of the resolution level used. We describe the efficient out-of-core computation of large-scale sparse pdf volumes, using a novel iterative simplification procedure of a mixture of 4D Gaussians. Finally, our data structure is optimized to facilitate interactive multi-resolution volume rendering on GPUs. PMID:26146475
DOE Office of Scientific and Technical Information (OSTI.GOV)
Downar, Thomas
This report summarizes the current status of VERA-CS Verification and Validation for PWR Core Follow operation and proposes a multi-phase plan for continuing VERA-CS V&V in FY17 and FY18. The proposed plan recognizes the hierarchical nature of a multi-physics code system such as VERA-CS and the importance of first achieving an acceptable level of V&V on each of the single physics codes before focusing on the V&V of the coupled physics solution. The report summarizes the V&V of each of the single physics codes systems currently used for core follow analysis (ie MPACT, CTF, Multigroup Cross Section Generation, and BISONmore » / Fuel Temperature Tables) and proposes specific actions to achieve a uniformly acceptable level of V&V in FY17. The report also recognizes the ongoing development of other codes important for PWR Core Follow (e.g. TIAMAT, MAMBA3D) and proposes Phase II (FY18) VERA-CS V&V activities in which those codes will also reach an acceptable level of V&V. The report then summarizes the current status of VERA-CS multi-physics V&V for PWR Core Follow and the ongoing PWR Core Follow V&V activities for FY17. An automated procedure and output data format is proposed for standardizing the output for core follow calculations and automatically generating tables and figures for the VERA-CS Latex file. A set of acceptance metrics is also proposed for the evaluation and assessment of core follow results that would be used within the script to automatically flag any results which require further analysis or more detailed explanation prior to being added to the VERA-CS validation base. After the Automation Scripts have been completed and tested using BEAVRS, the VERA-CS plan proposes the Watts Bar cycle depletion cases should be performed with the new cross section library and be included in the first draft of the new VERA-CS manual for release at the end of PoR15. Also, within the constraints imposed by the proprietary nature of plant data, as many as possible of the FY17 AMA Plant Core Follow cases should also be included in the VERA-CS manual at the end of PoR15. After completion of the ongoing development of TIAMAT for fully coupled, full core calculations with VERA-CS / BISON 1.5D, and after the completion of the refactoring of MAMBA3D for CIPS analysis in FY17, selected cases from the VERA-CS validation based should be performed, beginning with the legacy cases of Watts Bar and BEAVRS in PoR16. Finally, as potential Phase III future work some additional considerations are identified for extending the VERA-CS V&V to other reactor types such as the BWR.« less
Study of multi-functional precision optical measuring system for large scale equipment
NASA Astrophysics Data System (ADS)
Jiang, Wei; Lao, Dabao; Zhou, Weihu; Zhang, Wenying; Jiang, Xingjian; Wang, Yongxi
2017-10-01
The effective application of high performance measurement technology can greatly improve the large-scale equipment manufacturing ability. Therefore, the geometric parameters measurement, such as size, attitude and position, requires the measurement system with high precision, multi-function, portability and other characteristics. However, the existing measuring instruments, such as laser tracker, total station, photogrammetry system, mostly has single function, station moving and other shortcomings. Laser tracker needs to work with cooperative target, but it can hardly meet the requirement of measurement in extreme environment. Total station is mainly used for outdoor surveying and mapping, it is hard to achieve the demand of accuracy in industrial measurement. Photogrammetry system can achieve a wide range of multi-point measurement, but the measuring range is limited and need to repeatedly move station. The paper presents a non-contact opto-electronic measuring instrument, not only it can work by scanning the measurement path but also measuring the cooperative target by tracking measurement. The system is based on some key technologies, such as absolute distance measurement, two-dimensional angle measurement, automatically target recognition and accurate aiming, precision control, assembly of complex mechanical system and multi-functional 3D visualization software. Among them, the absolute distance measurement module ensures measurement with high accuracy, and the twodimensional angle measuring module provides precision angle measurement. The system is suitable for the case of noncontact measurement of large-scale equipment, it can ensure the quality and performance of large-scale equipment throughout the process of manufacturing and improve the manufacturing ability of large-scale and high-end equipment.
Amaya, N; Irfan, M; Zervas, G; Nejabati, R; Simeonidou, D; Sakaguchi, J; Klaus, W; Puttnam, B J; Miyazawa, T; Awaji, Y; Wada, N; Henning, I
2013-04-08
We present the first elastic, space division multiplexing, and multi-granular network based on two 7-core MCF links and four programmable optical nodes able to switch traffic utilising the space, frequency and time dimensions with over 6000-fold bandwidth granularity. Results show good end-to-end performance on all channels with power penalties between 0.75 dB and 3.7 dB.
Nguyen, Dai Hai; Lee, Jung Seok; Choi, Jong Hoon; Park, Kyung Min; Lee, Yunki; Park, Ki Dong
2016-04-15
Nanoparticle-based imaging and therapy are of interest for theranostic nanomedicine. In particular, superparamagnetic iron oxide (SPIO) nanoparticles (NPs) have attracted much attention in cancer imaging, diagnostics, and treatment because of their superior imagability and biocompatibility (approved by the Food and Drug Administration). Here, we developed SPIO nanoparticles (NPs) that self-assembled into magnetic nanoclusters (SAMNs) in aqueous environments as a theranostic nano-system. To generate multi-functional SPIO NPs, we covalently conjugated β-cyclodextrin (β-CD) to SPIO NPs using metal-adhesive dopamine groups. Polyethylene glycol (PEG) and paclitaxel (PTX) were hosted in the β-CD cavity through high affinity complexation. The core-shell structure of the magnetic nanoclusters was elucidated based on the condensed SPIO core and a PEG shell using electron microscopy and the composition was analyzed by thermogravimetric analysis (TGA). Our results indicate that nanocluster size could be readily controlled by changing the SPIO/PEG ratio in the assemblies. Interestingly, we observed a significant enhancement in magnetic resonance contrast due to the large cluster size and dense iron oxide core. In addition, tethering a tumor-targeting peptide to the SAMNs enhanced their uptake into tumor cells. PTX was efficiently loaded into β-CDs and released in a controlled manner when exposed to competitive guest molecules. These results strongly indicate that the SAMNs developed in this study possess great potential for application in image-guided cancer chemotherapy. In this study, we developed multi-functional SPIO NPs that self-assembled into magnetic nanoclusters (SAMNs) in aqueous conditions as a theranostic nano-system. The beta-cyclodextrin (β-CD) was immobilized on the surfaces of SPIO NPs and RGD-conjugated polyethylene glycol (PEG) and paclitaxel (PTX) were hosted in the β-CD cavity through high affinity complexation. We found that nanocluster size could be readily controlled by varying the SPIO/PEG ratio in the assemblies, and also demonstrated significant improvement of the functional nanoparticles for theranostic systems; enhanced magnetic resonance, improved cellular uptake, and efficient PTX loading and sustained release at the desired time point. These results strongly indicate that the SAMNs developed in this study possess great potential for application in image-guided cancer chemotherapy. Copyright © 2016 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.
A Mode of Combined ERP and KMS Knowledge Management System Construction
NASA Astrophysics Data System (ADS)
Yuena, Kang; Yangeng, Wen; Qun, Zhou
The core of ERP and knowledge management is quite similar; both will send appropriate knowledge (goods, funds) to the right people (position) at the right time. It is reasonable to believe that increase the knowledge management system in ERP will help companies achieve their goals better. This paper compares the concept of logical point of hall three-dimensional structure of the knowledge management system and the ERP in methodology level. And found they are very similar in the time dimension, logic dimension and knowledge dimension. This laid the basis of methodology in the simultaneous planning, implementation and applications. And then proposed a knowledge-based ERP Multi-Agent Management System Model. Finally, the paper described the process from planning to implementation of knowledge management ERP system with multi-Agent interaction and impact from three concepts, management thinking, software and system.
Zhou, Junhe; Wu, Jianjie; Hu, Qinsong
2018-02-05
In this paper, we propose a novel tunable unitary transformer, which can achieve arbitrary discrete unitary transforms. The unitary transformer is composed of multiple sections of multi-core fibers with closely aligned coupled cores. Phase shifters are inserted before and after the sections to control the phases of the waves in the cores. A simple algorithm is proposed to find the optimal phase setup for the phase shifters to realize the desired unitary transforms. The proposed device is fiber based and is particularly suitable for the mode division multiplexing systems. A tunable mode MUX/DEMUX for a three-mode fiber is designed based on the proposed structure.
NASA Astrophysics Data System (ADS)
Passilly, Nicolas; Perrin, Stéphane; Lullin, Justine; Albero, Jorge; Bargiel, Sylwester; Froehly, Luc; Gorecki, Christophe; Krauter, Johann; Osten, Wolfgang; Wang, Wei-Shan; Wiemer, Maik
2016-04-01
Some of the critical limitations for widespread use in medical applications of optical devices, such as confocal or optical coherence tomography (OCT) systems, are related to their cost and large size. Indeed, although quite efficient systems are available on the market, e.g. in dermatology, they equip only a few hospitals and hence, are far from being used as an early detection tool, for instance in screening of patients for early detection of cancers. In this framework, the VIAMOS project aims at proposing a concept of miniaturized, batch-fabricated and lower-cost, OCT system dedicated to non-invasive skin inspection. In order to image a large skin area, the system is based on a full-field approach. Moreover, since it relies on micro-fabricated devices whose fields of view are limited, 16 small interferometers are arranged in a dense array to perform multi-channel simultaneous imaging. Gaps between each channel are then filled by scanning of the system followed by stitching. This approach allows imaging a large area without the need of large optics. It also avoids the use of very fast and often expensive laser sources, since instead of a single point detector, almost 250 thousands pixels are used simultaneously. The architecture is then based on an array of Mirau interferometers which are interesting for their vertical arrangement compatible with vertical assembly at the wafer-level. Each array is consequently a local part of a stack of seven wafers. This stack includes a glass lens doublet, an out-of-plane actuated micro-mirror for phase shifting, a spacer and a planar beam-splitter. Consequently, different materials, such as silicon and glass, are bonded together and well-aligned thanks to lithographic-based fabrication processes.
Chikkagoudar, Satish; Wang, Kai; Li, Mingyao
2011-05-26
Gene-gene interaction in genetic association studies is computationally intensive when a large number of SNPs are involved. Most of the latest Central Processing Units (CPUs) have multiple cores, whereas Graphics Processing Units (GPUs) also have hundreds of cores and have been recently used to implement faster scientific software. However, currently there are no genetic analysis software packages that allow users to fully utilize the computing power of these multi-core devices for genetic interaction analysis for binary traits. Here we present a novel software package GENIE, which utilizes the power of multiple GPU or CPU processor cores to parallelize the interaction analysis. GENIE reads an entire genetic association study dataset into memory and partitions the dataset into fragments with non-overlapping sets of SNPs. For each fragment, GENIE analyzes: 1) the interaction of SNPs within it in parallel, and 2) the interaction between the SNPs of the current fragment and other fragments in parallel. We tested GENIE on a large-scale candidate gene study on high-density lipoprotein cholesterol. Using an NVIDIA Tesla C1060 graphics card, the GPU mode of GENIE achieves a speedup of 27 times over its single-core CPU mode run. GENIE is open-source, economical, user-friendly, and scalable. Since the computing power and memory capacity of graphics cards are increasing rapidly while their cost is going down, we anticipate that GENIE will achieve greater speedups with faster GPU cards. Documentation, source code, and precompiled binaries can be downloaded from http://www.cceb.upenn.edu/~mli/software/GENIE/.
2011-01-01
Background Gene-gene interaction in genetic association studies is computationally intensive when a large number of SNPs are involved. Most of the latest Central Processing Units (CPUs) have multiple cores, whereas Graphics Processing Units (GPUs) also have hundreds of cores and have been recently used to implement faster scientific software. However, currently there are no genetic analysis software packages that allow users to fully utilize the computing power of these multi-core devices for genetic interaction analysis for binary traits. Findings Here we present a novel software package GENIE, which utilizes the power of multiple GPU or CPU processor cores to parallelize the interaction analysis. GENIE reads an entire genetic association study dataset into memory and partitions the dataset into fragments with non-overlapping sets of SNPs. For each fragment, GENIE analyzes: 1) the interaction of SNPs within it in parallel, and 2) the interaction between the SNPs of the current fragment and other fragments in parallel. We tested GENIE on a large-scale candidate gene study on high-density lipoprotein cholesterol. Using an NVIDIA Tesla C1060 graphics card, the GPU mode of GENIE achieves a speedup of 27 times over its single-core CPU mode run. Conclusions GENIE is open-source, economical, user-friendly, and scalable. Since the computing power and memory capacity of graphics cards are increasing rapidly while their cost is going down, we anticipate that GENIE will achieve greater speedups with faster GPU cards. Documentation, source code, and precompiled binaries can be downloaded from http://www.cceb.upenn.edu/~mli/software/GENIE/. PMID:21615923
NASA Astrophysics Data System (ADS)
Rahman, P. A.
2018-05-01
This scientific paper deals with the model of the knapsack optimization problem and method of its solving based on directed combinatorial search in the boolean space. The offered by the author specialized mathematical model of decomposition of the search-zone to the separate search-spheres and the algorithm of distribution of the search-spheres to the different cores of the multi-core processor are also discussed. The paper also provides an example of decomposition of the search-zone to the several search-spheres and distribution of the search-spheres to the different cores of the quad-core processor. Finally, an offered by the author formula for estimation of the theoretical maximum of the computational acceleration, which can be achieved due to the parallelization of the search-zone to the search-spheres on the unlimited number of the processor cores, is also given.
NASA Technical Reports Server (NTRS)
Groom, Nelson J.
1997-01-01
The expanded equations for torque and force on a cylindrical permanent magnet core in a large-gap magnetic suspension system are presented. The core is assumed to be uniformly magnetized, and equations are developed for two orientations of the magnetization vector. One orientation is parallel to the axis of symmetry, and the other is perpendicular to this axis. Fields and gradients produced by suspension system electromagnets are assumed to be calculated at a point in inertial space which coincides with the origin of the core axis system in its initial alignment. Fields at a given point in the core are defined by expanding the fields produced at the origin as a Taylor series. The assumption is made that the fields can be adequately defined by expansion up to second-order terms. Examination of the expanded equations for the case where the magnetization vector is perpendicular to the axis of symmetry reveals that some of the second-order gradient terms provide a method of generating torque about the axis of magnetization and therefore provide the ability to produce six-degree-of-freedom control.
Large-area super-resolution optical imaging by using core-shell microfibers
NASA Astrophysics Data System (ADS)
Liu, Cheng-Yang; Lo, Wei-Chieh
2017-09-01
We first numerically and experimentally report large-area super-resolution optical imaging achieved by using core-shell microfibers. The particular spatial electromagnetic waves for different core-shell microfibers are studied by using finite-difference time-domain and ray tracing calculations. The focusing properties of photonic nanojets are evaluated in terms of intensity profile and full width at half-maximum along propagation and transversal directions. In experiment, the general optical fiber is chemically etched down to 6 μm diameter and coated with different metallic thin films by using glancing angle deposition. The direct imaging of photonic nanojets for different core-shell microfibers is performed with a scanning optical microscope system. We show that the intensity distribution of a photonic nanojet is highly related to the metallic shell due to the surface plasmon polaritons. Furthermore, large-area super-resolution optical imaging is performed by using different core-shell microfibers placed over the nano-scale grating with 150 nm line width. The core-shell microfiber-assisted imaging is achieved with super-resolution and hundreds of times the field-of-view in contrast to microspheres. The possible applications of these core-shell optical microfibers include real-time large-area micro-fluidics and nano-structure inspections.
2010-07-03
Benny Hopson from the Barrow (Alaska) Arctic Science Consortium drills a core sample from sea ice in the Chukchi Sea on July 4, 2010. The core is sliced up into puck-sized sections and stored onboard the U.S. Coast Guard Healy for analysis in the ship's lab. Impacts of Climate change on the Eco-Systems and Chemistry of the Arctic Pacific Environment (ICESCAPE) is a multi-year NASA shipborne project. The bulk of the research will take place in the Beaufort and Chukchi Sea’s in summer of 2010 and fall of 2011. Photo Credit: (NASA/Kathryn Hansen)
Hybrid Computational Architecture for Multi-Scale Modeling of Materials and Devices
2016-01-03
Equivalent: Total Number: Sub Contractors (DD882) Names of Faculty Supported Names of Under Graduate students supported Names of Personnel receiving masters...GHz, 20 cores (40 with hyper-threading ( HT )) Single node performance Node # of cores Total CPU time User CPU time System CPU time Elapsed time...INTEL20 40 (with HT ) 534.785 529.984 4.800 541.179 20 468.873 466.119 2.754 476.878 10 671.798 669.653 2.145 680.510 8 772.269 770.256 2.013
GeauxDock: Accelerating Structure-Based Virtual Screening with Heterogeneous Computing
Fang, Ye; Ding, Yun; Feinstein, Wei P.; Koppelman, David M.; Moreno, Juana; Jarrell, Mark; Ramanujam, J.; Brylinski, Michal
2016-01-01
Computational modeling of drug binding to proteins is an integral component of direct drug design. Particularly, structure-based virtual screening is often used to perform large-scale modeling of putative associations between small organic molecules and their pharmacologically relevant protein targets. Because of a large number of drug candidates to be evaluated, an accurate and fast docking engine is a critical element of virtual screening. Consequently, highly optimized docking codes are of paramount importance for the effectiveness of virtual screening methods. In this communication, we describe the implementation, tuning and performance characteristics of GeauxDock, a recently developed molecular docking program. GeauxDock is built upon the Monte Carlo algorithm and features a novel scoring function combining physics-based energy terms with statistical and knowledge-based potentials. Developed specifically for heterogeneous computing platforms, the current version of GeauxDock can be deployed on modern, multi-core Central Processing Units (CPUs) as well as massively parallel accelerators, Intel Xeon Phi and NVIDIA Graphics Processing Unit (GPU). First, we carried out a thorough performance tuning of the high-level framework and the docking kernel to produce a fast serial code, which was then ported to shared-memory multi-core CPUs yielding a near-ideal scaling. Further, using Xeon Phi gives 1.9× performance improvement over a dual 10-core Xeon CPU, whereas the best GPU accelerator, GeForce GTX 980, achieves a speedup as high as 3.5×. On that account, GeauxDock can take advantage of modern heterogeneous architectures to considerably accelerate structure-based virtual screening applications. GeauxDock is open-sourced and publicly available at www.brylinski.org/geauxdock and https://figshare.com/articles/geauxdock_tar_gz/3205249. PMID:27420300
GeauxDock: Accelerating Structure-Based Virtual Screening with Heterogeneous Computing.
Fang, Ye; Ding, Yun; Feinstein, Wei P; Koppelman, David M; Moreno, Juana; Jarrell, Mark; Ramanujam, J; Brylinski, Michal
2016-01-01
Computational modeling of drug binding to proteins is an integral component of direct drug design. Particularly, structure-based virtual screening is often used to perform large-scale modeling of putative associations between small organic molecules and their pharmacologically relevant protein targets. Because of a large number of drug candidates to be evaluated, an accurate and fast docking engine is a critical element of virtual screening. Consequently, highly optimized docking codes are of paramount importance for the effectiveness of virtual screening methods. In this communication, we describe the implementation, tuning and performance characteristics of GeauxDock, a recently developed molecular docking program. GeauxDock is built upon the Monte Carlo algorithm and features a novel scoring function combining physics-based energy terms with statistical and knowledge-based potentials. Developed specifically for heterogeneous computing platforms, the current version of GeauxDock can be deployed on modern, multi-core Central Processing Units (CPUs) as well as massively parallel accelerators, Intel Xeon Phi and NVIDIA Graphics Processing Unit (GPU). First, we carried out a thorough performance tuning of the high-level framework and the docking kernel to produce a fast serial code, which was then ported to shared-memory multi-core CPUs yielding a near-ideal scaling. Further, using Xeon Phi gives 1.9× performance improvement over a dual 10-core Xeon CPU, whereas the best GPU accelerator, GeForce GTX 980, achieves a speedup as high as 3.5×. On that account, GeauxDock can take advantage of modern heterogeneous architectures to considerably accelerate structure-based virtual screening applications. GeauxDock is open-sourced and publicly available at www.brylinski.org/geauxdock and https://figshare.com/articles/geauxdock_tar_gz/3205249.
C 3, A Command-line Catalog Cross-match Tool for Large Astrophysical Catalogs
NASA Astrophysics Data System (ADS)
Riccio, Giuseppe; Brescia, Massimo; Cavuoti, Stefano; Mercurio, Amata; di Giorgio, Anna Maria; Molinari, Sergio
2017-02-01
Modern Astrophysics is based on multi-wavelength data organized into large and heterogeneous catalogs. Hence, the need for efficient, reliable and scalable catalog cross-matching methods plays a crucial role in the era of the petabyte scale. Furthermore, multi-band data have often very different angular resolution, requiring the highest generality of cross-matching features, mainly in terms of region shape and resolution. In this work we present C 3 (Command-line Catalog Cross-match), a multi-platform application designed to efficiently cross-match massive catalogs. It is based on a multi-core parallel processing paradigm and conceived to be executed as a stand-alone command-line process or integrated within any generic data reduction/analysis pipeline, providing the maximum flexibility to the end-user, in terms of portability, parameter configuration, catalog formats, angular resolution, region shapes, coordinate units and cross-matching types. Using real data, extracted from public surveys, we discuss the cross-matching capabilities and computing time efficiency also through a direct comparison with some publicly available tools, chosen among the most used within the community, and representative of different interface paradigms. We verified that the C 3 tool has excellent capabilities to perform an efficient and reliable cross-matching between large data sets. Although the elliptical cross-match and the parametric handling of angular orientation and offset are known concepts in the astrophysical context, their availability in the presented command-line tool makes C 3 competitive in the context of public astronomical tools.
Rare-earth-doped optical-fiber core deposition using full vapor-phase SPCVD process
NASA Astrophysics Data System (ADS)
Barnini, A.; Robin, T.; Cadier, B.; Aka, G.; Caurant, D.; Gotter, T.; Guyon, C.; Pinsard, E.; Guitton, P.; Laurent, A.; Montron, R.
2017-02-01
One key parameter in the race toward ever-higher power fiber lasers remains the rare earth doped optical core quality. Modern Large Mode Area (LMA) fibers require a fine radial control of the core refractive index (RI) close to the silica level. These low RI are achieved with multi-component materials that cannot be readily obtained using conventional solution doping based Modified Chemical Vapor Deposition (MCVD) technology. This paper presents a study of such optical material obtained through a full-vapor phase Surface Plasma Chemical Vapor Deposition (SPCVD). The SPCVD process generates straight glassy films on the inner surface of a thermally regulated synthetic silica tube under vacuum. The first part of the presented results points out the feasibility of ytterbium-doped aluminosilicate fibers by this process. In the second part we describe the challenge controlling the refractive index throughout the core diameter when using volatile fluorine to create efficient LMA fiber profiles. It has been demonstrated that it is possible to counter-act the loss of fluorine at the center of the core by adjusting the core composition locally. Our materials yielded, when used in optical fibers with numerical apertures ranging from 0.07 to 0.09, power conversion efficiency up to 76% and low background losses below 20 dB/km at 1100nm. Photodarkening has been measured to be similar to equivalent MCVD based fibers. The use of cerium as a co-dopant allowed for a complete mitigation of this laser lifetime detrimental effect. The SPCVD process enables high capacity preforms and is particularly versatile when it comes to radial tailoring of both rare earth doping level and RI. Large core diameter preforms - up to 4mm - were successfully produced.
Charlton, Bruce G
2006-01-01
It is plausible to assume that in the future science will form the compulsory core element both of school curricula and multi-disciplinary undergraduate degrees. But for this to happen entails a shift in the emphasis and methods of science teaching, away from the traditional concern with educating specialists and professionals. Traditional science teaching was essentially vocational, designed to provide precise and comprehensive scientific knowledge for practical application. By contrast, future science teaching will be a general education, hence primarily conceptual. Its aim should be to provide an education in flexible rationality. Vocational science teaching was focused on a single-discipline undergraduate degree, but a general education in abstract systematic thinking is best inculcated by studying several scientific disciplines. In this sense, 'science' is understood as mathematics and the natural sciences, but also the abstract and systematic aspects of disciplines such as economics, linguistics, music theory, history, sociology, political science and management science. Such a wide variety of science options in a multi-disciplinary degree will increase the possibility of student motivation and aptitude. Specialist vocational science education will progressively be shifted to post-graduate level, in Masters and Doctoral programs. A multi-disciplinary and conceptually-based science core curriculum should provide an appropriate preparation for dealing with the demands of modern societies; their complex and rapidly changing social systems; and the need for individual social and professional mobility. Training in rational conceptual thinking also has potential benefits to human health and happiness, since it allows people to over-ride inappropriate instincts, integrate conflicting desires and pursue long-term goals.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sreepathi, Sarat; Sripathi, Vamsi; Mills, Richard T
2013-01-01
Inefficient parallel I/O is known to be a major bottleneck among scientific applications employed on supercomputers as the number of processor cores grows into the thousands. Our prior experience indicated that parallel I/O libraries such as HDF5 that rely on MPI-IO do not scale well beyond 10K processor cores, especially on parallel file systems (like Lustre) with single point of resource contention. Our previous optimization efforts for a massively parallel multi-phase and multi-component subsurface simulator (PFLOTRAN) led to a two-phase I/O approach at the application level where a set of designated processes participate in the I/O process by splitting themore » I/O operation into a communication phase and a disk I/O phase. The designated I/O processes are created by splitting the MPI global communicator into multiple sub-communicators. The root process in each sub-communicator is responsible for performing the I/O operations for the entire group and then distributing the data to rest of the group. This approach resulted in over 25X speedup in HDF I/O read performance and 3X speedup in write performance for PFLOTRAN at over 100K processor cores on the ORNL Jaguar supercomputer. This research describes the design and development of a general purpose parallel I/O library, SCORPIO (SCalable block-ORiented Parallel I/O) that incorporates our optimized two-phase I/O approach. The library provides a simplified higher level abstraction to the user, sitting atop existing parallel I/O libraries (such as HDF5) and implements optimized I/O access patterns that can scale on larger number of processors. Performance results with standard benchmark problems and PFLOTRAN indicate that our library is able to maintain the same speedups as before with the added flexibility of being applicable to a wider range of I/O intensive applications.« less
NASA Astrophysics Data System (ADS)
Zhang, Ziyang; Fiebrandt, Julia; Haynes, Dionne; Sun, Kai; Madhav, Kalaga; Stoll, Andreas; Makan, Kirill; Makan, Vadim; Roth, Martin
2018-03-01
Three-dimensional multi-mode interference devices are demonstrated using a single-mode fiber (SMF) center-spliced to a section of polygon-shaped core multimode fiber (MMF). This simple structure can effectively generate well-localized self-focusing spots that match to the layout of a chosen multi-core fiber (MCF) as a launcher device. An optimized hexagon-core MMF can provide efficient coupling from a SMF to a 7-core MCF with an insertion loss of 0.6 dB and a power imbalance of 0.5 dB, while a square-core MMF can form a self-imaging pattern with symmetrically distributed 2 × 2, 3 × 3 or 4 × 4 spots. These spots can be directly received by a two-dimensional detector array. The device can work as a vector curvature sensor by comparing the relative power among the spots with a resolution of ∼0.1° over a 1.8 mm-long MMF.
Elzoghby, Ahmed O; Mostafa, Shaimaa K; Helmy, Maged W; ElDemellawy, Maha A; Sheweita, Salah A
2017-09-01
In the current work, we propose a combined delivery nanoplatform for letrozole (LTZ) and celecoxib (CXB). Multi-reservoir nanocarriers were developed by enveloping protamine nanocapsules (PRM-NCs) within drug-phospholipid complex bilayer. Encapsulation of NCs within phospholipid bilayer was confirmed by both size increase from 109.7 to 179.8 nm and reduction of surface charge from +19.0 to +7.78 mV. The multi-compartmental core-shell structure enabled biphasic CXB release with initial fast release induced by complexation with phospholipid shell followed by prolonged release from oily core. Moreover, phospholipid coating provided protection for cationic PRM-NCs against interaction with RBCs and serum proteins enabling their systemic administration. Pharmacokinetic analysis demonstrated prolonged circulation and delayed clearance of both drugs after intravenous administration into rats. The superior anti-tumor efficacy of multi-reservoir NCs was manifested as powerful cytotoxicity against MCF-7 breast cancer cells and marked reduction in the mammary tumor volume in Ehrlich ascites bearing mice compared with free LTZ-CXB combination. Moreover, the NCs induced apoptotic caspase activation and marked inhibition of aromatase expression and angiogenic marker, VEGF as well as inhibition of both NFκB and TNFα. Multi-reservoir phospholipid shell coating PRM-NCs could serve as a promising nanocarrier for parenteral combined delivery of LTZ and CXB.
Evolution of the ATLAS Software Framework towards Concurrency
NASA Astrophysics Data System (ADS)
Jones, R. W. L.; Stewart, G. A.; Leggett, C.; Wynne, B. M.
2015-05-01
The ATLAS experiment has successfully used its Gaudi/Athena software framework for data taking and analysis during the first LHC run, with billions of events successfully processed. However, the design of Gaudi/Athena dates from early 2000 and the software and the physics code has been written using a single threaded, serial design. This programming model has increasing difficulty in exploiting the potential of current CPUs, which offer their best performance only through taking full advantage of multiple cores and wide vector registers. Future CPU evolution will intensify this trend, with core counts increasing and memory per core falling. Maximising performance per watt will be a key metric, so all of these cores must be used as efficiently as possible. In order to address the deficiencies of the current framework, ATLAS has embarked upon two projects: first, a practical demonstration of the use of multi-threading in our reconstruction software, using the GaudiHive framework; second, an exercise to gather requirements for an updated framework, going back to the first principles of how event processing occurs. In this paper we report on both these aspects of our work. For the hive based demonstrators, we discuss what changes were necessary in order to allow the serially designed ATLAS code to run, both to the framework and to the tools and algorithms used. We report on what general lessons were learned about the code patterns that had been employed in the software and which patterns were identified as particularly problematic for multi-threading. These lessons were fed into our considerations of a new framework and we present preliminary conclusions on this work. In particular we identify areas where the framework can be simplified in order to aid the implementation of a concurrent event processing scheme. Finally, we discuss the practical difficulties involved in migrating a large established code base to a multi-threaded framework and how this can be achieved for LHC Run 3.
Nuclear Engine System Simulation (NESS). Volume 1: Program user's guide
NASA Astrophysics Data System (ADS)
Pelaccio, Dennis G.; Scheil, Christine M.; Petrosky, Lyman J.
1993-03-01
A Nuclear Thermal Propulsion (NTP) engine system design analysis tool is required to support current and future Space Exploration Initiative (SEI) propulsion and vehicle design studies. Currently available NTP engine design models are those developed during the NERVA program in the 1960's and early 1970's and are highly unique to that design or are modifications of current liquid propulsion system design models. To date, NTP engine-based liquid design models lack integrated design of key NTP engine design features in the areas of reactor, shielding, multi-propellant capability, and multi-redundant pump feed fuel systems. Additionally, since the SEI effort is in the initial development stage, a robust, verified NTP analysis design tool could be of great use to the community. This effort developed an NTP engine system design analysis program (tool), known as the Nuclear Engine System Simulation (NESS) program, to support ongoing and future engine system and stage design study efforts. In this effort, Science Applications International Corporation's (SAIC) NTP version of the Expanded Liquid Engine Simulation (ELES) program was modified extensively to include Westinghouse Electric Corporation's near-term solid-core reactor design model. The ELES program has extensive capability to conduct preliminary system design analysis of liquid rocket systems and vehicles. The program is modular in nature and is versatile in terms of modeling state-of-the-art component and system options as discussed. The Westinghouse reactor design model, which was integrated in the NESS program, is based on the near-term solid-core ENABLER NTP reactor design concept. This program is now capable of accurately modeling (characterizing) a complete near-term solid-core NTP engine system in great detail, for a number of design options, in an efficient manner. The following discussion summarizes the overall analysis methodology, key assumptions, and capabilities associated with the NESS presents an example problem, and compares the results to related NTP engine system designs. Initial installation instructions and program disks are in Volume 2 of the NESS Program User's Guide.
Nuclear Engine System Simulation (NESS). Volume 1: Program user's guide
NASA Technical Reports Server (NTRS)
Pelaccio, Dennis G.; Scheil, Christine M.; Petrosky, Lyman J.
1993-01-01
A Nuclear Thermal Propulsion (NTP) engine system design analysis tool is required to support current and future Space Exploration Initiative (SEI) propulsion and vehicle design studies. Currently available NTP engine design models are those developed during the NERVA program in the 1960's and early 1970's and are highly unique to that design or are modifications of current liquid propulsion system design models. To date, NTP engine-based liquid design models lack integrated design of key NTP engine design features in the areas of reactor, shielding, multi-propellant capability, and multi-redundant pump feed fuel systems. Additionally, since the SEI effort is in the initial development stage, a robust, verified NTP analysis design tool could be of great use to the community. This effort developed an NTP engine system design analysis program (tool), known as the Nuclear Engine System Simulation (NESS) program, to support ongoing and future engine system and stage design study efforts. In this effort, Science Applications International Corporation's (SAIC) NTP version of the Expanded Liquid Engine Simulation (ELES) program was modified extensively to include Westinghouse Electric Corporation's near-term solid-core reactor design model. The ELES program has extensive capability to conduct preliminary system design analysis of liquid rocket systems and vehicles. The program is modular in nature and is versatile in terms of modeling state-of-the-art component and system options as discussed. The Westinghouse reactor design model, which was integrated in the NESS program, is based on the near-term solid-core ENABLER NTP reactor design concept. This program is now capable of accurately modeling (characterizing) a complete near-term solid-core NTP engine system in great detail, for a number of design options, in an efficient manner. The following discussion summarizes the overall analysis methodology, key assumptions, and capabilities associated with the NESS presents an example problem, and compares the results to related NTP engine system designs. Initial installation instructions and program disks are in Volume 2 of the NESS Program User's Guide.
NASA Astrophysics Data System (ADS)
Buono, A. S.; Dasgupta, R.; Walker, D.
2011-12-01
Secular cooling of terrestrial planets is known to cause crystallization of a solid inner core from metallic liquid core. Fractionation of light and siderophile elements is important during such crystallization for evolution of outer core and possible core-mantle interaction. Thus far studies focused on a pure Fe inner core in simple binary systems but the effects of possible formation of a carbide inner core component on siderophile element partitioning in a multi-component system has yet to be looked at in detail. We investigated the effects of pressure and S content on partition coefficients (D) between cohenite and liquid in the Fe-Ni-S-C system. Multi-anvil experiments were performed at 3 and 6 GPa at 1150 °C, in an Fe-rich mix containing a constant C and Ni to which S contents of 0, 5, and 14 wt.% were added. All the mixes were doped with W, Re, Os, Pt, and Co. Samples were imaged and analyzed for Fe, Ni, S, and C using an EPMA. Fe, Ni, and trace elements were analyzed using a LA-ICP-MS. All the experiments produced cohenite and Fe-Ni-C±S liquid. Compared to solid-Fe/melt Ds [1-2], cohenite/melt Ds are lower for all elements except W. The light element (S+C) content of the liquid is the dominant controlling factor in siderophile element partitioning between cohenite and liquid as it is between crystalline Fe and liquid. In the cohenite-metallic melt experiments, D Ni decreases as S+C increases. Ni is excluded from the crystallizing solid if the solid is cohenite. We also find that in the Fe-Ni-S-C system, cohenite is stabilized to higher P than in the Fe-S-C system [3-5]. Similar to the Fe-metallic liquid systems the non-metal avoidance model [6] is applicable to the Fe3C-metallic liquid system studied here. Our study has implications for both the cores of smaller planets and the mantles of larger planets. If inner core forms a cohenite layer we would predict that depletions in the outer core will be less than they might be for Fe metal crystallization. For the mantle of the earth, which is thought to become Fe-Ni metal-saturated as shallow as 250 km, the sub-system Fe-Ni + C + S becomes relevant and Fe-Ni carbide rather than metallic Fe-Ni alloy may become the crystalline phase of interest. Our study implies that because the partition coefficients between cohenite and Fe-C-S melts are significantly lower than those between Fe-metal and S-rich liquid, in the presence of cohenite and Fe-C-S melt in the mantle, the mantle budget of Ni, Co, and Pt may be dominated by Fe-C-S liquid. W, Re, and Os will also be slightly enriched in C-rich Fe-Ni liquid over cohenite if the metal sub-system of interest is S-free. [1] Chabot et al., GCA 70, 1322-1335, 2006 [2] Chabot et al., GCA 72, 4146-4158, 2008 [3] Chabot et al., Meteorit. Planet. Sci. 42, 1735-1750, 2007 [4] Stewart et al., EPSL 284, 302-309, 2009 [5] Van Orman et al., EPSL 274, 250-257, 2008 [6] Jones, J.H., Malvin, D.J., Metall Mater Trans B 21, 697-706, 1990
Yoo, Sun K; Kim, Dong Keun; Kim, Jung C; Park, Youn Jung; Chang, Byung Chul
2008-01-01
With the increase in demand for high quality medical services, the need for an innovative hospital information system has become essential. An improved system has been implemented in all hospital units of the Yonsei University Health System. Interoperability between multi-units required appropriate hardware infrastructure and software architecture. This large-scale hospital information system encompassed PACS (Picture Archiving and Communications Systems), EMR (Electronic Medical Records) and ERP (Enterprise Resource Planning). It involved two tertiary hospitals and 50 community hospitals. The monthly data production rate by the integrated hospital information system is about 1.8 TByte and the total quantity of data produced so far is about 60 TByte. Large scale information exchange and sharing will be particularly useful for telemedicine applications.
Flow characteristics of Korea multi-purpose research reactor
DOE Office of Scientific and Technical Information (OSTI.GOV)
Heonil Kim; Hee Taek Chae; Byung Jin Jun
1995-09-01
The construction of Korea Multi-purpose Research Reactor (KMRR), a 30 MW{sub th} open-tank-in-pool type, is completed. Various thermal-hydraulic experiments have been conducted to verify the design characteristics of the KMRR. This paper describes the commissioning experiments to determine the flow distribution of KMRR core and the flow characteristics inside the chimney which stands on top of the core. The core flow is distributed to within {+-}6% of the average values, which is sufficiently flat in the sense that the design velocity in the fueled region is satisfied. The role of core bypass flow to confine the activated core coolant inmore » the chimney structure is confirmed.« less
NASA Technical Reports Server (NTRS)
Goodyer, M. J.; Britcher, C. P.
1983-01-01
The results of experimental demonstrations of a superconducting solenoid model core in the Southampton University Magnetic Suspension and Balance System are detailed. Technology and techniques relevant to large-scale wind tunnel MSBSs comprise the long term goals. The magnetic moment of solenoids, difficulties peculiar to superconducting solenoid cores, lift force and pitching moment, dynamic lift calibration, and helium boil-off measurements are discussed.
NASA Astrophysics Data System (ADS)
Wingo, S. M.; Petersen, W. A.; Gatlin, P. N.; Marks, D. A.; Wolff, D. B.; Pabla, C. S.
2017-12-01
The versatile SIMBA (System for Integrating Multi-platform data to Build the Atmospheric column) precipitation data-fusion framework produces an atmospheric column data product with multi-platform observations set into a common 3-D grid, affording an efficient starting point for multi-sensor comparisons and analysis that can be applied to any region. Supported data sources include: ground-based scanning and profiling radars (S-, X-, Ku-, K-, and Ka-band), multiple types of disdrometers and rain gauges, the GPM Core Observatory's Microwave Imager (GMI, 10-183 GHz) and Dual-frequency Precipitation Radar (DPR, Ka/Ku-band), as well as thermodynamic soundings and the Multi-Radar/Multi-Sensor QPE product. SIMBA column data files provide a unique way to evaluate the complete vertical profile of precipitation. Two post-launch (GPM Core in orbit) field campaigns focused on different facets of the GPM mission: the Olympic Mountains Experiment (OLYMPEX) was geared toward winter season (November-February) precipitation in Pacific frontal systems and their transition from the coastal to mountainous terrain of northwest Washington, while the Integrated Precipitation and Hydrology Experiment (IPHEx) sampled warm season (April-June) precipitation and supported hydrologic applications in the southern Appalachians and eastern North Carolina. Both campaigns included multiple orographic precipitation enhancement episodes. SIMBA column products generated for select OLYMPEX and IPHEx events will be used to evaluate spatial variability and vertical profiles of precipitation and drop size distribution parameters derived and/or observed by space- and ground-based sensors. Results will provide a cursory view of how well the space-based measurements represent what is observed from the ground below and an indication to how the terrain in both regions impacts the characteristics of precipitation within the column and reaching the ground.
NASA Astrophysics Data System (ADS)
Wingo, S. M.; Petersen, W. A.; Gatlin, P. N.; Marks, D. A.; Wolff, D. B.; Pabla, C. S.
2016-12-01
The versatile SIMBA (System for Integrating Multi-platform data to Build the Atmospheric column) precipitation data-fusion framework produces an atmospheric column data product with multi-platform observations set into a common 3-D grid, affording an efficient starting point for multi-sensor comparisons and analysis that can be applied to any region. Supported data sources include: ground-based scanning and profiling radars (S-, X-, Ku-, K-, and Ka-band), multiple types of disdrometers and rain gauges, the GPM Core Observatory's Microwave Imager (GMI, 10-183 GHz) and Dual-frequency Precipitation Radar (DPR, Ka/Ku-band), as well as thermodynamic soundings and the Multi-Radar/Multi-Sensor QPE product. SIMBA column data files provide a unique way to evaluate the complete vertical profile of precipitation. Two post-launch (GPM Core in orbit) field campaigns focused on different facets of the GPM mission: the Olympic Mountains Experiment (OLYMPEX) was geared toward winter season (November-February) precipitation in Pacific frontal systems and their transition from the coastal to mountainous terrain of northwest Washington, while the Integrated Precipitation and Hydrology Experiment (IPHEx) sampled warm season (April-June) precipitation and supported hydrologic applications in the southern Appalachians and eastern North Carolina. Both campaigns included multiple orographic precipitation enhancement episodes. SIMBA column products generated for select OLYMPEX and IPHEx events will be used to evaluate spatial variability and vertical profiles of precipitation and drop size distribution parameters derived and/or observed by space- and ground-based sensors. Results will provide a cursory view of how well the space-based measurements represent what is observed from the ground below and an indication to how the terrain in both regions impacts the characteristics of precipitation within the column and reaching the ground.
Prospect of Continuous VLBI Measurement of Earth Rotation in Monitoring Geophysical Fluids
NASA Technical Reports Server (NTRS)
Chao, Benjamin F.; Ma, Chopo; Clark, Thomas
1998-01-01
Large-scale mass transports in the geophysical fluids of the Earth system excite Earth's rotational variations in both length-of-day and polar motion. The excitation process is via the conservation of angular momentum. Therefore Earth rotation observations contain information about the integrated angular momentum (consisting of both the mass term and the motion term) of the geophysical fluids, which include atmosphere, hydrosphere, mantle, and the outer and inner cores. Such global information is often important and otherwise unattainable depending on the nature of the mass transport, its magnitude and time scale. The last few years have seen great advances in VLBI measurement of Earth rotation in precision and temporal resolution. These advances have opened new. areas in geophysical fluid studies, such as oceanic tidal angular momentum, atmospheric tides, Earth librations, and rapid atmospheric angular momentum fluctuations. Precision of 10 microseconds in UTI and 200 microarcseconds in polar motion can now be achieved on hourly basis. Building upon this heritage, the multi-network geodetic VLBI project, Continuous Observation of the Rotation of the Earth (CORE), promises to further these studies and to make possible studies on elusive but tell-tale geophysical processes such as oscillatory modes in the core and in the atmosphere. Currently the early phase of CORE is underway. Within a few years into the new mellinnium, the upcoming space gravity missions (such as GRACE) will measure the temporal variations in Earth's gravitational field, thus providing complementary information to that from Earth rotation study for a better understanding of global geophysical fluid processes.
Adaptive MCMC in Bayesian phylogenetics: an application to analyzing partitioned data in BEAST.
Baele, Guy; Lemey, Philippe; Rambaut, Andrew; Suchard, Marc A
2017-06-15
Advances in sequencing technology continue to deliver increasingly large molecular sequence datasets that are often heavily partitioned in order to accurately model the underlying evolutionary processes. In phylogenetic analyses, partitioning strategies involve estimating conditionally independent models of molecular evolution for different genes and different positions within those genes, requiring a large number of evolutionary parameters that have to be estimated, leading to an increased computational burden for such analyses. The past two decades have also seen the rise of multi-core processors, both in the central processing unit (CPU) and Graphics processing unit processor markets, enabling massively parallel computations that are not yet fully exploited by many software packages for multipartite analyses. We here propose a Markov chain Monte Carlo (MCMC) approach using an adaptive multivariate transition kernel to estimate in parallel a large number of parameters, split across partitioned data, by exploiting multi-core processing. Across several real-world examples, we demonstrate that our approach enables the estimation of these multipartite parameters more efficiently than standard approaches that typically use a mixture of univariate transition kernels. In one case, when estimating the relative rate parameter of the non-coding partition in a heterochronous dataset, MCMC integration efficiency improves by > 14-fold. Our implementation is part of the BEAST code base, a widely used open source software package to perform Bayesian phylogenetic inference. guy.baele@kuleuven.be. Supplementary data are available at Bioinformatics online. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com
Astakhov, Vadim
2009-01-01
Interest in simulation of large-scale metabolic networks, species development, and genesis of various diseases requires new simulation techniques to accommodate the high complexity of realistic biological networks. Information geometry and topological formalisms are proposed to analyze information processes. We analyze the complexity of large-scale biological networks as well as transition of the system functionality due to modification in the system architecture, system environment, and system components. The dynamic core model is developed. The term dynamic core is used to define a set of causally related network functions. Delocalization of dynamic core model provides a mathematical formalism to analyze migration of specific functions in biosystems which undergo structure transition induced by the environment. The term delocalization is used to describe these processes of migration. We constructed a holographic model with self-poetic dynamic cores which preserves functional properties under those transitions. Topological constraints such as Ricci flow and Pfaff dimension were found for statistical manifolds which represent biological networks. These constraints can provide insight on processes of degeneration and recovery which take place in large-scale networks. We would like to suggest that therapies which are able to effectively implement estimated constraints, will successfully adjust biological systems and recover altered functionality. Also, we mathematically formulate the hypothesis that there is a direct consistency between biological and chemical evolution. Any set of causal relations within a biological network has its dual reimplementation in the chemistry of the system environment.
Application of Advanced Multi-Core Processor Technologies to Oceanographic Research
2014-09-30
Jordan Stanway are taking on the work of analyzing their code, and we are working on the Robot Operating System (ROS) and MOOS-DB systems to evaluate...Linux/GNU operating system that should reduce the time required to build the kernel and userspace significantly. This part of the work is vital to...the platform to be used not only as a service, but also as a private deployable package. As much as possible, this system was built using operating
Multi-core fiber amplifier arrays for intra-satellite links
NASA Astrophysics Data System (ADS)
Kechagias, Marios; Crabb, Jonathan; Stampoulidis, Leontios; Farzana, Jihan; Kehayas, Efstratios; Filipowicz, Marta; Napierala, Marek; Murawski, Michal; Nasilowski, Tomasz; Barbero, Juan
2017-09-01
In this paper we present erbium doped fibre (EDF) aimed at signal amplification within satellite photonic payload systems operating in C telecommunication band. In such volume-hungry applications, the use of advanced optical transmission techniques such as space division multiplexing (SDM) can be advantageous to reduce the component and cable count.
Effectiveness of a Class-Wide Peer-Mediated Elementary Math Differentiation Strategy
ERIC Educational Resources Information Center
Lloyd, Jason D.
2017-01-01
Approximately 60% of classroom students have insufficient math skills. Within a Multi-Tiered Systems of Support (MTSS) framework, teachers can implement core differentiation strategies targeted at improving math skills of an entire class of students. Differentiation programs are developed in order to target academic skills of groups of students…
Exploration and Evaluation of Nanometer Low-power Multi-core VLSI Computer Architectures
2015-03-01
ICC, the Milkway database was created using the command: milkyway –galaxy –nogui –tcl –log memory.log one.tcl As stated previously, it is...EDA tools. Typically, Synopsys® tools use Milkway databases, whereas, Cadence Design System® use Layout Exchange Format (LEF) formats. To help
HACC: Extreme Scaling and Performance Across Diverse Architectures
NASA Astrophysics Data System (ADS)
Habib, Salman; Morozov, Vitali; Frontiere, Nicholas; Finkel, Hal; Pope, Adrian; Heitmann, Katrin
2013-11-01
Supercomputing is evolving towards hybrid and accelerator-based architectures with millions of cores. The HACC (Hardware/Hybrid Accelerated Cosmology Code) framework exploits this diverse landscape at the largest scales of problem size, obtaining high scalability and sustained performance. Developed to satisfy the science requirements of cosmological surveys, HACC melds particle and grid methods using a novel algorithmic structure that flexibly maps across architectures, including CPU/GPU, multi/many-core, and Blue Gene systems. We demonstrate the success of HACC on two very different machines, the CPU/GPU system Titan and the BG/Q systems Sequoia and Mira, attaining unprecedented levels of scalable performance. We demonstrate strong and weak scaling on Titan, obtaining up to 99.2% parallel efficiency, evolving 1.1 trillion particles. On Sequoia, we reach 13.94 PFlops (69.2% of peak) and 90% parallel efficiency on 1,572,864 cores, with 3.6 trillion particles, the largest cosmological benchmark yet performed. HACC design concepts are applicable to several other supercomputer applications.
Multi-photon excited luminescence of magnetic FePt core-shell nanoparticles.
Seemann, K M; Kuhn, B
2014-07-01
We present magnetic FePt nanoparticles with a hydrophilic, inert, and biocompatible silico-tungsten oxide shell. The particles can be functionalized, optically detected, and optically manipulated. To show the functionalization the fluorescent dye NOPS was bound to the FePt core-shell nanoparticles with propyl-triethoxy-silane linkers and fluorescence of the labeled particles were observed in ethanol (EtOH). In aqueous dispersion the NOPS fluorescence is quenched making them invisible using 1-photon excitation. However, we observe bright luminescence of labeled and even unlabeled magnetic core-shell nanoparticles with multi-photon excitation. Luminescence can be detected in the near ultraviolet and the full visible spectral range by near infrared multi-photon excitation. For optical manipulation, we were able to drag clusters of particles, and maybe also single particles, by a focused laser beam that acts as optical tweezers by inducing an electric dipole in the insulated metal nanoparticles. In a first application, we show that the luminescence of the core-shell nanoparticles is bright enough for in vivo multi-photon imaging in the mouse neocortex down to cortical layer 5.
The Emergence of Dominant Design(s) in Large Scale Cyber-Infrastructure Systems
ERIC Educational Resources Information Center
Diamanti, Eirini Ilana
2012-01-01
Cyber-infrastructure systems are integrated large-scale IT systems designed with the goal of transforming scientific practice by enabling multi-disciplinary, cross-institutional collaboration. Their large scale and socio-technical complexity make design decisions for their underlying architecture practically irreversible. Drawing on three…
Large-scale quantum transport calculations for electronic devices with over ten thousand atoms
NASA Astrophysics Data System (ADS)
Lu, Wenchang; Lu, Yan; Xiao, Zhongcan; Hodak, Miro; Briggs, Emil; Bernholc, Jerry
The non-equilibrium Green's function method (NEGF) has been implemented in our massively parallel DFT software, the real space multigrid (RMG) code suite. Our implementation employs multi-level parallelization strategies and fully utilizes both multi-core CPUs and GPU accelerators. Since the cost of the calculations increases dramatically with the number of orbitals, an optimal basis set is crucial for including a large number of atoms in the ``active device'' part of the simulations. In our implementation, the localized orbitals are separately optimized for each principal layer of the device region, in order to obtain an accurate and optimal basis set. As a large example, we calculated the transmission characteristics of a Si nanowire p-n junction. The nanowire is along (110) direction in order to minimize the number dangling bonds that are saturated by H atoms. Its diameter is 3 nm. The length of 24 nm is necessary because of the long-range screening length in Si. Our calculations clearly show the I-V characteristics of a diode, i.e., the current increases exponentially with forward bias and is near zero with backward bias. Other examples will also be presented, including three-terminal transistors and large sensor structures.
MAD Adaptive Optics Imaging of High-luminosity Quasars: A Pilot Project
NASA Astrophysics Data System (ADS)
Liuzzo, E.; Falomo, R.; Paiano, S.; Treves, A.; Uslenghi, M.; Arcidiacono, C.; Baruffolo, A.; Diolaiti, E.; Farinato, J.; Lombini, M.; Moretti, A.; Ragazzoni, R.; Brast, R.; Donaldson, R.; Kolb, J.; Marchetti, E.; Tordo, S.
2016-08-01
We present near-IR images of five luminous quasars at z ˜ 2 and one at z ˜ 4 obtained with an experimental adaptive optics (AO) instrument at the European Southern Observatory Very Large Telescope. The observations are part of a program aimed at demonstrating the capabilities of multi-conjugated adaptive optics imaging combined with the use of natural guide stars for high spatial resolution studies on large telescopes. The observations were mostly obtained under poor seeing conditions but in two cases. In spite of these nonoptimal conditions, the resulting images of point sources have cores of FWHM ˜ 0.2 arcsec. We are able to characterize the host galaxy properties for two sources and set stringent upper limits to the galaxy luminosity for the others. We also report on the expected capabilities for investigating the host galaxies of distant quasars with AO systems coupled with future Extremely Large Telescopes. Detailed simulations show that it will be possible to characterize compact (2-3 kpc) quasar host galaxies for quasi-stellar objects at z = 2 with nucleus K-magnitude spanning from 15 to 20 (corresponding to absolute magnitude -31 to -26) and host galaxies that are 4 mag fainter than their nuclei.
Impact of compressibility on heat transport characteristics of large terrestrial planets
NASA Astrophysics Data System (ADS)
Čížková, Hana; van den Berg, Arie; Jacobs, Michel
2017-07-01
We present heat transport characteristics for mantle convection in large terrestrial exoplanets (M ⩽ 8M⊕) . Our thermal convection model is based on a truncated anelastic liquid approximation (TALA) for compressible fluids and takes into account a selfconsistent thermodynamic description of material properties derived from mineral physics based on a multi-Einstein vibrational approach. We compare heat transport characteristics in compressible models with those obtained with incompressible models based on the classical- and extended Boussinesq approximation (BA and EBA respectively). Our scaling analysis shows that heat flux scales with effective dissipation number as Nu ∼Dieff-0.71 and with Rayleigh number as Nu ∼Raeff0.27. The surface heat flux of the BA models strongly overestimates the values from the corresponding compressible models, whereas the EBA models systematically underestimate the heat flux by ∼10%-15% with respect to a corresponding compressible case. Compressible models are also systematically warmer than the EBA models. Compressibility effects are therefore important for mantle dynamic processes, especially for large rocky exoplanets and consequently also for formation of planetary atmospheres, through outgassing, and the existence of a magnetic field, through thermal coupling of mantle and core dynamic systems.
Drilling Polar Oceans with the European Research Icebreaker AURORA BOREALIS: the IODP Context
NASA Astrophysics Data System (ADS)
Lembke-Jene, Lester; Wolff-Boenisch, Bonnie; Azzolini, Roberto; Thiede, Joern; Biebow, Nicole; Eldholm, Olav; Egerton, Paul
2010-05-01
Polar oceans are characterized by extreme environmental conditions for humans and materials, and have remained the least accessible regions to scientists of the IODP. DSDP and ODP have for long faced specific technical and logistical problems when attempting to drill in ice-covered polar deep-sea basins. The Arctic Ocean and large areas of the high-latitude Southern Ocean remained largely un-sampled by ODP and remain one of the major scientific and technological challenges for IODP. Drilling in these regions has been discussed and anticipated for decades and the scientific rationales are reflected in the science plans of the international Nansen Arctic Drilling Program (NAD) or the Arctic Program Planning Group (APPG) of ODP/IODP, amongst others. More recently, the rationale to investigate the polar oceans in a holistic approach has been outlined by workshops, leading to strategic assessments of the scientific potential and new drilling proposals. The European Polar Board took the initiative to develop a plan for a novel and dedicated research icebreaker with technical capabilities hitherto unrealised. This research icebreaker will enable autonomous operations in the central Arctic Ocean and the Southern Ocean, even during the severest ice conditions in the deep winter, serving all marine disciplines of polar research including scientific drilling: The European Research Icebreaker and Deep-Sea Drilling Vessel AURORA BOREALIS. AURORA BOREALIS is presently planned as a multi-purpose vessel. The ship can be deployed as a research icebreaker in all polar waters during any season of the year, as it shall meet the specifications of the highest ice-class attainable (IACS Polar Code 1) for icebreakers. During the times when it is not employed for drilling, it will operate as the most technically advanced multi-disciplinary research vessel in the Arctic or polar Southern Ocean. AURORA BOREALIS will be a "European scientific flagship facility" (fully open to non-European partners), a multidisciplinary platform for studies ranging from the sub-seafloor into the atmosphere. AURORA BOREALIS was planned for her role in deep-sea drilling in consultation with engineers and technical experts familiar with the program and the operation of these vessels. All techniques currently deployed on IODP expeditions can be implemented onboard the vessel under polar weather and ice conditions, including the full range of re-entry, casing and cementing, and instrumentation options and the entire suite of downhole logging tools. Due to sufficient laboratory space, a full analytical workflow can be easily established comparable to existing permanent platforms, including clean rooms, diverse scanning and logging or incubation facilities. While the vessel is equipped with a dedicated deep-sea drilling rig, other coring and drilling techniques can be employed if needed (e.g. Rockdrill, MEBO, large diameter Kasten cores). AURORA BOREALIS is fitted to operate a CALYPSO Piston Coring System in polar waters. Future mud-return systems under consideration and testing for IODP to provide controlled borehole conditions in difficult facies are compatible with the layout of AURORA BOREALIS. The berthing capacity of 120 personnel total (scientists, technical support and crew) allows to accommodate a sufficient number of science party members offshore. The present scientific implementation documents plan for about one polar scientific drilling expedition per year in a to-be-determined configuration. As the vessel is a multi-dsiciplinary platform, operations for the entire year are not dependant on drilling operations alone. While principal access to the vessel will be based on a competitive proposal review and evaluation system, the allocation of timeslots specifically for drilling would preferably be given over to IODP handling and planning systems in a cooperative mode using the strengths and capacitites of the future program. Depending on interests and needs of the scientific communities a preferential focus in non-drilling expedition planning could be established e.g. for dedicated geophysical pre-site survey works in areas inaccessible by other vessels to secure critical data needed for later drilling expeditions. Based on ongoing expert consultations, it is safe to assume that the average costs for an Arctic or polar drilling expedition will be considerably lower than with an otherwise necessary multi-ship setup based on modelled expedition scenarios and annual operational cost calculations. Still, AURORA BOREALIS shall provide substantially enhanced scientific, operational, personnel and technical capacities offshore.
An Approach for Autonomy: A Collaborative Communication Framework for Multi-Agent Systems
NASA Technical Reports Server (NTRS)
Dufrene, Warren Russell, Jr.
2005-01-01
Research done during the last three years has studied the emersion properties of Complex Adaptive Systems (CAS). The deployment of Artificial Intelligence (AI) techniques applied to remote Unmanned Aerial Vehicles has led the author to investigate applications of CAS within the field of Autonomous Multi-Agent Systems. The core objective of current research efforts is focused on the simplicity of Intelligent Agents (IA) and the modeling of these agents within complex systems. This research effort looks at the communication, interaction, and adaptability of multi-agents as applied to complex systems control. The embodiment concept applied to robotics has application possibilities within multi-agent frameworks. A new framework for agent awareness within a virtual 3D world concept is possible where the vehicle is composed of collaborative agents. This approach has many possibilities for applications to complex systems. This paper describes the development of an approach to apply this virtual framework to the NASA Goddard Space Flight Center (GSFC) tetrahedron structure developed under the Autonomous Nano Technology Swarm (ANTS) program and the Super Miniaturized Addressable Reconfigurable Technology (SMART) architecture program. These projects represent an innovative set of novel concepts deploying adaptable, self-organizing structures composed of many tetrahedrons. This technology is pushing current applied Agents Concepts to new levels of requirements and adaptability.
Proton core-beam system in the expanding solar wind: Hybrid simulations
NASA Astrophysics Data System (ADS)
Hellinger, Petr; Trávníček, Pavel M.
2011-11-01
Results of a two-dimensional hybrid expanding box simulation of a proton beam-core system in the solar wind are presented. The expansion with a strictly radial magnetic field leads to a decrease of the ratio between the proton perpendicular and parallel temperatures as well as to an increase of the ratio between the beam-core differential velocity and the local Alfvén velocity creating a free energy for many different instabilities. The system is indeed most of the time marginally stable with respect to the parallel magnetosonic, oblique Alfvén, proton cyclotron and parallel fire hose instabilities which determine the system evolution counteracting some effects of the expansion and interacting with each other. Nonlinear evolution of these instabilities leads to large modifications of the proton velocity distribution function. The beam and core protons are slowed with respect to each other and heated, and at later stages of the evolution the two populations are not clearly distinguishable. On the macroscopic level the instabilities cause large departures from the double adiabatic prediction leading to an efficient isotropization of effective proton temperatures in agreement with Helios observations.
A novel medical image data-based multi-physics simulation platform for computational life sciences.
Neufeld, Esra; Szczerba, Dominik; Chavannes, Nicolas; Kuster, Niels
2013-04-06
Simulating and modelling complex biological systems in computational life sciences requires specialized software tools that can perform medical image data-based modelling, jointly visualize the data and computational results, and handle large, complex, realistic and often noisy anatomical models. The required novel solvers must provide the power to model the physics, biology and physiology of living tissue within the full complexity of the human anatomy (e.g. neuronal activity, perfusion and ultrasound propagation). A multi-physics simulation platform satisfying these requirements has been developed for applications including device development and optimization, safety assessment, basic research, and treatment planning. This simulation platform consists of detailed, parametrized anatomical models, a segmentation and meshing tool, a wide range of solvers and optimizers, a framework for the rapid development of specialized and parallelized finite element method solvers, a visualization toolkit-based visualization engine, a Python scripting interface for customized applications, a coupling framework, and more. Core components are cross-platform compatible and use open formats. Several examples of applications are presented: hyperthermia cancer treatment planning, tumour growth modelling, evaluating the magneto-haemodynamic effect as a biomarker and physics-based morphing of anatomical models.
Fast data reconstructed method of Fourier transform imaging spectrometer based on multi-core CPU
NASA Astrophysics Data System (ADS)
Yu, Chunchao; Du, Debiao; Xia, Zongze; Song, Li; Zheng, Weijian; Yan, Min; Lei, Zhenggang
2017-10-01
Imaging spectrometer can gain two-dimensional space image and one-dimensional spectrum at the same time, which shows high utility in color and spectral measurements, the true color image synthesis, military reconnaissance and so on. In order to realize the fast reconstructed processing of the Fourier transform imaging spectrometer data, the paper designed the optimization reconstructed algorithm with OpenMP parallel calculating technology, which was further used for the optimization process for the HyperSpectral Imager of `HJ-1' Chinese satellite. The results show that the method based on multi-core parallel computing technology can control the multi-core CPU hardware resources competently and significantly enhance the calculation of the spectrum reconstruction processing efficiency. If the technology is applied to more cores workstation in parallel computing, it will be possible to complete Fourier transform imaging spectrometer real-time data processing with a single computer.
RTEMS SMP and MTAPI for Efficient Multi-Core Space Applications on LEON3/LEON4 Processors
NASA Astrophysics Data System (ADS)
Cederman, Daniel; Hellstrom, Daniel; Sherrill, Joel; Bloom, Gedare; Patte, Mathieu; Zulianello, Marco
2015-09-01
This paper presents the final result of an European Space Agency (ESA) activity aimed at improving the software support for LEON processors used in SMP configurations. One of the benefits of using a multicore system in a SMP configuration is that in many instances it is possible to better utilize the available processing resources by load balancing between cores. This however comes with the cost of having to synchronize operations between cores, leading to increased complexity. While in an AMP system one can use multiple instances of operating systems that are only uni-processor capable, a SMP system requires the operating system to be written to support multicore systems. In this activity we have improved and extended the SMP support of the RTEMS real-time operating system and ensured that it fully supports the multicore capable LEON processors. The targeted hardware in the activity has been the GR712RC, a dual-core core LEON3FT processor, and the functional prototype of ESA's Next Generation Multiprocessor (NGMP), a quad core LEON4 processor. The final version of the NGMP is now available as a product under the name GR740. An implementation of the Multicore Task Management API (MTAPI) has been developed as part of this activity to aid in the parallelization of applications for RTEMS SMP. It allows for simplified development of parallel applications using the task-based programming model. An existing space application, the Gaia Video Processing Unit, has been ported to RTEMS SMP using the MTAPI implementation to demonstrate the feasibility and usefulness of multicore processors for space payload software. The activity is funded by ESA under contract 4000108560/13/NL/JK. Gedare Bloom is supported in part by NSF CNS-0934725.
Multiscale Multiphysics Caprock Seal Analysis: A Case Study of the Farnsworth Unit, Texas, USA
NASA Astrophysics Data System (ADS)
Heath, J. E.; Dewers, T. A.; Mozley, P.
2015-12-01
Caprock sealing behavior depends on coupled processes that operate over a variety of length and time scales. Capillary sealing behavior depends on nanoscale pore throats and interfacial fluid properties. Larger-scale sedimentary architecture, fractures, and faults may govern properties of potential "seal-bypass" systems. We present the multiscale multiphysics investigation of sealing integrity of the caprock system that overlies the Morrow Sandstone reservoir, Farnsworth Unit, Texas. The Morrow Sandstone is the target injection unit for an on-going combined enhanced oil recovery-CO2 storage project by the Southwest Regional Partnership on Carbon Sequestration (SWP). Methods include small-to-large scale measurement techniques, including: focused ion beam-scanning electron microscopy; laser scanning confocal microscopy; electron and optical petrography; core examinations of sedimentary architecture and fractures; geomechanical testing; and a noble gas profile through sealing lithologies into the reservoir, as preserved from fresh core. The combined data set is used as part of a performance assessment methodology. The authors gratefully acknowledge the U.S. Department of Energy's (DOE) National Energy Technology Laboratory for sponsoring this project through the SWP under Award No. DE-FC26-05NT42591. Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-AC04-94AL85000.
Tolbert, Jeremy R; Kabali, Pratik; Brar, Simeranjit; Mukhopadhyay, Saibal
2009-01-01
We present a digital system for adaptive data compression for low power wireless transmission of Electroencephalography (EEG) data. The proposed system acts as a base-band processor between the EEG analog-to-digital front-end and RF transceiver. It performs a real-time accuracy energy trade-off for multi-channel EEG signal transmission by controlling the volume of transmitted data. We propose a multi-core digital signal processor for on-chip processing of EEG signals, to detect signal information of each channel and perform real-time adaptive compression. Our analysis shows that the proposed approach can provide significant savings in transmitter power with minimal impact on the overall signal accuracy.
Nagaoka, Tomoaki; Watanabe, Soichi
2012-01-01
Electromagnetic simulation with anatomically realistic computational human model using the finite-difference time domain (FDTD) method has recently been performed in a number of fields in biomedical engineering. To improve the method's calculation speed and realize large-scale computing with the computational human model, we adapt three-dimensional FDTD code to a multi-GPU cluster environment with Compute Unified Device Architecture and Message Passing Interface. Our multi-GPU cluster system consists of three nodes. The seven GPU boards (NVIDIA Tesla C2070) are mounted on each node. We examined the performance of the FDTD calculation on multi-GPU cluster environment. We confirmed that the FDTD calculation on the multi-GPU clusters is faster than that on a multi-GPU (a single workstation), and we also found that the GPU cluster system calculate faster than a vector supercomputer. In addition, our GPU cluster system allowed us to perform the large-scale FDTD calculation because were able to use GPU memory of over 100 GB.
Parallel Computer System for 3D Visualization Stereo on GPU
NASA Astrophysics Data System (ADS)
Al-Oraiqat, Anas M.; Zori, Sergii A.
2018-03-01
This paper proposes the organization of a parallel computer system based on Graphic Processors Unit (GPU) for 3D stereo image synthesis. The development is based on the modified ray tracing method developed by the authors for fast search of tracing rays intersections with scene objects. The system allows significant increase in the productivity for the 3D stereo synthesis of photorealistic quality. The generalized procedure of 3D stereo image synthesis on the Graphics Processing Unit/Graphics Processing Clusters (GPU/GPC) is proposed. The efficiency of the proposed solutions by GPU implementation is compared with single-threaded and multithreaded implementations on the CPU. The achieved average acceleration in multi-thread implementation on the test GPU and CPU is about 7.5 and 1.6 times, respectively. Studying the influence of choosing the size and configuration of the computational Compute Unified Device Archi-tecture (CUDA) network on the computational speed shows the importance of their correct selection. The obtained experimental estimations can be significantly improved by new GPUs with a large number of processing cores and multiprocessors, as well as optimized configuration of the computing CUDA network.
Revealing Thermal Instabilities in the Core of the Phoenix Cluster
NASA Astrophysics Data System (ADS)
McDonald, Michael
2017-08-01
The Phoenix cluster is the most relaxed cluster known, and hosts the strongest cool core of any cluster yet discovered. At the center of this cluster is a massive starburst galaxy, with a SFR of 500-1000 Msun/yr, seemingly satisfying the early cooling flow predictions, despite the presence of strong AGN feedback from the central supermassive black hole. Here we propose deep narrow-band imaging of the central 120 kpc of the cluster, to map the warm (10^4K) ionized gas via the [O II] emission line. In low-z clusters, such as Perseus and Abell 1795, the warm, ionized phase is of critical importance to map out thermal instabilities in the hot gas, and maps of Halpha and [O II] have been used for decades to understand how (and how not) cooling proceeds in the intracluster medium. The data proposed for here, combined with deep ALMA data, a recently-approved Large Chandra Program, and recently-approved multi-frequency JVLA data, will allow us to probe the cooling ICM, the cool, filamentary gas, the cold molecular gas, the star-forming population, and the AGN jets all on scales of <10 kpc. This multi-observatory campaign, focusing on the most extreme cooling cluster, will lead to a more complete understanding of how and why thermal instabilities develop in the hot ICM of cool core clusters.
Fast Image Subtraction Using Multi-cores and GPUs
NASA Astrophysics Data System (ADS)
Hartung, Steven; Shukla, H.
2013-01-01
Many important image processing techniques in astronomy require a massive number of computations per pixel. Among them is an image differencing technique known as Optimal Image Subtraction (OIS), which is very useful for detecting and characterizing transient phenomena. Like many image processing routines, OIS computations increase proportionally with the number of pixels being processed, and the number of pixels in need of processing is increasing rapidly. Utilizing many-core graphical processing unit (GPU) technology in a hybrid conjunction with multi-core CPU and computer clustering technologies, this work presents a new astronomy image processing pipeline architecture. The chosen OIS implementation focuses on the 2nd order spatially-varying kernel with the Dirac delta function basis, a powerful image differencing method that has seen limited deployment in part because of the heavy computational burden. This tool can process standard image calibration and OIS differencing in a fashion that is scalable with the increasing data volume. It employs several parallel processing technologies in a hierarchical fashion in order to best utilize each of their strengths. The Linux/Unix based application can operate on a single computer, or on an MPI configured cluster, with or without GPU hardware. With GPU hardware available, even low-cost commercial video cards, the OIS convolution and subtraction times for large images can be accelerated by up to three orders of magnitude.
NASA Astrophysics Data System (ADS)
Joglekar, Prasad; Shastry, Karthik; Hulbert, Steven; Weiss, Alex
2014-03-01
Auger Photoelectron Coincidence Spectroscopy (APECS), in which the Auger spectra is measured in coincidence with the core level photoelectron, is capable of pulling difficult to observe low energy Auger peaks out of a large background due mostly to inelastically scattered valence band photoelectrons. However the APECS method alone cannot eliminate the background due to valence band VB photoemission processes in which the initial photon energy is shared by 2 or more electrons and one of the electrons is in the energy range of the core level photoemission peak. Here we describe an experimental method for estimating the contributions from these background processes in the case of an Ag N23VV Auger spectra obtained in coincidence with the 4p photoemission peak. A beam of 180eV photons was incident on a Ag sample and a series of coincidence measurements were made with one cylindrical mirror analyzer (CMA) set at a fixed energies between the core and the valence band and the other CMA scanned over a range corresponding to electrons leaving the surface between 0eV and the 70eV. The spectra obtained were then used to obtain an estimate of the background in the APECS spectra due to multi-electron and inelastic VB photoemission processes. NSF, Welch Foundation.
EINSTEIN-HOME DISCOVERY OF 24 PULSARS IN THE PARKES MULTI-BEAM PULSAR SURVEY
DOE Office of Scientific and Technical Information (OSTI.GOV)
Knispel, B.; Kim, H.; Allen, B.
2013-09-10
We have conducted a new search for radio pulsars in compact binary systems in the Parkes multi-beam pulsar survey (PMPS) data, employing novel methods to remove the Doppler modulation from binary motion. This has yielded unparalleled sensitivity to pulsars in compact binaries. The required computation time of Almost-Equal-To 17, 000 CPU core years was provided by the distributed volunteer computing project Einstein-Home, which has a sustained computing power of about 1 PFlop s{sup -1}. We discovered 24 new pulsars in our search, 18 of which were isolated pulsars, and 6 were members of binary systems. Despite the wide filterbank channelsmore » and relatively slow sampling time of the PMPS data, we found pulsars with very large ratios of dispersion measure (DM) to spin period. Among those is PSR J1748-3009, the millisecond pulsar with the highest known DM ( Almost-Equal-To 420 pc cm{sup -3}). We also discovered PSR J1840-0643, which is in a binary system with an orbital period of 937 days, the fourth largest known. The new pulsar J1750-2536 likely belongs to the rare class of intermediate-mass binary pulsars. Three of the isolated pulsars show long-term nulling or intermittency in their emission, further increasing this growing family. Our discoveries demonstrate the value of distributed volunteer computing for data-driven astronomy and the importance of applying new analysis methods to extensively searched data.« less
SAXS analysis of single- and multi-core iron oxide magnetic nanoparticles
Szczerba, Wojciech; Costo, Rocio; Morales, Maria del Puerto; Thünemann, Andreas F.
2017-01-01
This article reports on the characterization of four superparamagnetic iron oxide nanoparticles stabilized with dimercaptosuccinic acid, which are suitable candidates for reference materials for magnetic properties. Particles p1 and p2 are single-core particles, while p3 and p4 are multi-core particles. Small-angle X-ray scattering analysis reveals a lognormal type of size distribution for the iron oxide cores of the particles. Their mean radii are 6.9 nm (p1), 10.6 nm (p2), 5.5 nm (p3) and 4.1 nm (p4), with narrow relative distribution widths of 0.08, 0.13, 0.08 and 0.12. The cores are arranged as a clustered network in the form of dense mass fractals with a fractal dimension of 2.9 in the multi-core particles p3 and p4, but the cores are well separated from each other by a protecting organic shell. The radii of gyration of the mass fractals are 48 and 44 nm, and each network contains 117 and 186 primary particles, respectively. The radius distributions of the primary particle were confirmed with transmission electron microscopy. All particles contain purely maghemite, as shown by X-ray absorption fine structure spectroscopy. PMID:28381973
Shen, Wenfeng; Wei, Daming; Xu, Weimin; Zhu, Xin; Yuan, Shizhong
2010-10-01
Biological computations like electrocardiological modelling and simulation usually require high-performance computing environments. This paper introduces an implementation of parallel computation for computer simulation of electrocardiograms (ECGs) in a personal computer environment with an Intel CPU of Core (TM) 2 Quad Q6600 and a GPU of Geforce 8800GT, with software support by OpenMP and CUDA. It was tested in three parallelization device setups: (a) a four-core CPU without a general-purpose GPU, (b) a general-purpose GPU plus 1 core of CPU, and (c) a four-core CPU plus a general-purpose GPU. To effectively take advantage of a multi-core CPU and a general-purpose GPU, an algorithm based on load-prediction dynamic scheduling was developed and applied to setting (c). In the simulation with 1600 time steps, the speedup of the parallel computation as compared to the serial computation was 3.9 in setting (a), 16.8 in setting (b), and 20.0 in setting (c). This study demonstrates that a current PC with a multi-core CPU and a general-purpose GPU provides a good environment for parallel computations in biological modelling and simulation studies. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.
2005-09-01
paleoceanographic and terrestrial climate proxies . Greenland ice cores, in particular, provide evidence of large amplitude, very rapid climate change during...received the most attention because it is the largest Holocene excursion in the GISP2 810 record [Alley et al., 1997]. Multiple proxies in Greenland ice...latitude North Atlantic foraminiferal-based proxies such as modem analogue technique [Marchal et al., 2002; Risebrobakken et al., 2003], but
Improved multi-beam laser interference lithography system by vibration analysis model
NASA Astrophysics Data System (ADS)
Lin, Te Hsun; Yang, Yin-Kuang; Mai, Hsuan-Ying; Fu, Chien-Chung
2017-03-01
This paper has developed the multi-beam laser interference lithography (LIL) system for nano/micro pattern sapphire substrate process (PSS/NPSS). However, the multi-beam LIL system is very sensitive to the light source and the vibration. When there is a vibration source in the exposure environment, the standing wave distribution on the substrate will be affected by the vibration and move in a certain angle. As a result, Moiré fringe defects occur on the exposure result. In order to eliminate the effect of the vibration, we use the software ANSYS to analyze the resonant frequencies of our multi-beam LIL system. Therefore, we need to design new multi-beam LIL system to raise the value of resonant frequencies. The new design of the multi-beam LIL system has higher resonant frequencies and successfully eliminates the bending and rotating effect of the resonant frequencies. As a result, the new multi-beam LIL system can fabricate large area and defects free period structures.
Single-mode tunable laser emission in the single-exciton regime from colloidal nanocrystals
Grivas, Christos; Li, Chunyong; Andreakou, Peristera; Wang, Pengfei; Ding, Ming; Brambilla, Gilberto; Manna, Liberato; Lagoudakis, Pavlos
2013-01-01
Whispering-gallery-mode resonators have been extensively used in conjunction with different materials for the development of a variety of photonic devices. Among the latter, hybrid structures, consisting of dielectric microspheres and colloidal core/shell semiconductor nanocrystals as gain media, have attracted interest for the development of microlasers and studies of cavity quantum electrodynamic effects. Here we demonstrate single-exciton, single-mode, spectrally tuned lasing from ensembles of optical antenna-designed, colloidal core/shell CdSe/CdS quantum rods deposited on silica microspheres. We obtain single-exciton emission by capitalizing on the band structure of the specific core/shell architecture that strongly localizes holes in the core, and the two-dimensional quantum confinement of electrons across the elongated shell. This creates a type-II conduction band alignment driven by coulombic repulsion that eliminates non-radiative multi-exciton Auger recombination processes, thereby inducing a large exciton–bi-exciton energy shift. Their ultra-low thresholds and single-mode, single-exciton emission make these hybrid lasers appealing for various applications, including quantum information processing. PMID:23974520
Stefik, Morgan; Mahajan, Surbhi; Sai, Hiroaki; Epps, Thomas H.; Bates, Frank S.; Gruner, Sol M; DiSalvo, Francis J.; Wiesner, Ulrich
2009-01-01
We report the first use of a non-frustrated block terpolymer for the synthesis of highly ordered oxide nanocomposites containing multiple plies. The morphological behavior of 15 ISO-oxide nanocomposites was investigated spanning a large range of compositions along the ƒI=ƒS isopleth using aluminosilicate and niobia sols. Morphologies were determined by TEM and SAXS measurements. Four morphologies were identified, including core-shell hexagonal, core-shell double gyroid, three-domain lamellae, and core-shell inverse-hexagonal, in order of increasing O+oxide vol fraction. All of the resulting nanocomposites had three- or five-ply morphologies containing domains that were continuous in one, two, or three dimensions. The five-ply core-shell double gyroid phase was only found to be stable when the O+oxide domain was a minority. Removal of the polymer enabled simple and direct synthesis of mesoporous oxide materials while retaining the ordered network structure. We believe that advances in the synthesis of multi-ply nanocomposites will lead to advanced materials and devices containing multiple plies of functional materials. PMID:20209023
Stefik, Morgan; Mahajan, Surbhi; Sai, Hiroaki; Epps, Thomas H; Bates, Frank S; Gruner, Sol M; Disalvo, Francis J; Wiesner, Ulrich
2009-11-24
We report the first use of a non-frustrated block terpolymer for the synthesis of highly ordered oxide nanocomposites containing multiple plies. The morphological behavior of 15 ISO-oxide nanocomposites was investigated spanning a large range of compositions along the ƒ(I)=ƒ(S) isopleth using aluminosilicate and niobia sols. Morphologies were determined by TEM and SAXS measurements. Four morphologies were identified, including core-shell hexagonal, core-shell double gyroid, three-domain lamellae, and core-shell inverse-hexagonal, in order of increasing O+oxide vol fraction. All of the resulting nanocomposites had three- or five-ply morphologies containing domains that were continuous in one, two, or three dimensions. The five-ply core-shell double gyroid phase was only found to be stable when the O+oxide domain was a minority. Removal of the polymer enabled simple and direct synthesis of mesoporous oxide materials while retaining the ordered network structure. We believe that advances in the synthesis of multi-ply nanocomposites will lead to advanced materials and devices containing multiple plies of functional materials.
Wetzel, Lucas; Jörg, David J.; Pollakis, Alexandros; Rave, Wolfgang; Fettweis, Gerhard; Jülicher, Frank
2017-01-01
Self-organized synchronization occurs in a variety of natural and technical systems but has so far only attracted limited attention as an engineering principle. In distributed electronic systems, such as antenna arrays and multi-core processors, a common time reference is key to coordinate signal transmission and processing. Here we show how the self-organized synchronization of mutually coupled digital phase-locked loops (DPLLs) can provide robust clocking in large-scale systems. We develop a nonlinear phase description of individual and coupled DPLLs that takes into account filter impulse responses and delayed signal transmission. Our phase model permits analytical expressions for the collective frequencies of synchronized states, the analysis of stability properties and the time scale of synchronization. In particular, we find that signal filtering introduces stability transitions that are not found in systems without filtering. To test our theoretical predictions, we designed and carried out experiments using networks of off-the-shelf DPLL integrated circuitry. We show that the phase model can quantitatively predict the existence, frequency, and stability of synchronized states. Our results demonstrate that mutually delay-coupled DPLLs can provide robust and self-organized synchronous clocking in electronic systems. PMID:28207779
Accelerating cardiac bidomain simulations using graphics processing units.
Neic, A; Liebmann, M; Hoetzl, E; Mitchell, L; Vigmond, E J; Haase, G; Plank, G
2012-08-01
Anatomically realistic and biophysically detailed multiscale computer models of the heart are playing an increasingly important role in advancing our understanding of integrated cardiac function in health and disease. Such detailed simulations, however, are computationally vastly demanding, which is a limiting factor for a wider adoption of in-silico modeling. While current trends in high-performance computing (HPC) hardware promise to alleviate this problem, exploiting the potential of such architectures remains challenging since strongly scalable algorithms are necessitated to reduce execution times. Alternatively, acceleration technologies such as graphics processing units (GPUs) are being considered. While the potential of GPUs has been demonstrated in various applications, benefits in the context of bidomain simulations where large sparse linear systems have to be solved in parallel with advanced numerical techniques are less clear. In this study, the feasibility of multi-GPU bidomain simulations is demonstrated by running strong scalability benchmarks using a state-of-the-art model of rabbit ventricles. The model is spatially discretized using the finite element methods (FEM) on fully unstructured grids. The GPU code is directly derived from a large pre-existing code, the Cardiac Arrhythmia Research Package (CARP), with very minor perturbation of the code base. Overall, bidomain simulations were sped up by a factor of 11.8 to 16.3 in benchmarks running on 6-20 GPUs compared to the same number of CPU cores. To match the fastest GPU simulation which engaged 20 GPUs, 476 CPU cores were required on a national supercomputing facility.
Accelerating Cardiac Bidomain Simulations Using Graphics Processing Units
Neic, Aurel; Liebmann, Manfred; Hoetzl, Elena; Mitchell, Lawrence; Vigmond, Edward J.; Haase, Gundolf
2013-01-01
Anatomically realistic and biophysically detailed multiscale computer models of the heart are playing an increasingly important role in advancing our understanding of integrated cardiac function in health and disease. Such detailed simulations, however, are computationally vastly demanding, which is a limiting factor for a wider adoption of in-silico modeling. While current trends in high-performance computing (HPC) hardware promise to alleviate this problem, exploiting the potential of such architectures remains challenging since strongly scalable algorithms are necessitated to reduce execution times. Alternatively, acceleration technologies such as graphics processing units (GPUs) are being considered. While the potential of GPUs has been demonstrated in various applications, benefits in the context of bidomain simulations where large sparse linear systems have to be solved in parallel with advanced numerical techniques are less clear. In this study, the feasibility of multi-GPU bidomain simulations is demonstrated by running strong scalability benchmarks using a state-of-the-art model of rabbit ventricles. The model is spatially discretized using the finite element methods (FEM) on fully unstructured grids. The GPU code is directly derived from a large pre-existing code, the Cardiac Arrhythmia Research Package (CARP), with very minor perturbation of the code base. Overall, bidomain simulations were sped up by a factor of 11.8 to 16.3 in benchmarks running on 6–20 GPUs compared to the same number of CPU cores. To match the fastest GPU simulation which engaged 20GPUs, 476 CPU cores were required on a national supercomputing facility. PMID:22692867
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lichtner, Peter C.; Hammond, Glenn E.; Lu, Chuan
PFLOTRAN solves a system of generally nonlinear partial differential equations describing multi-phase, multicomponent and multiscale reactive flow and transport in porous materials. The code is designed to run on massively parallel computing architectures as well as workstations and laptops (e.g. Hammond et al., 2011). Parallelization is achieved through domain decomposition using the PETSc (Portable Extensible Toolkit for Scientific Computation) libraries for the parallelization framework (Balay et al., 1997). PFLOTRAN has been developed from the ground up for parallel scalability and has been run on up to 218 processor cores with problem sizes up to 2 billion degrees of freedom. Writtenmore » in object oriented Fortran 90, the code requires the latest compilers compatible with Fortran 2003. At the time of this writing this requires gcc 4.7.x, Intel 12.1.x and PGC compilers. As a requirement of running problems with a large number of degrees of freedom, PFLOTRAN allows reading input data that is too large to fit into memory allotted to a single processor core. The current limitation to the problem size PFLOTRAN can handle is the limitation of the HDF5 file format used for parallel IO to 32 bit integers. Noting that 2 32 = 4; 294; 967; 296, this gives an estimate of the maximum problem size that can be currently run with PFLOTRAN. Hopefully this limitation will be remedied in the near future.« less
NASA Astrophysics Data System (ADS)
Chakraborty, Sovan; Mirizzi, Alessandro; Saviano, Ninetta; Seixas, David de Sousa
2014-05-01
It has been recently pointed out that by removing the axial symmetry in the "multi-angle effects" associated with the neutrino-neutrino interactions for supernova (SN) neutrinos a new multi-azimuthal-angle (MAA) instability would arise. In particular, for a flux ordering Fνe>Fν ¯e>Fνx, as expected during the SN accretion phase, this instability occurs in the normal neutrino mass hierarchy. However, during this phase, the ordinary matter density can be larger than the neutrino one, suppressing the self-induced conversions. In this regard, we investigate the matter suppression of the MAA effects, performing a linearized stability analysis of the neutrino equations of motion, in the presence of realistic SN density profiles. We compare these results with the numerical solution of the SN neutrino nonlinear evolution equations. Assuming axially symmetric distributions of neutrino momenta, we find that the large matter term strongly inhibits the MAA effects. In particular, the hindrance becomes stronger including realistic forward-peaked neutrino angular distributions. As a result, in our model for a 10.8 M⊙ iron-core SNe, MAA instability does not trigger any flavor conversion during the accretion phase. Instead, for a 8.8 M⊙ O-Ne-Mg core SN model, with lower matter density profile and less forward-peaked angular distributions, flavor conversions are possible also at early times.
MUTILS - a set of efficient modeling tools for multi-core CPUs implemented in MEX
NASA Astrophysics Data System (ADS)
Krotkiewski, Marcin; Dabrowski, Marcin
2013-04-01
The need for computational performance is common in scientific applications, and in particular in numerical simulations, where high resolution models require efficient processing of large amounts of data. Especially in the context of geological problems the need to increase the model resolution to resolve physical and geometrical complexities seems to have no limits. Alas, the performance of new generations of CPUs does not improve any longer by simply increasing clock speeds. Current industrial trends are to increase the number of computational cores. As a result, parallel implementations are required in order to fully utilize the potential of new processors, and to study more complex models. We target simulations on small to medium scale shared memory computers: laptops and desktop PCs with ~8 CPU cores and up to tens of GB of memory to high-end servers with ~50 CPU cores and hundereds of GB of memory. In this setting MATLAB is often the environment of choice for scientists that want to implement their own models with little effort. It is a useful general purpose mathematical software package, but due to its versatility some of its functionality is not as efficient as it could be. In particular, the challanges of modern multi-core architectures are not fully addressed. We have developed MILAMIN 2 - an efficient FEM modeling environment written in native MATLAB. Amongst others, MILAMIN provides functions to define model geometry, generate and convert structured and unstructured meshes (also through interfaces to external mesh generators), compute element and system matrices, apply boundary conditions, solve the system of linear equations, address non-linear and transient problems, and perform post-processing. MILAMIN strives to combine the ease of code development and the computational efficiency. Where possible, the code is optimized and/or parallelized within the MATLAB framework. Native MATLAB is augmented with the MUTILS library - a set of MEX functions that implement the computationally intensive, performance critical parts of the code, which we have identified to be bottlenecks. Here, we discuss the functionality and performance of the MUTILS library. Currently, it includes: 1. time and memory efficient assembly of sparse matrices for FEM simulations 2. parallel sparse matrix - vector product with optimizations speficic to symmetric matrices and multiple degrees of freedom per node 3. parallel point in triangle location and point in tetrahedron location for unstructured, adaptive 2D and 3D meshes (useful for 'marker in cell' type of methods) 4. parallel FEM interpolation for 2D and 3D meshes of elements of different types and orders, and for different number of degrees of freedom per node 5. a stand-alone, MEX implementation of the Conjugate Gradients iterative solver 6. interface to METIS graph partitioning and a fast implementation of RCM reordering
ERIC Educational Resources Information Center
Rutland Consulting Group Ltd.
The report presents summaries of evaluations of the Coordinated Assessment and Program Planning for Education (CAPE) Program and the Coordinated Rehabilitation and Education (CORE) program for multi-handicapped sensory impaired and/or communication and behavior disordered children and their families in Alberta, Canada. Each program is evaluated…
CQPSO scheduling algorithm for heterogeneous multi-core DAG task model
NASA Astrophysics Data System (ADS)
Zhai, Wenzheng; Hu, Yue-Li; Ran, Feng
2017-07-01
Efficient task scheduling is critical to achieve high performance in a heterogeneous multi-core computing environment. The paper focuses on the heterogeneous multi-core directed acyclic graph (DAG) task model and proposes a novel task scheduling method based on an improved chaotic quantum-behaved particle swarm optimization (CQPSO) algorithm. A task priority scheduling list was built. A processor with minimum cumulative earliest finish time (EFT) was acted as the object of the first task assignment. The task precedence relationships were satisfied and the total execution time of all tasks was minimized. The experimental results show that the proposed algorithm has the advantage of optimization abilities, simple and feasible, fast convergence, and can be applied to the task scheduling optimization for other heterogeneous and distributed environment.
NASA Astrophysics Data System (ADS)
Ross, P.-S.; Bourke, A.
2017-01-01
Physical property measurements are increasingly important in mining exploration. For density determinations on rocks, one method applicable on exploration drill cores relies on gamma ray attenuation. This non-destructive method is ideal because each measurement takes only 10 s, making it suitable for high-resolution logging. However calibration has been problematic. In this paper we present new empirical, site-specific correction equations for whole NQ and BQ cores. The corrections force back the gamma densities to the "true" values established by the immersion method. For the NQ core caliber, the density range extends to high values (massive pyrite, 5 g/cm3) and the correction is thought to be very robust. We also present additional empirical correction factors for cut cores which take into account the missing material. These "cut core correction factors", which are not site-specific, were established by making gamma density measurements on truncated aluminum cylinders of various residual thicknesses. Finally we show two examples of application for the Abitibi Greenstone Belt in Canada. The gamma ray attenuation measurement system is part of a multi-sensor core logger which also determines magnetic susceptibility, geochemistry and mineralogy on rock cores, and performs line-scan imaging.
NASA Astrophysics Data System (ADS)
Gu, Yanchao; Fan, Dongming; You, Wei
2017-07-01
Eleven GPS crustal vertical displacement (CVD) solutions for 110 IGS08/IGS14 core stations provided by the International Global Navigation Satellite Systems Service Analysis Centers are compared with seven Gravity Recovery and Climate Experiment (GRACE)-modeled CVD solutions. The results of the internal comparison of the GPS solutions from multiple institutions imply large uncertainty in the GPS postprocessing. There is also evidence that GRACE solutions from both different institutions and different processing approaches (mascon and traditional spherical harmonic coefficients) show similar results, suggesting that GRACE can provide CVD results of good internal consistency. When the uncertainty of the GPS data is accounted for, the GRACE data can explain as much as 50% of the actual signals and more than 80% of the GPS annual signals. Our study strongly indicates that GRACE data have great potential to correct the nontidal loading in GPS time series.
Developments of scintillator-based soft x-ray diagnostic in LHD with CsI:Tl and P47 scintillators.
Bando, T; Ohdachi, S; Suzuki, Y
2016-11-01
Multi-channel soft x-ray (SX) diagnostic has been used in the large helical device (LHD) to research magnetohydrodynamic equilibria and activities. However, in the coming deuterium plasma experiments of LHD, it will be difficult to use semiconductor systems near LHD. Therefore, a new type of SX diagnostic, a scintillator-based type diagnostic, has been investigated in order to avoid damage from the radiation. A fiber optic plate coated by P47 scintillator will be used to detect SX emission. Scintillation light will be transferred by pure silica core optical fibers and detected by photomultiplier tubes. A vertically elongated section of LHD will be covered by a 13 ch. array. Effects from the Deuterium Deuterium neutrons can be negligible when the scintillator is covered by a Pb plate 4 cm in thickness to avoid gamma-rays.
Developments of scintillator-based soft x-ray diagnostic in LHD with CsI:Tl and P47 scintillators
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bando, T., E-mail: bando.takahiro@nifs.ac.jp; Ohdachi, S.; Suzuki, Y.
2016-11-15
Multi-channel soft x-ray (SX) diagnostic has been used in the large helical device (LHD) to research magnetohydrodynamic equilibria and activities. However, in the coming deuterium plasma experiments of LHD, it will be difficult to use semiconductor systems near LHD. Therefore, a new type of SX diagnostic, a scintillator-based type diagnostic, has been investigated in order to avoid damage from the radiation. A fiber optic plate coated by P47 scintillator will be used to detect SX emission. Scintillation light will be transferred by pure silica core optical fibers and detected by photomultiplier tubes. A vertically elongated section of LHD will bemore » covered by a 13 ch. array. Effects from the Deuterium Deuterium neutrons can be negligible when the scintillator is covered by a Pb plate 4 cm in thickness to avoid gamma-rays.« less
Ultra-sensitive all-fibre photothermal spectroscopy with large dynamic range
Jin, Wei; Cao, Yingchun; Yang, Fan; Ho, Hoi Lut
2015-01-01
Photothermal interferometry is an ultra-sensitive spectroscopic means for trace chemical detection in gas- and liquid-phase materials. Previous photothermal interferometry systems used free-space optics and have limitations in efficiency of light–matter interaction, size and optical alignment, and integration into photonic circuits. Here we exploit photothermal-induced phase change in a gas-filled hollow-core photonic bandgap fibre, and demonstrate an all-fibre acetylene gas sensor with a noise equivalent concentration of 2 p.p.b. (2.3 × 10−9 cm−1 in absorption coefficient) and an unprecedented dynamic range of nearly six orders of magnitude. The realization of photothermal interferometry with low-cost near infrared semiconductor lasers and fibre-based technology allows a class of optical sensors with compact size, ultra sensitivity and selectivity, applicability to harsh environment, and capability for remote and multiplexed multi-point detection and distributed sensing. PMID:25866015
NASA Astrophysics Data System (ADS)
Francis, T.
2003-04-01
HYACINTH is the acronym for "Development of HYACE tools in new tests on Hydrates". The project is being carried out by a consortium of six companies and academic institutions from Germany, The Netherlands and the United Kingdom. It is a European Framework Five project whose objective is to bring the pressure corers developed in the earlier HYACE project, together with new core handling technology developed in the HYACINTH project, to the operational stage. Our philosophy is that if all one does with a pressure core is to bleed off the gas it contains, a major scientific opportunity has been missed. The current system enables pressure cores to be acquired, then transferred, without loss of pressure, into laboratory chambers so that they can be geophysically logged. The suite of equipment - HYACE Rotary Corer (HRC), Fugro Pressure Corer (FPC), Shear Transfer Chamber (STC), Logging Chamber (LC), Storage Chamber (SC) and Vertical Multi-Sensor Core Logger (V-MSCL) - will be briefly described. Other developments currently in progress to extend the capabilities of the system will be summarised: - to allow electrical resistivity logging of the pressure cores - to enable pressurised sub-samples to be taken from the cores - to facilitate microbiological experiments on pressurised sub-samples The first scientific results obtained with the HYACE/HYACINTH technology were achieved on ODP Leg 204 and are the subject of another talk at this meeting.
Miller, Julie M; Dewey, Marc; Vavere, Andrea L; Rochitte, Carlos E; Niinuma, Hiroyuki; Arbab-Zadeh, Armin; Paul, Narinder; Hoe, John; de Roos, Albert; Yoshioka, Kunihiro; Lemos, Pedro A; Bush, David E; Lardo, Albert C; Texter, John; Brinker, Jeffery; Cox, Christopher; Clouse, Melvin E; Lima, João A C
2009-04-01
Multislice computed tomography (MSCT) for the noninvasive detection of coronary artery stenoses is a promising candidate for widespread clinical application because of its non-invasive nature and high sensitivity and negative predictive value as found in several previous studies using 16 to 64 simultaneous detector rows. A multi-centre study of CT coronary angiography using 16 simultaneous detector rows has shown that 16-slice CT is limited by a high number of nondiagnostic cases and a high false-positive rate. A recent meta-analysis indicated a significant interaction between the size of the study sample and the diagnostic odds ratios suggestive of small study bias, highlighting the importance of evaluating MSCT using 64 simultaneous detector rows in a multi-centre approach with a larger sample size. In this manuscript we detail the objectives and methods of the prospective "CORE-64" trial ("Coronary Evaluation Using Multidetector Spiral Computed Tomography Angiography using 64 Detectors"). This multi-centre trial was unique in that it assessed the diagnostic performance of 64-slice CT coronary angiography in nine centres worldwide in comparison to conventional coronary angiography. In conclusion, the multi-centre, multi-institutional and multi-continental trial CORE-64 has great potential to ultimately assess the per-patient diagnostic performance of coronary CT angiography using 64 simultaneous detector rows.
Final Report: Enabling Exascale Hardware and Software Design through Scalable System Virtualization
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bridges, Patrick G.
2015-02-01
In this grant, we enhanced the Palacios virtual machine monitor to increase its scalability and suitability for addressing exascale system software design issues. This included a wide range of research on core Palacios features, large-scale system emulation, fault injection, perfomrance monitoring, and VMM extensibility. This research resulted in large number of high-impact publications in well-known venues, the support of a number of students, and the graduation of two Ph.D. students and one M.S. student. In addition, our enhanced version of the Palacios virtual machine monitor has been adopted as a core element of the Hobbes operating system under active DOE-fundedmore » research and development.« less
Nanomechanical Optical Fiber with Embedded Electrodes Actuated by Joule Heating.
Lian, Zhenggang; Segura, Martha; Podoliak, Nina; Feng, Xian; White, Nicholas; Horak, Peter
2014-07-31
Nanomechanical optical fibers with metal electrodes embedded in the jacket were fabricated by a multi-material co-draw technique. At the center of the fibers, two glass cores suspended by thin membranes and surrounded by air form a directional coupler that is highly temperature-dependent. We demonstrate optical switching between the two fiber cores by Joule heating of the electrodes with as little as 0.4 W electrical power, thereby demonstrating an electrically actuated all-fiber microelectromechanical system (MEMS). Simulations show that the main mechanism for optical switching is the transverse thermal expansion of the fiber structure.
NASA Astrophysics Data System (ADS)
Dvorak, Steven L.; Sternberg, Ben K.; Feng, Wanjie
2017-03-01
In this paper we discuss the design and verification of wide-band, multi-frequency, tuning circuits for large-moment Transmitter (TX) loops. Since these multi-frequency, tuned-TX loops allow for the simultaneous transmission of multiple frequencies at high-current levels, they are ideally suited for frequency-domain geophysical systems that collect data while moving, such as helicopter mounted systems. Furthermore, since multi-frequency tuners use the same TX loop for all frequencies, instead of using separate tuned-TX loops for each frequency, they allow for the use of larger moment TX loops. In this paper we discuss the design and simulation of one- and three-frequency tuned TX loops and then present measurement results for a three-frequency, tuned-TX loop.
Performance of fully-coupled algebraic multigrid preconditioners for large-scale VMS resistive MHD
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lin, P. T.; Shadid, J. N.; Hu, J. J.
Here, we explore the current performance and scaling of a fully-implicit stabilized unstructured finite element (FE) variational multiscale (VMS) capability for large-scale simulations of 3D incompressible resistive magnetohydrodynamics (MHD). The large-scale linear systems that are generated by a Newton nonlinear solver approach are iteratively solved by preconditioned Krylov subspace methods. The efficiency of this approach is critically dependent on the scalability and performance of the algebraic multigrid preconditioner. Our study considers the performance of the numerical methods as recently implemented in the second-generation Trilinos implementation that is 64-bit compliant and is not limited by the 32-bit global identifiers of themore » original Epetra-based Trilinos. The study presents representative results for a Poisson problem on 1.6 million cores of an IBM Blue Gene/Q platform to demonstrate very large-scale parallel execution. Additionally, results for a more challenging steady-state MHD generator and a transient solution of a benchmark MHD turbulence calculation for the full resistive MHD system are also presented. These results are obtained on up to 131,000 cores of a Cray XC40 and one million cores of a BG/Q system.« less
Performance of fully-coupled algebraic multigrid preconditioners for large-scale VMS resistive MHD
Lin, P. T.; Shadid, J. N.; Hu, J. J.; ...
2017-11-06
Here, we explore the current performance and scaling of a fully-implicit stabilized unstructured finite element (FE) variational multiscale (VMS) capability for large-scale simulations of 3D incompressible resistive magnetohydrodynamics (MHD). The large-scale linear systems that are generated by a Newton nonlinear solver approach are iteratively solved by preconditioned Krylov subspace methods. The efficiency of this approach is critically dependent on the scalability and performance of the algebraic multigrid preconditioner. Our study considers the performance of the numerical methods as recently implemented in the second-generation Trilinos implementation that is 64-bit compliant and is not limited by the 32-bit global identifiers of themore » original Epetra-based Trilinos. The study presents representative results for a Poisson problem on 1.6 million cores of an IBM Blue Gene/Q platform to demonstrate very large-scale parallel execution. Additionally, results for a more challenging steady-state MHD generator and a transient solution of a benchmark MHD turbulence calculation for the full resistive MHD system are also presented. These results are obtained on up to 131,000 cores of a Cray XC40 and one million cores of a BG/Q system.« less
Multi-GPU Accelerated Admittance Method for High-Resolution Human Exposure Evaluation.
Xiong, Zubiao; Feng, Shi; Kautz, Richard; Chandra, Sandeep; Altunyurt, Nevin; Chen, Ji
2015-12-01
A multi-graphics processing unit (GPU) accelerated admittance method solver is presented for solving the induced electric field in high-resolution anatomical models of human body when exposed to external low-frequency magnetic fields. In the solver, the anatomical model is discretized as a three-dimensional network of admittances. The conjugate orthogonal conjugate gradient (COCG) iterative algorithm is employed to take advantage of the symmetric property of the complex-valued linear system of equations. Compared against the widely used biconjugate gradient stabilized method, the COCG algorithm can reduce the solving time by 3.5 times and reduce the storage requirement by about 40%. The iterative algorithm is then accelerated further by using multiple NVIDIA GPUs. The computations and data transfers between GPUs are overlapped in time by using asynchronous concurrent execution design. The communication overhead is well hidden so that the acceleration is nearly linear with the number of GPU cards. Numerical examples show that our GPU implementation running on four NVIDIA Tesla K20c cards can reach 90 times faster than the CPU implementation running on eight CPU cores (two Intel Xeon E5-2603 processors). The implemented solver is able to solve large dimensional problems efficiently. A whole adult body discretized in 1-mm resolution can be solved in just several minutes. The high efficiency achieved makes it practical to investigate human exposure involving a large number of cases with a high resolution that meets the requirements of international dosimetry guidelines.
First experience with particle-in-cell plasma physics code on ARM-based HPC systems
NASA Astrophysics Data System (ADS)
Sáez, Xavier; Soba, Alejandro; Sánchez, Edilberto; Mantsinen, Mervi; Mateo, Sergi; Cela, José M.; Castejón, Francisco
2015-09-01
In this work, we will explore the feasibility of porting a Particle-in-cell code (EUTERPE) to an ARM multi-core platform from the Mont-Blanc project. The used prototype is based on a system-on-chip Samsung Exynos 5 with an integrated GPU. It is the first prototype that could be used for High-Performance Computing (HPC), since it supports double precision and parallel programming languages.
Charge Weld Effects on High Cycle Fatigue Behavior of a Hollow Extruded AA6082 Profile
NASA Astrophysics Data System (ADS)
Nanninga, N.; White, C.; Dickson, R.
2011-10-01
Fatigue properties of specimens taken from different locations along the length of a hollow AA6082 extrusion, where charge weld (interface between successive billets in multi-billet extrusions) properties and the degree of coring (accumulation of highly sheared billet surface material at back end of billet) are expected to vary, have been evaluated. The fatigue strength of transverse specimens containing charge welds is lower near the front of the extrusion where the charge weld separation is relatively large. The relationship between fatigue failure and charge weld separation appears to be directly related to charge weld properties. The lower fatigue properties of the specimens are likely associated with early overload fatigue failure along the charge weld interface. Coring does not appear to have significantly affected fatigue behavior.
Multi-walled carbon/IF-WS2 nanoparticles with improved thermal properties.
Xu, Fang; Almeida, Trevor P; Chang, Hong; Xia, Yongde; Wears, M Lesley; Zhu, Yanqiu
2013-11-07
A unique new class of core-shell structured composite nanoparticles, C-coated inorganic fullerene-like WS2 (IF-WS2) hollow nanoparticles, has been created for the first time in large quantities, by a continuous chemical vapour deposition method using a rotary furnace. Transmission electron microscopy and Raman characterisations of the resulting samples reveal that the composite nanoparticles exhibited a uniform shell of carbon coating, ranging from 2-5 nm on the IF-WS2 core, with little or no agglomeration. Importantly, thermogravimetric analysis and differential scanning calorimetry analysis confirm that their thermal stability against oxidation in air has been improved by about 70 °C, compared to the pristine IF-WS2, making these new C-coated IF-WS2 nanoparticles more attractive for critical engineering applications.
Core-Collapse Supernovae Explored by Multi-D Boltzmann Hydrodynamic Simulations
NASA Astrophysics Data System (ADS)
Sumiyoshi, Kohsuke; Nagakura, Hiroki; Iwakami, Wakana; Furusawa, Shun; Matsufuru, Hideo; Imakura, Akira; Yamada, Shoichi
We report the latest results of numerical simulations of core-collapse supernovae by solving multi-D neutrino-radiation hydrodynamics with Boltzmann equations. One of the longstanding issues of the explosion mechanism of supernovae has been uncertainty in the approximations of the neutrino transfer in multi-D such as the diffusion approximation and ray-by-ray method. The neutrino transfer is essential, together with 2D/3D hydrodynamical instabilities, to evaluate the neutrino heating behind the shock wave for successful explosions and to predict the neutrino burst signals. We tackled this difficult problem by utilizing our solver of the 6D Boltzmann equation for neutrinos in 3D space and 3D neutrino momentum space coupled with multi-D hydrodynamics adding special and general relativistic extensions. We have performed a set of 2D core-collapse simulations from 11M ⊙ and 15M ⊙ stars on K-computer in Japan by following long-term evolution over 400 ms after bounce to reveal the outcome from the full Boltzmann hydrodynamic simulations with a sophisticated equation of state with multi-nuclear species and updated rates for electron captures on nuclei.
Widefield TSCSPC-systems with large-area-detectors: application in simultaneous multi-channel-FLIM
NASA Astrophysics Data System (ADS)
Stepanov, Sergei; Bakhlanov, Sergei; Drobchenko, Evgeny; Eckert, Hann-Jörg; Kemnitz, Klaus
2010-11-01
Novel proximity-type Time- and Space-Correlated Single Photon Counting (TSCSPC) crossed-delay-line (DL)- and multi-anode (MA)-systems of outstanding performance and homogeneity were developed, using large-area detector heads of 25 and 40 mm diameter. Instrument response functions IRF(space) = (60 +/- 5) μm FWHM and IRF(time) = (28 +/- 3) ps FWHM were achieved over the full 12 cm2 area of the detector. Deadtime at throughput of 105 cps is 10% for "high-resolution" system and 5% in the "video"-system at 106 cps, at slightly reduced time- and space resolution. A fluorescence lifetime of (3.5 +/- 1) ps can be recovered from multi-exponential dynamics of a single living cyanobacterium (Acaryochloris marina). The present large-area detectors are particularly useful in simultaneous multichannel applications, such as 2-colour anisotropy or 4-colour lifetime imaging, utilizing dual- or quad-view image splitters. The long-term stability, low- excitation-intensity (< 100 mW/cm2) widefield systems enable minimal-invasive observation, without significant bleaching or photodynamic reactions, thus allowing long-period observation of up to several hours in living cells.
NASA Astrophysics Data System (ADS)
Dee, S. G.; Parsons, L. A.; Loope, G. R.; Overpeck, J. T.; Ault, T. R.; Emile-Geay, J.
2017-10-01
The spectral characteristics of paleoclimate observations spanning the last millennium suggest the presence of significant low-frequency (multi-decadal to centennial scale) variability in the climate system. Since this low-frequency climate variability is critical for climate predictions on societally-relevant scales, it is essential to establish whether General Circulation models (GCMs) are able to simulate it faithfully. Recent studies find large discrepancies between models and paleoclimate data at low frequencies, prompting concerns surrounding the ability of GCMs to predict long-term, high-magnitude variability under greenhouse forcing (Laepple and Huybers, 2014a, 2014b). However, efforts to ground climate model simulations directly in paleoclimate observations are impeded by fundamental differences between models and the proxy data: proxy systems often record a multivariate and/or nonlinear response to climate, precluding a direct comparison to GCM output. In this paper we bridge this gap via a forward proxy modeling approach, coupled to an isotope-enabled GCM. This allows us to disentangle the various contributions to signals embedded in ice cores, speleothem calcite, coral aragonite, tree-ring width, and tree-ring cellulose. The paper addresses the following questions: (1) do forward-modeled ;pseudoproxies; exhibit variability comparable to proxy data? (2) if not, which processes alter the shape of the spectrum of simulated climate variability, and are these processes broadly distinguishable from climate? We apply our method to representative case studies, and broaden these insights with an analysis of the PAGES2k database (PAGES2K Consortium, 2013). We find that current proxy system models (PSMs) can help resolve model-data discrepancies on interannual to decadal timescales, but cannot account for the mismatch in variance on multi-decadal to centennial timescales. We conclude that, specific to this set of PSMs and isotope-enabled model, the paleoclimate record may exhibit larger low-frequency variability than GCMs currently simulate, indicative of incomplete physics and/or forcings.
NASA Astrophysics Data System (ADS)
Napoli, Jay
2016-05-01
Precision fiber optic gyroscopes (FOGs) are critical components for an array of platforms and applications ranging from stabilization and pointing orientation of payloads and platforms to navigation and control for unmanned and autonomous systems. In addition, FOG-based inertial systems provide extremely accurate data for geo-referencing systems. Significant improvements in the performance of FOGs and FOG-based inertial systems at KVH are due, in large part, to advancements in the design and manufacture of optical fiber, as well as in manufacturing operations and signal processing. Open loop FOGs, such as those developed and manufactured by KVH Industries, offer tactical-grade performance in a robust, small package. The success of KVH FOGs and FOG-based inertial systems is due to innovations in key fields, including the development of proprietary D-shaped fiber with an elliptical core, and KVH's unique ThinFiber. KVH continually improves its FOG manufacturing processes and signal processing, which result in improved accuracies across its entire FOG product line. KVH acquired its FOG capabilities, including its patented E•Core fiber, when the company purchased Andrew Corporation's Fiber Optic Group in 1997. E•Core fiber is unique in that the light-guiding core - critical to the FOG's performance - is elliptically shaped. The elliptical core produces a fiber that has low loss and high polarization-maintaining ability. In 2010, KVH developed its ThinFiber, a 170-micron diameter fiber that retains the full performance characteristics of E•Core fiber. ThinFiber has enabled the development of very compact, high-performance open-loop FOGs, which are also used in a line of FOG-based inertial measurement units and inertial navigation systems.
Parallelized Kalman-Filter-Based Reconstruction of Particle Tracks on Many-Core Processors and GPUs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cerati, Giuseppe; Elmer, Peter; Krutelyov, Slava
2017-01-01
For over a decade now, physical and energy constraints have limited clock speed improvements in commodity microprocessors. Instead, chipmakers have been pushed into producing lower-power, multi-core processors such as Graphical Processing Units (GPU), ARM CPUs, and Intel MICs. Broad-based efforts from manufacturers and developers have been devoted to making these processors user-friendly enough to perform general computations. However, extracting performance from a larger number of cores, as well as specialized vector or SIMD units, requires special care in algorithm design and code optimization. One of the most computationally challenging problems in high-energy particle experiments is finding and fitting the charged-particlemore » tracks during event reconstruction. This is expected to become by far the dominant problem at the High-Luminosity Large Hadron Collider (HL-LHC), for example. Today the most common track finding methods are those based on the Kalman filter. Experience with Kalman techniques on real tracking detector systems has shown that they are robust and provide high physics performance. This is why they are currently in use at the LHC, both in the trigger and offine. Previously we reported on the significant parallel speedups that resulted from our investigations to adapt Kalman filters to track fitting and track building on Intel Xeon and Xeon Phi. Here, we discuss our progresses toward the understanding of these processors and the new developments to port the Kalman filter to NVIDIA GPUs.« less
NASA Astrophysics Data System (ADS)
Jollivet, C.; Farley, K.; Conroy, M.; Abramczyk, J.; Belke, S.; Becker, F.; Tankala, K.
2016-03-01
Single-mode (SM) kW-class fiber lasers are the tools of choice for material processing applications such as sheet metal cutting and welding. However, application requirements include a flat-top intensity profile and specific beam parameter product (BPP). Here, Nufern introduces a novel specialty fiber technology capable of converting a SM laser beam into a flat-top beam suited for these applications. The performances are demonstrated using a specialty fiber with 100 μm pure silica core, 0.22 NA surrounded by a 120 μm fluorine-doped layer and a 360 μm pure silica cladding, which was designed to match the conventional beam delivery fibers. A SM fiber laser operating at a wavelength of 1.07 μm and terminated with a large-mode area (LMA) fiber with 20 μm core and 0.06 NA was directly coupled in the core of the flat-top specialty fiber using conventional splicing technique. The output beam profile and BPP were characterized first with a low-power source and confirmed using a 2 kW laser and we report a beam transformation from a SM beam into a flat-top intensity profile beam with a 3.8 mm*mrad BPP. This is, to the best of our knowledge, the first successful beam transformation from SM to MM flat-top with controlled BPP in a single fiber integrated in a multi-kW all-fiber system architecture.
Parallelized Kalman-Filter-Based Reconstruction of Particle Tracks on Many-Core Processors and GPUs
NASA Astrophysics Data System (ADS)
Cerati, Giuseppe; Elmer, Peter; Krutelyov, Slava; Lantz, Steven; Lefebvre, Matthieu; Masciovecchio, Mario; McDermott, Kevin; Riley, Daniel; Tadel, Matevž; Wittich, Peter; Würthwein, Frank; Yagil, Avi
2017-08-01
For over a decade now, physical and energy constraints have limited clock speed improvements in commodity microprocessors. Instead, chipmakers have been pushed into producing lower-power, multi-core processors such as Graphical Processing Units (GPU), ARM CPUs, and Intel MICs. Broad-based efforts from manufacturers and developers have been devoted to making these processors user-friendly enough to perform general computations. However, extracting performance from a larger number of cores, as well as specialized vector or SIMD units, requires special care in algorithm design and code optimization. One of the most computationally challenging problems in high-energy particle experiments is finding and fitting the charged-particle tracks during event reconstruction. This is expected to become by far the dominant problem at the High-Luminosity Large Hadron Collider (HL-LHC), for example. Today the most common track finding methods are those based on the Kalman filter. Experience with Kalman techniques on real tracking detector systems has shown that they are robust and provide high physics performance. This is why they are currently in use at the LHC, both in the trigger and offine. Previously we reported on the significant parallel speedups that resulted from our investigations to adapt Kalman filters to track fitting and track building on Intel Xeon and Xeon Phi. Here, we discuss our progresses toward the understanding of these processors and the new developments to port the Kalman filter to NVIDIA GPUs.
Liu, Xing; Hou, Kun Mean; de Vaulx, Christophe; Xu, Jun; Yang, Jianfeng; Zhou, Haiying; Shi, Hongling; Zhou, Peng
2015-01-01
Memory and energy optimization strategies are essential for the resource-constrained wireless sensor network (WSN) nodes. In this article, a new memory-optimized and energy-optimized multithreaded WSN operating system (OS) LiveOS is designed and implemented. Memory cost of LiveOS is optimized by using the stack-shifting hybrid scheduling approach. Different from the traditional multithreaded OS in which thread stacks are allocated statically by the pre-reservation, thread stacks in LiveOS are allocated dynamically by using the stack-shifting technique. As a result, memory waste problems caused by the static pre-reservation can be avoided. In addition to the stack-shifting dynamic allocation approach, the hybrid scheduling mechanism which can decrease both the thread scheduling overhead and the thread stack number is also implemented in LiveOS. With these mechanisms, the stack memory cost of LiveOS can be reduced more than 50% if compared to that of a traditional multithreaded OS. Not is memory cost optimized, but also the energy cost is optimized in LiveOS, and this is achieved by using the multi-core “context aware” and multi-core “power-off/wakeup” energy conservation approaches. By using these approaches, energy cost of LiveOS can be reduced more than 30% when compared to the single-core WSN system. Memory and energy optimization strategies in LiveOS not only prolong the lifetime of WSN nodes, but also make the multithreaded OS feasible to run on the memory-constrained WSN nodes. PMID:25545264
A direction detective asymmetrical twin-core fiber curving sensor
NASA Astrophysics Data System (ADS)
An, Maowei; Geng, Tao; Yang, Wenlei; Zeng, Hongyi; Li, Jian
2015-10-01
Long period fiber gratings (LPFGs), which can couple the core mode to the forward propagating cladding modes of a fiber and have the advantage of small additional loss, no backward reflection, small size, which is widely used in optical fiber sensors and optical communication systems. LPFG has different fabricating methods, in order to write gratings on the twin-core at the same time effectively, we specially choose electric heating fused taper system to fabricate asymmetric dual-core long period fiber grating, because this kind of method can guarantee the similarity of gratings on the twin cores and obtain good geometric parameters of LPFG, such as cycle, cone waist. Then we use bending test platform to conduct bending test for each of the core of twin-core asymmetric long period fiber grating. Experiments show that: the sensitivity of asymmetrical twin-core long period fiber grating's central core under bending is -5.47nm·m, while the sensitivity of asymmetric twin-core long period fiber grating partial core changed with the relative position of screw micrometer. The sensitivity at 0°, 30°, 90° direction is -4.22nm·m, -9.84nm·m, -11.44nm·m respectively. The experiment results strongly demonstrate the properties of rim sensing of asymmetrical twin-core fiber gratings which provides the possibility of simultaneously measuring the bending magnitude and direction and solving the problem of cross sensing when multi-parameter measuring. In other words, we can detect temperature and bend at the same time by this sensor. As our knowledge, it is the first time simultaneously measuring bend and temperature using this structure of fiber sensors.
Liu, Bo; Zhang, Lijia; Xin, Xiangjun
2018-03-19
This paper proposes and demonstrates an enhanced secure 4-D modulation optical generalized filter bank multi-carrier (GFBMC) system based on joint constellation and Stokes vector scrambling. The constellation and Stokes vectors are scrambled by using different scrambling parameters. A multi-scroll Chua's circuit map is adopted as the chaotic model. Large secure key space can be obtained due to the multi-scroll attractors and independent operability of subcarriers. A 40.32Gb/s encrypted optical GFBMC signal with 128 parallel subcarriers is successfully demonstrated in the experiment. The results show good resistance against the illegal receiver and indicate a potential way for the future optical multi-carrier system.
LOSITAN: a workbench to detect molecular adaptation based on a Fst-outlier method.
Antao, Tiago; Lopes, Ana; Lopes, Ricardo J; Beja-Pereira, Albano; Luikart, Gordon
2008-07-28
Testing for selection is becoming one of the most important steps in the analysis of multilocus population genetics data sets. Existing applications are difficult to use, leaving many non-trivial, error-prone tasks to the user. Here we present LOSITAN, a selection detection workbench based on a well evaluated Fst-outlier detection method. LOSITAN greatly facilitates correct approximation of model parameters (e.g., genome-wide average, neutral Fst), provides data import and export functions, iterative contour smoothing and generation of graphics in a easy to use graphical user interface. LOSITAN is able to use modern multi-core processor architectures by locally parallelizing fdist, reducing computation time by half in current dual core machines and with almost linear performance gains in machines with more cores. LOSITAN makes selection detection feasible to a much wider range of users, even for large population genomic datasets, by both providing an easy to use interface and essential functionality to complete the whole selection detection process.
C3: A Command-line Catalogue Cross-matching tool for modern astrophysical survey data
NASA Astrophysics Data System (ADS)
Riccio, Giuseppe; Brescia, Massimo; Cavuoti, Stefano; Mercurio, Amata; di Giorgio, Anna Maria; Molinari, Sergio
2017-06-01
In the current data-driven science era, it is needed that data analysis techniques has to quickly evolve to face with data whose dimensions has increased up to the Petabyte scale. In particular, being modern astrophysics based on multi-wavelength data organized into large catalogues, it is crucial that the astronomical catalog cross-matching methods, strongly dependant from the catalogues size, must ensure efficiency, reliability and scalability. Furthermore, multi-band data are archived and reduced in different ways, so that the resulting catalogues may differ each other in formats, resolution, data structure, etc, thus requiring the highest generality of cross-matching features. We present C 3 (Command-line Catalogue Cross-match), a multi-platform application designed to efficiently cross-match massive catalogues from modern surveys. Conceived as a stand-alone command-line process or a module within generic data reduction/analysis pipeline, it provides the maximum flexibility, in terms of portability, configuration, coordinates and cross-matching types, ensuring high performance capabilities by using a multi-core parallel processing paradigm and a sky partitioning algorithm.
Rapid Calculation of Max-Min Fair Rates for Multi-Commodity Flows in Fat-Tree Networks
Mollah, Md Atiqul; Yuan, Xin; Pakin, Scott; ...
2017-08-29
Max-min fairness is often used in the performance modeling of interconnection networks. Existing methods to compute max-min fair rates for multi-commodity flows have high complexity and are computationally infeasible for large networks. In this paper, we show that by considering topological features, this problem can be solved efficiently for the fat-tree topology that is widely used in data centers and high performance compute clusters. Several efficient new algorithms are developed for this problem, including a parallel algorithm that can take advantage of multi-core and shared-memory architectures. Using these algorithms, we demonstrate that it is possible to find the max-min fairmore » rate allocation for multi-commodity flows in fat-tree networks that support tens of thousands of nodes. We evaluate the run-time performance of the proposed algorithms and show improvement in orders of magnitude over the previously best known method. Finally, we further demonstrate a new application of max-min fair rate allocation that is only computationally feasible using our new algorithms.« less
Rapid Calculation of Max-Min Fair Rates for Multi-Commodity Flows in Fat-Tree Networks
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mollah, Md Atiqul; Yuan, Xin; Pakin, Scott
Max-min fairness is often used in the performance modeling of interconnection networks. Existing methods to compute max-min fair rates for multi-commodity flows have high complexity and are computationally infeasible for large networks. In this paper, we show that by considering topological features, this problem can be solved efficiently for the fat-tree topology that is widely used in data centers and high performance compute clusters. Several efficient new algorithms are developed for this problem, including a parallel algorithm that can take advantage of multi-core and shared-memory architectures. Using these algorithms, we demonstrate that it is possible to find the max-min fairmore » rate allocation for multi-commodity flows in fat-tree networks that support tens of thousands of nodes. We evaluate the run-time performance of the proposed algorithms and show improvement in orders of magnitude over the previously best known method. Finally, we further demonstrate a new application of max-min fair rate allocation that is only computationally feasible using our new algorithms.« less
Shape sensing using multi-core fiber optic cable and parametric curve solutions.
Moore, Jason P; Rogge, Matthew D
2012-01-30
The shape of a multi-core optical fiber is calculated by numerically solving a set of Frenet-Serret equations describing the path of the fiber in three dimensions. Included in the Frenet-Serret equations are curvature and bending direction functions derived from distributed fiber Bragg grating strain measurements in each core. The method offers advantages over prior art in that it determines complex three-dimensional fiber shape as a continuous parametric solution rather than an integrated series of discrete planar bends. Results and error analysis of the method using a tri-core optical fiber is presented. Maximum error expressed as a percentage of fiber length was found to be 7.2%.
He, Jiale; Li, Borui; Deng, Lei; Tang, Ming; Gan, Lin; Fu, Songnian; Shum, Perry Ping; Liu, Deming
2016-06-13
In this paper, the feasibility of space division multiplexing for optical wireless fronthaul systems is experimentally demonstrated by implementing high speed MIMO-OFDM/OQAM radio signals over 20km 7-core fiber and 0.4m wireless link. Moreover, the impact of optical inter-core crosstalk in multicore fibers on the proposed MIMO-OFDM/OQAM radio over fiber system is experimentally evaluated in both SISO and MIMO configurations for comparison. The experimental results show that the inter-core crosstalk tolerance of the proposed radio over fiber system can be relaxed to -10 dB by using the proposed MIMO-OFDM/OQAM processing. These results could guide high density multicore fiber design to support a large number of antenna modules and a higher density of radio-access points for potential applications in 5G cellular system.
NASA Astrophysics Data System (ADS)
Feist, S.; Maclachlan, J. C.; Reinhardt, E. G.; McNeill-Jewer, C.; Eyles, C.
2016-12-01
Hidden River Cave is part of a cave system hydrogeologically related to Mammoth Cave in Kentucky and is a multi-level active cave system with 25km of mapped passages. Upper levels experience flow during flood events and lower levels have continuously flowing water. Improper industrial and domestic waste disposal and poor understanding of local hydrogeology lead to contamination of Hidden River Cave in the early 1940s. Previously used for hydroelectric power generation and as a source of potable water the cave was closed to the public for almost 50 years. A new sewage treatment plant and remediation efforts since 1989 have improved the cave system's health. This project focuses on sedimentological studies in the Hidden River Cave system. Water and sediment transport in the cave are being investigated using sediment cores, surface sediment samples and water level data. An Itrax core scanner is used to analyze sediment cores for elemental concentrations, magnetic susceptibility, radiography, and high resolution photography. Horizons of metal concentrations in the core allow correlation of sedimentation events in the cave system. Thecamoebian (testate amoebae) microfossils identified in surface samples allow for further constraint of sediment sources, sedimentation rates, and paleoclimatic analysis. Dive recorders monitor water levels, providing data to further understand the movement of sediment through the cave system. A general time constraint on the sediment's age is based on the presence of microplastic in the surface samples and sediment cores, and data from radiocarbon and lead-210 dating. The integration of various sedimentological data allows for better understanding of sedimentation processes and their record of paleoenvironmental change in the cave system. Sediment studies and methodologies from this project can be applied to other karst systems, and have important applications for communities living on karst landscapes and their water management policies.
Hemmi, Akihide; Mizumura, Ryosuke; Kawanishi, Ryuta; Nakajima, Hizuru; Zeng, Hulie; Uchiyama, Katsumi; Kaneki, Noriaki; Imato, Toshihiko
2013-01-01
A novel two dimensional surface plasmon resonance (SPR) sensor system with a multi-point sensing region is described. The use of multiplied beam splitting optics, as a core technology, permitted multi-point sensing to be achieved. This system was capable of simultaneously measuring nine sensing points. Calibration curves for sucrose obtained on nine sensing points were linear in the range of 0–10% with a correlation factor of 0.996–0.998 with a relative standard deviation of 0.090–4.0%. The detection limits defined as S/N = 3 were 1.98 × 10−6–3.91 × 10−5 RIU. This sensitivity is comparable to that of conventional SPR sensors. PMID:23299626
Reversible Parallel Discrete-Event Execution of Large-scale Epidemic Outbreak Models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Perumalla, Kalyan S; Seal, Sudip K
2010-01-01
The spatial scale, runtime speed and behavioral detail of epidemic outbreak simulations together require the use of large-scale parallel processing. In this paper, an optimistic parallel discrete event execution of a reaction-diffusion simulation model of epidemic outbreaks is presented, with an implementation over themore » $$\\mu$$sik simulator. Rollback support is achieved with the development of a novel reversible model that combines reverse computation with a small amount of incremental state saving. Parallel speedup and other runtime performance metrics of the simulation are tested on a small (8,192-core) Blue Gene / P system, while scalability is demonstrated on 65,536 cores of a large Cray XT5 system. Scenarios representing large population sizes (up to several hundred million individuals in the largest case) are exercised.« less
NASA Astrophysics Data System (ADS)
Lin, Shian-Jiann; Harris, Lucas; Chen, Jan-Huey; Zhao, Ming
2014-05-01
A multi-scale High-Resolution Atmosphere Model (HiRAM) is being developed at NOAA/Geophysical Fluid Dynamics Laboratory. The model's dynamical framework is the non-hydrostatic extension of the vertically Lagrangian finite-volume dynamical core (Lin 2004, Monthly Wea. Rev.) constructed on a stretchable (via Schmidt transformation) cubed-sphere grid. Physical parametrizations originally designed for IPCC-type climate predictions are in the process of being modified and made more "scale-aware", in an effort to make the model suitable for multi-scale weather-climate applications, with horizontal resolution ranging from 1 km (near the target high-resolution region) to as low as 400 km (near the antipodal point). One of the main goals of this development is to enable simulation of high impact weather phenomena (such as tornadoes, thunderstorms, category-5 hurricanes) within an IPCC-class climate modeling system previously thought impossible. We will present preliminary results, covering a very wide spectrum of temporal-spatial scales, ranging from simulation of tornado genesis (hours), Madden-Julian Oscillations (intra-seasonal), topical cyclones (seasonal), to Quasi Biennial Oscillations (intra-decadal), using the same global multi-scale modeling system.
Large-scale and Long-duration Simulation of a Multi-stage Eruptive Solar Event
NASA Astrophysics Data System (ADS)
Jiang, chaowei; Hu, Qiang; Wu, S. T.
2015-04-01
We employ a data-driven 3D MHD active region evolution model by using the Conservation Element and Solution Element (CESE) numerical method. This newly developed model retains the full MHD effects, allowing time-dependent boundary conditions and time evolution studies. The time-dependent simulation is driven by measured vector magnetograms and the method of MHD characteristics on the bottom boundary. We have applied the model to investigate the coronal magnetic field evolution of AR11283 which was characterized by a pre-existing sigmoid structure in the core region and multiple eruptions, both in relatively small and large scales. We have succeeded in producing the core magnetic field structure and the subsequent eruptions of flux-rope structures (see https://dl.dropboxusercontent.com/u/96898685/large.mp4 for an animation) as the measured vector magnetograms on the bottom boundary evolve in time with constant flux emergence. The whole process, lasting for about an hour in real time, compares well with the corresponding SDO/AIA and coronagraph imaging observations. From these results, we show the capability of the model, largely data-driven, that is able to simulate complex, topological, and highly dynamic active region evolutions. (We acknowledge partial support of NSF grants AGS 1153323 and AGS 1062050, and data support from SDO/HMI and AIA teams).
Quantum Monte Carlo for electronic structure: Recent developments and applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rodriquez, Maria Milagos Soto
Quantum Monte Carlo (QMC) methods have been found to give excellent results when applied to chemical systems. The main goal of the present work is to use QMC to perform electronic structure calculations. In QMC, a Monte Carlo simulation is used to solve the Schroedinger equation, taking advantage of its analogy to a classical diffusion process with branching. In the present work the author focuses on how to extend the usefulness of QMC to more meaningful molecular systems. This study is aimed at questions concerning polyatomic and large atomic number systems. The accuracy of the solution obtained is determined bymore » the accuracy of the trial wave function`s nodal structure. Efforts in the group have given great emphasis to finding optimized wave functions for the QMC calculations. Little work had been done by systematically looking at a family of systems to see how the best wave functions evolve with system size. In this work the author presents a study of trial wave functions for C, CH, C 2H and C 2H 2. The goal is to study how to build wave functions for larger systems by accumulating knowledge from the wave functions of its fragments as well as gaining some knowledge on the usefulness of multi-reference wave functions. In a MC calculation of a heavy atom, for reasonable time steps most moves for core electrons are rejected. For this reason true equilibration is rarely achieved. A method proposed by Batrouni and Reynolds modifies the way the simulation is performed without altering the final steady-state solution. It introduces an acceleration matrix chosen so that all coordinates (i.e., of core and valence electrons) propagate at comparable speeds. A study of the results obtained using their proposed matrix suggests that it may not be the optimum choice. In this work the author has found that the desired mixing of coordinates between core and valence electrons is not achieved when using this matrix. A bibliography of 175 references is included.« less
Compact objects at the heart of outflows in large and small systems
NASA Astrophysics Data System (ADS)
Sell, Paul Harrison
2013-12-01
This thesis focuses on studying and assessing high-energy feedback generated by both stellar mass and supermassive compact objects. From these two perspectives, I help bridge the gap in understanding how jets and winds can transform their much larger environments in thousands to millions of years, astronomically short timescales. I have acquired X-ray and optical data that aim to elucidate the role these objects play in powering parsec-scale shockwaves in the ISM and in driving kiloparsec-scale outflows in galaxies. I present Chandra X-ray imaging, Hubble Space Telescope imaging, and WIYN Hydra multi-object optical spectroscopic observations. The data reveal the morphologies of the systems and constrain on a range of interesting parameters: power, outflow velocity, density, accretion efficiency, and timescale. My analysis provides perspective on the importance of black holes, both large and small, and neutron stars for driving outflows into the interstellar and intergalactic medium. On kiloparsec scales, I explore the nature of what appear to be merging or recently merging post-starburst galaxies with very high-velocity winds. This work is part of a multiwavelength effort to characterize the niche these galaxies fill in the larger scheme of galaxy evolution. My focus is on the accretion activity of the coalescing supermassive black holes in their cores. This work leads us to compare the relative importance of a massive starburst to the supermassive black holes in the cores of the galaxies. On parsec scales, I present case studies of two prominent microquasars, Galactic X-ray binaries with jets, Circinus X-1 and Cygnus X-1. In the case of Circinus X-1, I present very deep follow-up observations of parsec-scale shock plumes driven by a powerful, bipolar jet. In the case of Cygnus X-1, I present follow-up observations to probe a recently discovered outflow near the binary. I calculate robust, physically motivated limits on the total power needed to drive the outflows in both of these systems.
NASA Astrophysics Data System (ADS)
Rhodes, Rachael H.; Faïn, Xavier; Stowasser, Christopher; Blunier, Thomas; Chappellaz, Jérôme; McConnell, Joseph R.; Romanini, Daniele; Mitchell, Logan E.; Brook, Edward J.
2013-04-01
Ancient air trapped inside bubbles in ice cores can now be analysed for methane concentration utilising a laser spectrometer coupled to a continuous melter system. We present a new ultra-high resolution record of atmospheric methane variability over the last 1800 yr obtained from continuous analysis of a shallow ice core from the North Greenland Eemian project (NEEM-2011-S1) during a 4-week laboratory-based measurement campaign. Our record faithfully replicates the form and amplitudes of multi-decadal oscillations previously observed in other ice cores and demonstrates the detailed depth resolution (5.3 cm), rapid acquisition time (30 m day-1) and good long-term reproducibility (2.6%, 2σ) of the continuous measurement technique. In addition, we report the detection of high frequency ice core methane signals of non-atmospheric origin. Firstly, measurements of air from the firn-ice transition region and an interval of ice core dating from 1546-1560 AD (gas age) resolve apparently quasi-annual scale methane oscillations. Traditional gas chromatography measurements on discrete ice samples confirm these signals and indicate peak-to-peak amplitudes of ca. 22 parts per billion (ppb). We hypothesise that these oscillations result from staggered bubble close-off between seasonal layers of contrasting density during time periods of sustained multi-year atmospheric methane change. Secondly, we report the detection of abrupt (20-100 cm depth interval), high amplitude (35-80 ppb excess) methane spikes in the NEEM ice that are reproduced by discrete measurements. We show for the first time that methane spikes present in thin and infrequent layers in polar, glacial ice are accompanied by elevated concentrations of carbon- and nitrogen-based chemical impurities, and suggest that biological in-situ production may be responsible.
Extensive Core Microbiome in Drone-Captured Whale Blow Supports a Framework for Health Monitoring
Miller, Carolyn A.; Moore, Michael J.; Durban, John W.; Fearnbach, Holly; Barrett-Lennard, Lance G.
2017-01-01
ABSTRACT The pulmonary system is a common site for bacterial infections in cetaceans, but very little is known about their respiratory microbiome. We used a small, unmanned hexacopter to collect exhaled breath condensate (blow) from two geographically distinct populations of apparently healthy humpback whales (Megaptera novaeangliae), sampled in the Massachusetts coastal waters off Cape Cod (n = 17) and coastal waters around Vancouver Island (n = 9). Bacterial and archaeal small-subunit rRNA genes were amplified and sequenced from blow samples, including many of sparse volume, as well as seawater and other controls, to characterize the associated microbial community. The blow microbiomes were distinct from the seawater microbiomes and included 25 phylogenetically diverse bacteria common to all sampled whales. This core assemblage comprised on average 36% of the microbiome, making it one of the more consistent animal microbiomes studied to date. The closest phylogenetic relatives of 20 of these core microbes were previously detected in marine mammals, suggesting that this core microbiome assemblage is specialized for marine mammals and may indicate a healthy, noninfected pulmonary system. Pathogen screening was conducted on the microbiomes at the genus level, which showed that all blow and few seawater microbiomes contained relatives of bacterial pathogens; no known cetacean respiratory pathogens were detected in the blow. Overall, the discovery of a shared large core microbiome in humpback whales is an important advancement for health and disease monitoring of this species and of other large whales. IMPORTANCE The conservation and management of large whales rely in part upon health monitoring of individuals and populations, and methods generally necessitate invasive sampling. Here, we used a small, unmanned hexacopter drone to noninvasively fly above humpback whales from two populations, capture their exhaled breath (blow), and examine the associated microbiome. In the first extensive examination of the large-whale blow microbiome, we present surprising results about the discovery of a large core microbiome that was shared across individual whales from geographically separated populations in two ocean basins. We suggest that this core microbiome, in addition to other microbiome characteristics, could be a useful feature for health monitoring of large whales worldwide. PMID:29034331
Extensive Core Microbiome in Drone-Captured Whale Blow Supports a Framework for Health Monitoring.
Apprill, Amy; Miller, Carolyn A; Moore, Michael J; Durban, John W; Fearnbach, Holly; Barrett-Lennard, Lance G
2017-01-01
The pulmonary system is a common site for bacterial infections in cetaceans, but very little is known about their respiratory microbiome. We used a small, unmanned hexacopter to collect exhaled breath condensate (blow) from two geographically distinct populations of apparently healthy humpback whales ( Megaptera novaeangliae ), sampled in the Massachusetts coastal waters off Cape Cod ( n = 17) and coastal waters around Vancouver Island ( n = 9). Bacterial and archaeal small-subunit rRNA genes were amplified and sequenced from blow samples, including many of sparse volume, as well as seawater and other controls, to characterize the associated microbial community. The blow microbiomes were distinct from the seawater microbiomes and included 25 phylogenetically diverse bacteria common to all sampled whales. This core assemblage comprised on average 36% of the microbiome, making it one of the more consistent animal microbiomes studied to date. The closest phylogenetic relatives of 20 of these core microbes were previously detected in marine mammals, suggesting that this core microbiome assemblage is specialized for marine mammals and may indicate a healthy, noninfected pulmonary system. Pathogen screening was conducted on the microbiomes at the genus level, which showed that all blow and few seawater microbiomes contained relatives of bacterial pathogens; no known cetacean respiratory pathogens were detected in the blow. Overall, the discovery of a shared large core microbiome in humpback whales is an important advancement for health and disease monitoring of this species and of other large whales. IMPORTANCE The conservation and management of large whales rely in part upon health monitoring of individuals and populations, and methods generally necessitate invasive sampling. Here, we used a small, unmanned hexacopter drone to noninvasively fly above humpback whales from two populations, capture their exhaled breath (blow), and examine the associated microbiome. In the first extensive examination of the large-whale blow microbiome, we present surprising results about the discovery of a large core microbiome that was shared across individual whales from geographically separated populations in two ocean basins. We suggest that this core microbiome, in addition to other microbiome characteristics, could be a useful feature for health monitoring of large whales worldwide.
NASA Astrophysics Data System (ADS)
Stewart, J.; Tolliver, R.; Field, D. B.; Young, C.; Stafford, G.; Day, R. D.
2016-12-01
Monitoring of the physiological/ecological response of marine calcifying organisms to the combination of lower pH and toxic metal pollutants (e.g. Cu and Sn from boat anti-fouling paints) into the oceans requires detailed knowledge of the rates and spatial distribution of ocean acidification (OA) and trace metal composition over time. Yet, measurement of metal concentrations and carbonate system parameters in the modern ocean from seawater bottle data is patchy (e.g. CDIAC/WOCE Carbon Data; http://cdiac.ornl.gov) and there remain few long-term surface water pH monitoring stations; the two longest continuous records of ocean pH extend back less than 30 years (Bermuda - BATS, 31°40'N, 64°10'W; Hawaii - HOTs, 22°45'N, 158°00'W). Much attention has therefore been focused on trace metal and ocean carbonate system proxy development to allow reconstruction of seawater metal content and pH in the past. Of particular promise is the boron isotope (δ11B) pH-proxy measured in marine calcifying organisms such as coral that can be cored enabling multi-decadal, annual-resolution, records of trace element incorporation and seawater pH to be generated. Here we present continuous Cu/Ca and Sn/Ca records in addition to δ11B data from three coral cores of Porites lutea. collected from waters proximal to Oahu, Hawaii. The diagenetic integrity of samples is verified using X-ray diffraction to assess the degree of calcite replacement. These cores reach a maximum depth of 80 cm and represent approximately 80 years of coral growth and seawater chemistry.
Bender, P.; Bogart, L. K.; Posth, O.; Szczerba, W.; Rogers, S. E.; Castro, A.; Nilsson, L.; Zeng, L. J.; Sugunan, A.; Sommertune, J.; Fornara, A.; González-Alonso, D.; Barquín, L. Fernández; Johansson, C.
2017-01-01
The structural and magnetic properties of magnetic multi-core particles were determined by numerical inversion of small angle scattering and isothermal magnetisation data. The investigated particles consist of iron oxide nanoparticle cores (9 nm) embedded in poly(styrene) spheres (160 nm). A thorough physical characterisation of the particles included transmission electron microscopy, X-ray diffraction and asymmetrical flow field-flow fractionation. Their structure was ultimately disclosed by an indirect Fourier transform of static light scattering, small angle X-ray scattering and small angle neutron scattering data of the colloidal dispersion. The extracted pair distance distribution functions clearly indicated that the cores were mostly accumulated in the outer surface layers of the poly(styrene) spheres. To investigate the magnetic properties, the isothermal magnetisation curves of the multi-core particles (immobilised and dispersed in water) were analysed. The study stands out by applying the same numerical approach to extract the apparent moment distributions of the particles as for the indirect Fourier transform. It could be shown that the main peak of the apparent moment distributions correlated to the expected intrinsic moment distribution of the cores. Additional peaks were observed which signaled deviations of the isothermal magnetisation behavior from the non-interacting case, indicating weak dipolar interactions. PMID:28397851
Hadwiger, M; Beyer, J; Jeong, Won-Ki; Pfister, H
2012-12-01
This paper presents the first volume visualization system that scales to petascale volumes imaged as a continuous stream of high-resolution electron microscopy images. Our architecture scales to dense, anisotropic petascale volumes because it: (1) decouples construction of the 3D multi-resolution representation required for visualization from data acquisition, and (2) decouples sample access time during ray-casting from the size of the multi-resolution hierarchy. Our system is designed around a scalable multi-resolution virtual memory architecture that handles missing data naturally, does not pre-compute any 3D multi-resolution representation such as an octree, and can accept a constant stream of 2D image tiles from the microscopes. A novelty of our system design is that it is visualization-driven: we restrict most computations to the visible volume data. Leveraging the virtual memory architecture, missing data are detected during volume ray-casting as cache misses, which are propagated backwards for on-demand out-of-core processing. 3D blocks of volume data are only constructed from 2D microscope image tiles when they have actually been accessed during ray-casting. We extensively evaluate our system design choices with respect to scalability and performance, compare to previous best-of-breed systems, and illustrate the effectiveness of our system for real microscopy data from neuroscience.
NASA Technical Reports Server (NTRS)
Rogge, Matthew D. (Inventor); Moore, Jason P. (Inventor)
2014-01-01
Shape of a multi-core optical fiber is determined by positioning the fiber in an arbitrary initial shape and measuring strain over the fiber's length using strain sensors. A three-coordinate p-vector is defined for each core as a function of the distance of the corresponding cores from a center point of the fiber and a bending angle of the cores. The method includes calculating, via a controller, an applied strain value of the fiber using the p-vector and the measured strain for each core, and calculating strain due to bending as a function of the measured and the applied strain values. Additionally, an apparent local curvature vector is defined for each core as a function of the calculated strain due to bending. Curvature and bend direction are calculated using the apparent local curvature vector, and fiber shape is determined via the controller using the calculated curvature and bend direction.
Life Cycle of Midlatitude Deep Convective Systems in a Lagrangian Framework
NASA Technical Reports Server (NTRS)
Feng, Zhe; Dong, Xiquan; Xie, Baike; McFarlane, Sally A.; Kennedy, Aaron; Lin, Bing; Minnis, Patrick
2012-01-01
Deep Convective Systems (DCSs) consist of intense convective cores (CC), large stratiform rain (SR) regions, and extensive non-precipitating anvil clouds (AC). This study focuses on the evolution of these three components and the factors that affect convective AC production. An automated satellite tracking method is used in conjunction with a recently developed multi-sensor hybrid classification to analyze the evolution of DCS structure in a Lagrangian framework over the central United States. Composite analysis from 4221 tracked DCSs during two warm seasons (May-August, 2010-2011) shows that maximum system size correlates with lifetime, and longer-lived DCSs have more extensive SR and AC. Maximum SR and AC area lag behind peak convective intensity and the lag increases linearly from approximately 1-hour for short-lived systems to more than 3-hours for long-lived ones. The increased lag, which depends on the convective environment, suggests that changes in the overall diabatic heating structure associated with the transition from CC to SR and AC could prolong the system lifetime by sustaining stratiform cloud development. Longer-lasting systems are associated with up to 60% higher mid-tropospheric relative humidity and up to 40% stronger middle to upper tropospheric wind shear. Regression analysis shows that the areal coverage of thick AC is strongly correlated with the size of CC, updraft strength, and SR area. Ambient upper tropospheric wind speed and wind shear also play an important role for convective AC production where for systems with large AC (radius greater than 120-km) they are 24% and 20% higher, respectively, than those with small AC (radius=20 km).
A Lattice Boltzmann Framework for the simulation of boiling hydrodynamics in BWRs.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jain, P. K.; Tentner, A.; Uddin, R.
2008-01-01
Multi phase and multi component flows are ubiquitous in nature as well as in many man-made processes. A specific example is the Boiling Water Reactor (BWR) core, in which the coolant enters the core as liquid, undergoes a phase change as it traverses the core and exits as a high quality two-phase mixture. Two-phase flows in BWRs typically manifest a wide variety of geometrical patterns of the co-existing phases depending on the local system conditions. Modeling of such flows currently relies on empirical correlations (for example, in the simulation of bubble nucleation, bubble growth and coalescence, and inter-phase surface topologymore » transitions) that hinder the accurate simulation of two-phase phenomena using Computational Fluid Dynamics (CFD) approaches. The Lattice Boltzmann Method (LBM) is in rapid development as a modeling tool to understand these macro-phenomena by coupling them with their underlying micro-dynamics. This paper presents a consistent LBM formulation for the simulation of a two-phase water-steam system. Results of initial model validation in a range of thermodynamic conditions typical for BWRs are also shown. The interface between the two coexisting phases is captured from the dynamics of the model itself, i.e., no interface tracking is needed. The model is based on the Peng-Robinson (P-R) non-ideal equation of state and can quantitatively approximate the phase-coexistence curve for water at different temperatures ranging from 125 to 325 oC. Consequently, coexisting phases with large density ratios (up to {approx}1000) may be simulated. Two-phase models in the 200-300 C temperature range are of significant importance to nuclear engineers since most BWRs operate under similar thermodynamic conditions. Simulation of bubbles and droplets in a gravity-free environment of the corresponding coexisting phase until steady state is reached satisfies Laplace law at different temperatures and thus, yield the surface tension of the fluid. Comparing the LBM surface tension thus calculated using the LBM to the corresponding experimental values for water, the LBM lattice unit (lu) can be scaled to the physical units. Using this approach, spatial scaling of the LBM emerges from the model itself and is not imposed externally.« less
A New Generation of Real-Time Systems in the JET Tokamak
NASA Astrophysics Data System (ADS)
Alves, Diogo; Neto, Andre C.; Valcarcel, Daniel F.; Felton, Robert; Lopez, Juan M.; Barbalace, Antonio; Boncagni, Luca; Card, Peter; De Tommasi, Gianmaria; Goodyear, Alex; Jachmich, Stefan; Lomas, Peter J.; Maviglia, Francesco; McCullen, Paul; Murari, Andrea; Rainford, Mark; Reux, Cedric; Rimini, Fernanda; Sartori, Filippo; Stephen, Adam V.; Vega, Jesus; Vitelli, Riccardo; Zabeo, Luca; Zastrow, Klaus-Dieter
2014-04-01
Recently, a new recipe for developing and deploying real-time systems has become increasingly adopted in the JET tokamak. Powered by the advent of x86 multi-core technology and the reliability of JET's well established Real-Time Data Network (RTDN) to handle all real-time I/O, an official Linux vanilla kernel has been demonstrated to be able to provide real-time performance to user-space applications that are required to meet stringent timing constraints. In particular, a careful rearrangement of the Interrupt ReQuests' (IRQs) affinities together with the kernel's CPU isolation mechanism allows one to obtain either soft or hard real-time behavior depending on the synchronization mechanism adopted. Finally, the Multithreaded Application Real-Time executor (MARTe) framework is used for building applications particularly optimised for exploring multi-core architectures. In the past year, four new systems based on this philosophy have been installed and are now part of JET's routine operation. The focus of the present work is on the configuration aspects that enable these new systems' real-time capability. Details are given about the common real-time configuration of these systems, followed by a brief description of each system together with results regarding their real-time performance. A cycle time jitter analysis of a user-space MARTe based application synchronizing over a network is also presented. The goal is to compare its deterministic performance while running on a vanilla and on a Messaging Real time Grid (MRG) Linux kernel.
2008-07-01
generation of process partitioning, a thread pipelining becomes possible. In this paper we briefly summarize the requirements and trends for FADEC based... FADEC environment, presenting a hypothetical realization of an example application. Finally we discuss the application of Time-Triggered...based control applications of the future. 15. SUBJECT TERMS Gas turbine, FADEC , Multi-core processing technology, disturbed based control
How temperature determines formation of maghemite nanoparticles
NASA Astrophysics Data System (ADS)
Girod, Matthias; Vogel, Stefanie; Szczerba, Wojciech; Thünemann, Andreas F.
2015-04-01
We report on the formation of polymer-stabilized superparamagnetic single-core and multi-core maghemite nanoparticles. The particle formation was carried out by coprecipitation of Fe(II) and Fe(III) sulfate in a continuous aqueous process using a micromixer system. Aggregates containing 50 primary particles with sizes of 2 nm were formed at a reaction temperature of 30 °C. These particles aggregated further with time and were not stable. In contrast, stable single-core particles with a diameter of 7 nm were formed at 80 °C as revealed by small-angle X-ray scattering (SAXS) coupled in-line with the micromixer for particle characterization. X-ray diffraction and TEM confirmed the SAXS results. X-ray absorption near-edge structure spectroscopy (XANES) identified the iron oxide phase as maghemite.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Luszczek, Piotr R; Tomov, Stanimire Z; Dongarra, Jack J
We present an efficient and scalable programming model for the development of linear algebra in heterogeneous multi-coprocessor environments. The model incorporates some of the current best design and implementation practices for the heterogeneous acceleration of dense linear algebra (DLA). Examples are given as the basis for solving linear systems' algorithms - the LU, QR, and Cholesky factorizations. To generate the extreme level of parallelism needed for the efficient use of coprocessors, algorithms of interest are redesigned and then split into well-chosen computational tasks. The tasks execution is scheduled over the computational components of a hybrid system of multi-core CPUs andmore » coprocessors using a light-weight runtime system. The use of lightweight runtime systems keeps scheduling overhead low, while enabling the expression of parallelism through otherwise sequential code. This simplifies the development efforts and allows the exploration of the unique strengths of the various hardware components.« less
NASA Astrophysics Data System (ADS)
Rosinski, A.; Morentz, J.; Beilin, P.
2017-12-01
The principal function of the California Earthquake Clearinghouse is to provide State and Federal disaster response managers, and the scientific and engineering communities, with prompt information on ground failure, structural damage, and other consequences from significant seismic events such as earthquakes and tsunamis. The overarching problem highlighted in discussions with Clearinghouse partners is the confusion and frustration of many of the Operational Area representatives, and some regional utilities throughout the state on what software applications they should be using and maintaining to meet State, Federal, and Local, requirements, and for what purposes, and how to deal with the limitations of these applications. This problem is getting in the way of making meaningful progress on developing multi-application interoperability and the necessary supporting cross-sector information-sharing procedures and dialogue on essential common operational information that entities need to share for different all hazards missions and related operational activities associated with continuity, security, and resilience. The XchangeCore based system the Clearinghouse is evolving helps deal with this problem, and does not compound it by introducing yet another end-user application; there is no end-user interface with which one views XchangeCore, all viewing of data provided through XchangeCore occurs in and on existing, third-party operational applications. The Clearinghouse efforts with XchangeCore are compatible with FEMA, which is currently using XchangeCore-provided data for regional and National Business Emergency Operations Center (source of business information sharing during emergencies) response. Also important, and should be emphasized, is that information-sharing is not just for response, but for preparedness, risk assessment/mitigation decision-making, and everyday operational needs for situational awareness. In other words, the benefits of the Clearinghouse information sharing efforts transcend emergency response. The Clearinghouse is in the process of developing an Information-Sharing System Guide and CONOPS/ templates, that should be aimed a multi-stakeholder, non-technical audience.
Siderophile Element Partitioning between Sulfide- and Silicate melts.
NASA Astrophysics Data System (ADS)
Hackler, S.; Rohrbach, A.; Loroch, D. C.; Klemme, S.; Berndt, J.
2017-12-01
Different theories concerning the formation of the Earth are debated. Either Earth accreted mostly `dry' or volatile elements were delivered late after core formation was largely inactive [1, 2], or volatile rich material was accreted during the main stages of accretion and core formation [3, 4, 5]. The partitioning behavior of siderophile volatile elements (SVE; S, Se, Te, Tl, Ag, Au, Cd, Bi, Pb, Sn, Cu, Ge, and In) may provide first order constraints whether these element concentrations in Earth's mantle were established before or after core-mantle differentiation or perhaps during both periods by multi stage core formation [6]. A special interest is laid into chalcophile element behavior with respective to the possible formation and segregation of a hadean matte [7]. To examine the influence of sulfur on SVE partitioning between metal-silicate melts, we performed experiments simulating a magma ocean stage evolving from sulfur poor- (low fO2) to more oxidizing sulfur rich- (Fe, Ni)-S melts ( 20 wt% S) towards the end of accretion. We carried out partitioning experiments under various P-T-fO2 conditions with a Bristol type end loaded piston cylinder apparatus (<3 GPa) and a 1000 t walker-type multi-anvil press (3-20 GPa). Our results will be presented at the meeting. References: [1] Albarède F. (2009) Nature, 461, 1227-1233. [2] Ballhaus C. et al. (2013) EPSL, 362, 237-245. [3] Fischer-Gödde M. and Kleine T. (2017) Nature, 541, 525 527. [4] Wade J. and Wood B. J. (2005) EPSL, 236, 78-95. [5] Rubie D. et al. (2016) Science, 253, 1141-1144. [6] Rubie D. et al. (2011) EPSL, 301, 31-42. [7] O'Neill H. St. C. (1991) GCA, 55, 1159-1172.
Photometric Calibration and Image Stitching for a Large Field of View Multi-Camera System
Lu, Yu; Wang, Keyi; Fan, Gongshu
2016-01-01
A new compact large field of view (FOV) multi-camera system is introduced. The camera is based on seven tiny complementary metal-oxide-semiconductor sensor modules covering over 160° × 160° FOV. Although image stitching has been studied extensively, sensor and lens differences have not been considered in previous multi-camera devices. In this study, we have calibrated the photometric characteristics of the multi-camera device. Lenses were not mounted on the sensor in the process of radiometric response calibration to eliminate the influence of the focusing effect of uniform light from an integrating sphere. Linearity range of the radiometric response, non-linearity response characteristics, sensitivity, and dark current of the camera response function are presented. The R, G, and B channels have different responses for the same illuminance. Vignetting artifact patterns have been tested. The actual luminance of the object is retrieved by sensor calibration results, and is used to blend images to make panoramas reflect the objective luminance more objectively. This compensates for the limitation of stitching images that are more realistic only through the smoothing method. The dynamic range limitation of can be resolved by using multiple cameras that cover a large field of view instead of a single image sensor with a wide-angle lens. The dynamic range is expanded by 48-fold in this system. We can obtain seven images in one shot with this multi-camera system, at 13 frames per second. PMID:27077857
Liu, Chang; Deng, Lei; He, Jiale; Li, Di; Fu, Songnian; Tang, Ming; Cheng, Mengfan; Liu, Deming
2017-07-24
In this paper, 4 × 4 multiple-input multiple-output (MIMO) radio over 7-core fiber system based on sparse code multiple access (SCMA) and OFDM/OQAM techniques is proposed. No cyclic prefix (CP) is required by properly designing the prototype filters in OFDM/OQAM modulator, and non-orthogonally overlaid codewords by using SCMA is help to serve more users simultaneously under the condition of using equal number of time and frequency resources compared with OFDMA, resulting in the increase of spectral efficiency (SE) and system capacity. In our experiment, 11.04 Gb/s 4 × 4 MIMO SCMA-OFDM/OQAM signal is successfully transmitted over 20 km 7-core fiber and 0.4 m air distance in both uplink and downlink. As a comparison, 6.681 Gb/s traditional MIMO-OFDM signal with the same occupied bandwidth has been evaluated for both uplink and downlink transmission. The experimental results show that SE could be increased by 65.2% with no bit error rate (BER) performance degradation compared with the traditional MIMO-OFDM technique.
The Midlatitude Continental Convective Clouds Experiment (MC3E)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jensen, Mark P.; Petersen, Walt A.; Bansemer, Aaron
The Midlatitude Continental Convective Clouds Experiment (MC3E), a field program jointly led by the U.S. Department of Energy’s Atmospheric Radiation Measurement program and the NASA Global Precipitation Measurement (GPM) Mission, was conducted in south-central Oklahoma during April – May 2011. MC3E science objectives were motivated by the need to improve understanding of midlatitude continental convective cloud system lifecycles, microphysics, and GPM precipitation retrieval algorithms. To achieve these objectives a multi-scale surface- and aircraft-based in situ and remote sensing observing strategy was employed. A variety of cloud and precipitation events were sampled during the MC3E, of which results from three deepmore » convective events are highlighted. Vertical structure, air motions, precipitation drop-size distributions and ice properties were retrieved from multi-wavelength radar, profiler, and aircraft observations for an MCS on 11 May. Aircraft observations for another MCS observed on 20 May were used to test agreement between observed radar reflectivities and those calculated with forward-modeled reflectivity and microwave brightness temperatures using in situ particle size distributions and ice water content. Multi-platform observations of a supercell that occurred on 23 May allowed for an integrated analysis of kinematic and microphysical interactions. A core updraft of 25 ms-1 supported growth of hail and large rain drops. Data collected during the MC3E campaign is being used in a number of current and ongoing research projects and is available through the DOE ARM and NASA data archives.« less
Photonic-Networks-on-Chip for High Performance Radiation Survivable Multi-Core Processor Systems
2013-12-01
Loss Spectra” Proceedings of SPIE 8255, (2012) and in a journal publication: M. T. Crowley, D. Murrell, N. Patel, M. Breivik , C.-Y. Lin, Y. Li, B.-O...Crowley, D. Murrell, N. Patel, M. Breivik , C.-Y. Lin, Y. Li, B.-O. Fimland and L. F. Lester, "Analytical Modeling of the Temperature Performance of
Zlotnik, V.A.; McGuire, V.L.
1998-01-01
Using the developed theory and modified Springer-Gelhar (SG) model, an identification method is proposed for estimating hydraulic conductivity from multi-level slug tests. The computerized algorithm calculates hydraulic conductivity from both monotonic and oscillatory well responses obtained using a double-packer system. Field verification of the method was performed at a specially designed fully penetrating well of 0.1-m diameter with a 10-m screen in a sand and gravel alluvial aquifer (MSEA site, Shelton, Nebraska). During well installation, disturbed core samples were collected every 0.6 m using a split-spoon sampler. Vertical profiles of hydraulic conductivity were produced on the basis of grain-size analysis of the disturbed core samples. These results closely correlate with the vertical profile of horizontal hydraulic conductivity obtained by interpreting multi-level slug test responses using the modified SG model. The identification method was applied to interpret the response from 474 slug tests in 156 locations at the MSEA site. More than 60% of responses were oscillatory. The method produced a good match to experimental data for both oscillatory and monotonic responses using an automated curve matching procedure. The proposed method allowed us to drastically increase the efficiency of each well used for aquifer characterization and to process massive arrays of field data. Recommendations generalizing this experience to massive application of the proposed method are developed.Using the developed theory and modified Springer-Gelhar (SG) model, an identification method is proposed for estimating hydraulic conductivity from multi-level slug tests. The computerized algorithm calculates hydraulic conductivity from both monotonic and oscillatory well responses obtained using a double-packer system. Field verification of the method was performed at a specially designed fully penetrating well of 0.1-m diameter with a 10-m screen in a sand and gravel alluvial aquifer (MSEA site, Shelton, Nebraska). During well installation, disturbed core samples were collected every 0.6 m using a split-spoon sampler. Vertical profiles of hydraulic conductivity were produced on the basis of grain-size analysis of the disturbed core samples. These results closely correlate with the vertical profile of horizontal hydraulic conductivity obtained by interpreting multi-level slug test responses using the modified SG model. The identification method was applied to interpret the response from 474 slug tests in 156 locations at the MSEA site. More than 60% of responses were oscillatory. The method produced a good match to experimental data for both oscillatory and monotonic responses using an automated curve matching procedure. The proposed method allowed us to drastically increase the efficiency of each well used for aquifer characterization and to process massive arrays of field data. Recommendations generalizing this experience to massive application of the proposed method are developed.
Research on Intelligent Control System of DC SQUID Magnetometer Parameters for Multi-channel System
NASA Astrophysics Data System (ADS)
Chen, Hua; Yang, Kang; Lu, Li; Kong, Xiangyan; Wang, Hai; Wu, Jun; Wang, Yongliang
2018-07-01
In a multi-channel SQUID measurement system, adjusting device parameters to optimal condition for all channels is time-consuming. In this paper, an intelligent control system is presented to determine the optimal working point of devices which is automatic and more efficient comparing to the manual one. An optimal working point searching algorithm is introduced as the core component of the control system. In this algorithm, the bias voltage V_bias is step scanned to obtain the maximal value of the peak-to-peak current value I_pp of the SQUID magnetometer modulation curve. We choose this point as the optimal one. Using the above control system, more than 30 weakly damped SQUID magnetometers with area of 5 × 5 mm^2 or 10 × 10 mm^2 are adjusted and a 36-channel magnetocardiography system perfectly worked in a magnetically shielded room. The average white flux noise is 15 {μ Φ }_0/Hz^{1/2}.
Research on Intelligent Control System of DC SQUID Magnetometer Parameters for Multi-channel System
NASA Astrophysics Data System (ADS)
Chen, Hua; Yang, Kang; Lu, Li; Kong, Xiangyan; Wang, Hai; Wu, Jun; Wang, Yongliang
2018-03-01
In a multi-channel SQUID measurement system, adjusting device parameters to optimal condition for all channels is time-consuming. In this paper, an intelligent control system is presented to determine the optimal working point of devices which is automatic and more efficient comparing to the manual one. An optimal working point searching algorithm is introduced as the core component of the control system. In this algorithm, the bias voltage V_bias is step scanned to obtain the maximal value of the peak-to-peak current value I_pp of the SQUID magnetometer modulation curve. We choose this point as the optimal one. Using the above control system, more than 30 weakly damped SQUID magnetometers with area of 5 × 5 mm^2 or 10 × 10 mm^2 are adjusted and a 36-channel magnetocardiography system perfectly worked in a magnetically shielded room. The average white flux noise is 15 μΦ_0/Hz^{1/2}.
NASA Astrophysics Data System (ADS)
Makowska, A.; Markiewicz, K.; Szostkiewicz, L.; Kolakowska, A.; Fidelus, J.; Stanczyk, T.; Wysokinski, K.; Budnicki, D.; Ostrowski, L.; Szymanski, M.; Makara, M.; Poturaj, K.; Tenderenda, T.; Mergo, P.; Nasilowski, T.
2018-02-01
Sensors based on fiber optics are irreplaceable wherever immunity to strong electro-magnetic fields or safe operation in explosive atmospheres is needed. Furthermore, it is often essential to be able to monitor high temperatures of over 500°C in such environments (e.g. in cooling systems or equipment monitoring in power plants). In order to meet this demand, we have designed and manufactured a fiber optic sensor with which temperatures up to 900°C can be measured. The sensor utilizes multi-core fibers which are recognized as the dedicated medium for telecommunication or shape sensing, but as we show may be also deployed advantageously in new types of fiber optic temperature sensors. The sensor presented in this paper is based on a dual-core microstructured fiber Michelson interferometer. The fiber is characterized by strongly coupled cores, hence it acts as an all-fiber coupler, but with an outer diameter significantly wider than a standard fused biconical taper coupler, which significantly increases the coupling region's mechanical reliability. Owing to the proposed interferometer imbalance, effective operation and high-sensitivity can be achieved. The presented sensor is designed to be used at high temperatures as a result of the developed low temperature chemical process of metal (copper or gold) coating. The hermetic metal coating can be applied directly to the silica cladding of the fiber or the fiber component. This operation significantly reduces the degradation of sensors due to hydrolysis in uncontrolled atmospheres and high temperatures.
Coccarelli, Alberto; Boileau, Etienne; Parthimos, Dimitris; Nithiarasu, Perumal
2016-10-01
In the present work, an elaborate one-dimensional thermofluid model for a human body is presented. By contrast to the existing pure conduction-/perfusion-based models, the proposed methodology couples the arterial fluid dynamics of a human body with a multi-segmental bioheat model of surrounding solid tissues. In the present configuration, arterial flow is included through a network of elastic vessels. More than a dozen solid segments are employed to represent the heat conduction in the surrounding tissues, and each segment is constituted by a multilayered circular cylinder. Such multi-layers allow flexible delineation of the geometry and incorporation of properties of different tissue types. The coupling of solid tissue and fluid models requires subdivision of the arterial circulation into large and small arteries. The heat exchange between tissues and arterial wall occurs by convection in large vessels and by perfusion in small arteries. The core region, including the heart, provides the inlet conditions for the fluid equations. In the proposed model, shivering, sweating, and perfusion changes constitute the basis of the thermoregulatory system. The equations governing flow and heat transfer in the circulatory system are solved using a locally conservative Galerkin approach, and the heat conduction in the surrounding tissues is solved using a standard implicit backward Euler method. To investigate the effectiveness of the proposed model, temperature field evolutions are monitored at different points of the arterial tree and in the surrounding tissue layers. To study the differences due to flow-induced convection effects on thermal balance, the results of the current model are compared against those of the widely used modelling methodologies. The results show that the convection significantly influences the temperature distribution of the solid tissues in the vicinity of the arteries. Thus, the inner convection has a more predominant role in the human body heat balance than previously thought. To demonstrate its capabilities, the proposed new model is used to study different scenarios, including thermoregulation inactivity and variation in surrounding atmospheric conditions.
Description of the Large Gap Magnetic Suspension System (LGMSS) ground-based experiment
NASA Technical Reports Server (NTRS)
Groom, Nelson J.
1991-01-01
A description of the Large Gap Magnetic Suspension System (LGMSS) ground-based experiment is presented. The LGMSS provides five degrees of freedom control of a cylindrical suspended element which is levitated above a floor-mounted array of air core electromagnets. The uncontrolled degree of freedom is rotation about the long axis of the cylinder (roll). Levitation and control forces are produced on a permanent magnet core which is embedded in the cylinder. The cylinder also contains light emitting diodes (LEDs), assorted electrons, and a power supply. The LEDs provide active targets for an optical position measurement system which is being developed in-house at the Langley Research Center. The optical position measurement system will provide six degrees of freedom position information for the LGMSS control system.
NASA Astrophysics Data System (ADS)
Ali, Amir R.; Kamel, Mohamed A.
2017-05-01
This paper studies the effect of the electrostriction force on the single optical dielectric core coated with multi-layers based on whispering gallery mode (WGM). The sensing element is a dielectric core made of polymeric material coated with multi-layers having different dielectric and mechanical properties. The external electric field deforming the sensing element causing shifts in its WGM spectrum. The multi-layer structures will enhance the body and the pressure forces acting on the core of the sensing element. Due to the gradient on the dielectric permittivity; pressure forces at the interface between every two layers will be created. Also, the gradient on Young's modulus will affect the overall stiffness of the optical sensor. In turn the sensitivity of the optical sensor to the electric field will be increased when the materials of each layer selected properly. A mathematical model is used to test the effect for that multi-layer structures. Two layering techniques are considered to increase the sensor's sensitivity; (i) Pressure force enhancement technique; and (ii) Young's modulus reduction technique. In the first technique, Young's modulus is kept constant for all layers, while the dielectric permittivity is varying. In this technique the results will be affected by the value dielectric permittivity of the outer medium surrounding the cavity. If the medium's dielectric permittivity is greater than that of the cavity, then the ascending ordered layers of the cavity will yield the highest sensitivity (the core will have the smallest dielectric permittivity) to the applied electric field and vice versa. In the second technique, Young's modulus is varying along the layers, while the dielectric permittivity has a certain constant value per layer. On the other hand, the descending order will enhance the sensitivity in the second technique. Overall, results show the multi-layer cavity based on these techniques will enhance the sensitivity compared to the typical polymeric optical sensor.
Contamination of arctic Fjord sediments by Pb-Zn mining at Maarmorilik in central West Greenland.
Perner, K; Leipe, Th; Dellwig, O; Kuijpers, A; Mikkelsen, N; Andersen, T J; Harff, J
2010-07-01
This study focuses on heavy metal contamination of arctic sediments from a small Fjord system adjacent to the Pb-Zn "Black Angel" mine (West Greenland) to investigate the temporal and spatial development of contamination and to provide baseline levels before the mines re-opening in January 2009. For this purpose we collected multi-cores along a transect from Affarlikassaa Fjord, which received high amounts of tailings from 1973 to 1990, to the mouth of Qaumarujuk Fjord. Along with radiochemical dating by (210)Pb and (137)Cs, geochemical analyses of heavy metals (e.g. As, Cd, Hg, Pb, and Zn) were carried out. Maximum contents were found at 12 cm depth in Affarlikassaa. After 17 years the mine last closed, specific local hydrographic conditions continue to disperse heavy metal enriched material derived from the Affarlikassaa into Qaumarujuk. Total Hg profiles from multi-cores along the transect clearly illustrate this transport and spatial distribution pattern of the contaminated material. Copyright 2010 Elsevier Ltd. All rights reserved.
A superconducting large-angle magnetic suspension
NASA Technical Reports Server (NTRS)
Downer, James; Goldie, James; Torti, Richard
1991-01-01
The component technologies were developed required for an advanced control moment gyro (CMG) type of slewing actuator for large payloads. The key component of the CMG is a large-angle magnetic suspension (LAMS). The LAMS combines the functions of the gimbal structure, torque motors, and rotor bearings of a CMG. The LAMS uses a single superconducting source coil and an array of cryoresistive control coils to produce a specific output torque more than an order of magnitude greater than conventional devices. The designed and tested LAMS system is based around an available superconducting solenoid, an array of twelve room-temperature normal control coils, and a multi-input, multi-output control system. The control laws were demonstrated for stabilizing and controlling the LAMS system.
Laurinaviciene, Aida; Plancoulaine, Benoit; Baltrusaityte, Indra; Meskauskas, Raimundas; Besusparis, Justinas; Lesciute-Krilaviciene, Daiva; Raudeliunas, Darius; Iqbal, Yasir; Herlin, Paulette; Laurinavicius, Arvydas
2014-01-01
Digital immunohistochemistry (IHC) is one of the most promising applications brought by new generation image analysis (IA). While conventional IHC staining quality is monitored by semi-quantitative visual evaluation of tissue controls, IA may require more sensitive measurement. We designed an automated system to digitally monitor IHC multi-tissue controls, based on SQL-level integration of laboratory information system with image and statistical analysis tools. Consecutive sections of TMA containing 10 cores of breast cancer tissue were used as tissue controls in routine Ki67 IHC testing. Ventana slide label barcode ID was sent to the LIS to register the serial section sequence. The slides were stained and scanned (Aperio ScanScope XT), IA was performed by the Aperio/Leica Colocalization and Genie Classifier/Nuclear algorithms. SQL-based integration ensured automated statistical analysis of the IA data by the SAS Enterprise Guide project. Factor analysis and plot visualizations were performed to explore slide-to-slide variation of the Ki67 IHC staining results in the control tissue. Slide-to-slide intra-core IHC staining analysis revealed rather significant variation of the variables reflecting the sample size, while Brown and Blue Intensity were relatively stable. To further investigate this variation, the IA results from the 10 cores were aggregated to minimize tissue-related variance. Factor analysis revealed association between the variables reflecting the sample size detected by IA and Blue Intensity. Since the main feature to be extracted from the tissue controls was staining intensity, we further explored the variation of the intensity variables in the individual cores. MeanBrownBlue Intensity ((Brown+Blue)/2) and DiffBrownBlue Intensity (Brown-Blue) were introduced to better contrast the absolute intensity and the colour balance variation in each core; relevant factor scores were extracted. Finally, tissue-related factors of IHC staining variance were explored in the individual tissue cores. Our solution enabled to monitor staining of IHC multi-tissue controls by the means of IA, followed by automated statistical analysis, integrated into the laboratory workflow. We found that, even in consecutive serial tissue sections, tissue-related factors affected the IHC IA results; meanwhile, less intense blue counterstain was associated with less amount of tissue, detected by the IA tools.