Science.gov

Sample records for highly scalable udp-based

  1. Highly Scalable, UDP-Based Network Transport Protocols for Lambda Grids and 10 GE Routed Networks

    SciTech Connect

    PI: Robert Grossman Co-PI: Stephen Eick

    2009-08-04

    Summary of Report In work prior to this grant, NCDM developed a high performance data transport protocol called SABUL. During this grant, we refined SABUL’s functionality, and then extended both the capabilities and functionality and incorporated them into a new protocol called UDP-based Data transport Protocol, or UDT. We also began preliminary work on Composable UDT, a version of UDT that allows the user to choose among different congestion control algorithms and implement the algorithm of his choice at the time he compiles the code. Specifically, we: · Investigated the theoretical foundations of protocols similar to SABUL and UDT. · Performed design and development work of UDT, a protocol that uses UDP in both the data and control channels. · Began design and development work of Composable UDT, a protocol that supports the use of different congestion control algorithms by simply including the appropriate library when compiling the code. · Performed experimental studies using UDT and Composable UDT using real world applications such as the Sloan Digital Sky Survey (SDSS) astronomical data sets. · Released several versions of UDT and Composable, the most recent being v3.1.

  2. Highly scalable coherent fiber combining

    NASA Astrophysics Data System (ADS)

    Antier, M.; Bourderionnet, J.; Larat, C.; Lallier, E.; Brignon, A.

    2015-10-01

    An architecture for active coherent fiber laser beam combining using an interferometric measurement is demonstrated. This technique allows measuring the exact phase errors of each fiber beam in a single shot. Therefore, this method is a promising candidate toward very large number of combined fibers. Our experimental system, composed of 16 independent fiber channels, is used to evaluate the achieved phase locking stability in terms of phase shift error and bandwidth. We show that only 8 pixels per fiber on the camera is required for a stable close loop operation with a residual phase error of λ/20 rms, which demonstrates the scalability of this concept. Furthermore we propose a beam shaping technique to increase the combining efficiency.

  3. Scalable resource management in high performance computers.

    SciTech Connect

    Frachtenberg, E.; Petrini, F.; Fernandez Peinador, J.; Coll, S.

    2002-01-01

    Clusters of workstations have emerged as an important platform for building cost-effective, scalable and highly-available computers. Although many hardware solutions are available today, the largest challenge in making large-scale clusters usable lies in the system software. In this paper we present STORM, a resource management tool designed to provide scalability, low overhead and the flexibility necessary to efficiently support and analyze a wide range of job scheduling algorithms. STORM achieves these feats by closely integrating the management daemons with the low-level features that are common in state-of-the-art high-performance system area networks. The architecture of STORM is based on three main technical innovations. First, a sizable part of the scheduler runs in the thread processor located on the network interface. Second, we use hardware collectives that are highly scalable both for implementing control heartbeats and to distribute the binary of a parallel job in near-constant time, irrespective of job and machine sizes. Third, we use an I/O bypass protocol that allows fast data movements from the file system to the communication buffers in the network interface and vice versa. The experimental results show that STORM can launch a job with a binary of 12MB on a 64 processor/32 node cluster in less than 0.25 sec on an empty network, in less than 0.45 sec when all the processors are busy computing other jobs, and in less than 0.65 sec when the network is flooded with a background traffic. This paper provides experimental and analytical evidence that these results scale to a much larger number of nodes. To the best of our knowledge, STORM is at least two orders of magnitude faster than existing production schedulers in launching jobs, performing resource management tasks and gang scheduling.

  4. Highly Scalable Matching Pursuit Signal Decomposition Algorithm

    NASA Technical Reports Server (NTRS)

    Christensen, Daniel; Das, Santanu; Srivastava, Ashok N.

    2009-01-01

    Matching Pursuit Decomposition (MPD) is a powerful iterative algorithm for signal decomposition and feature extraction. MPD decomposes any signal into linear combinations of its dictionary elements or atoms . A best fit atom from an arbitrarily defined dictionary is determined through cross-correlation. The selected atom is subtracted from the signal and this procedure is repeated on the residual in the subsequent iterations until a stopping criterion is met. The reconstructed signal reveals the waveform structure of the original signal. However, a sufficiently large dictionary is required for an accurate reconstruction; this in return increases the computational burden of the algorithm, thus limiting its applicability and level of adoption. The purpose of this research is to improve the scalability and performance of the classical MPD algorithm. Correlation thresholds were defined to prune insignificant atoms from the dictionary. The Coarse-Fine Grids and Multiple Atom Extraction techniques were proposed to decrease the computational burden of the algorithm. The Coarse-Fine Grids method enabled the approximation and refinement of the parameters for the best fit atom. The ability to extract multiple atoms within a single iteration enhanced the effectiveness and efficiency of each iteration. These improvements were implemented to produce an improved Matching Pursuit Decomposition algorithm entitled MPD++. Disparate signal decomposition applications may require a particular emphasis of accuracy or computational efficiency. The prominence of the key signal features required for the proper signal classification dictates the level of accuracy necessary in the decomposition. The MPD++ algorithm may be easily adapted to accommodate the imposed requirements. Certain feature extraction applications may require rapid signal decomposition. The full potential of MPD++ may be utilized to produce incredible performance gains while extracting only slightly less energy than the

  5. Technical Report: Scalable Parallel Algorithms for High Dimensional Numerical Integration

    SciTech Connect

    Masalma, Yahya; Jiao, Yu

    2010-10-01

    We implemented a scalable parallel quasi-Monte Carlo numerical high-dimensional integration for tera-scale data points. The implemented algorithm uses the Sobol s quasi-sequences to generate random samples. Sobol s sequence was used to avoid clustering effects in the generated random samples and to produce low-discrepancy random samples which cover the entire integration domain. The performance of the algorithm was tested. Obtained results prove the scalability and accuracy of the implemented algorithms. The implemented algorithm could be used in different applications where a huge data volume is generated and numerical integration is required. We suggest using the hyprid MPI and OpenMP programming model to improve the performance of the algorithms. If the mixed model is used, attention should be paid to the scalability and accuracy.

  6. Low power, scalable multichannel high voltage controller

    DOEpatents

    Stamps, James Frederick; Crocker, Robert Ward; Yee, Daniel Dadwa; Dils, David Wright

    2006-03-14

    A low voltage control circuit is provided for individually controlling high voltage power provided over bus lines to a multitude of interconnected loads. An example of a load is a drive for capillary channels in a microfluidic system. Control is distributed from a central high voltage circuit, rather than using a number of large expensive central high voltage circuits to enable reducing circuit size and cost. Voltage is distributed to each individual load and controlled using a number of high voltage controller channel switches connected to high voltage bus lines. The channel switches each include complementary pull up and pull down photo isolator relays with photo isolator switching controlled from the central high voltage circuit to provide a desired bus line voltage. Switching of the photo isolator relays is further controlled in each channel switch using feedback from a resistor divider circuit to maintain the bus voltage swing within desired limits. Current sensing is provided using a switched resistive load in each channel switch, with switching of the resistive loads controlled from the central high voltage circuit.

  7. Low power, scalable multichannel high voltage controller

    DOEpatents

    Stamps, James Frederick; Crocker, Robert Ward; Yee, Daniel Dadwa; Dils, David Wright

    2008-03-25

    A low voltage control circuit is provided for individually controlling high voltage power provided over bus lines to a multitude of interconnected loads. An example of a load is a drive for capillary channels in a microfluidic system. Control is distributed from a central high voltage circuit, rather than using a number of large expensive central high voltage circuits to enable reducing circuit size and cost. Voltage is distributed to each individual load and controlled using a number of high voltage controller channel switches connected to high voltage bus lines. The channel switches each include complementary pull up and pull down photo isolator relays with photo isolator switching controlled from the central high voltage circuit to provide a desired bus line voltage. Switching of the photo isolator relays is further controlled in each channel switch using feedback from a resistor divider circuit to maintain the bus voltage swing within desired limits. Current sensing is provided using a switched resistive load in each channel switch, with switching of the resistive loads controlled from the central high voltage circuit.

  8. Scalable, flexible and high resolution patterning of CVD graphene.

    PubMed

    Hofmann, Mario; Hsieh, Ya-Ping; Hsu, Allen L; Kong, Jing

    2014-01-01

    The unique properties of graphene make it a promising material for interconnects in flexible and transparent electronics. To increase the commercial impact of graphene in those applications, a scalable and economical method for producing graphene patterns is required. The direct synthesis of graphene from an area-selectively passivated catalyst substrate can generate patterned graphene of high quality. We here present a solution-based method for producing patterned passivation layers. Various deposition methods such as ink-jet deposition and microcontact printing were explored, that can satisfy application demands for low cost, high resolution and scalable production of patterned graphene. The demonstrated high quality and nanometer precision of grown graphene establishes the potential of this synthesis approach for future commercial applications of graphene. Finally, the ability to transfer high resolution graphene patterns onto complex three-dimensional surfaces affords the vision of graphene-based interconnects in novel electronics. PMID:24189709

  9. Scalable, flexible and high resolution patterning of CVD graphene

    NASA Astrophysics Data System (ADS)

    Hofmann, Mario; Hsieh, Ya-Ping; Hsu, Allen L.; Kong, Jing

    2013-12-01

    The unique properties of graphene make it a promising material for interconnects in flexible and transparent electronics. To increase the commercial impact of graphene in those applications, a scalable and economical method for producing graphene patterns is required. The direct synthesis of graphene from an area-selectively passivated catalyst substrate can generate patterned graphene of high quality. We here present a solution-based method for producing patterned passivation layers. Various deposition methods such as ink-jet deposition and microcontact printing were explored, that can satisfy application demands for low cost, high resolution and scalable production of patterned graphene. The demonstrated high quality and nanometer precision of grown graphene establishes the potential of this synthesis approach for future commercial applications of graphene. Finally, the ability to transfer high resolution graphene patterns onto complex three-dimensional surfaces affords the vision of graphene-based interconnects in novel electronics.The unique properties of graphene make it a promising material for interconnects in flexible and transparent electronics. To increase the commercial impact of graphene in those applications, a scalable and economical method for producing graphene patterns is required. The direct synthesis of graphene from an area-selectively passivated catalyst substrate can generate patterned graphene of high quality. We here present a solution-based method for producing patterned passivation layers. Various deposition methods such as ink-jet deposition and microcontact printing were explored, that can satisfy application demands for low cost, high resolution and scalable production of patterned graphene. The demonstrated high quality and nanometer precision of grown graphene establishes the potential of this synthesis approach for future commercial applications of graphene. Finally, the ability to transfer high resolution graphene patterns onto

  10. Scalable Multiprocessor for High-Speed Computing in Space

    NASA Technical Reports Server (NTRS)

    Lux, James; Lang, Minh; Nishimoto, Kouji; Clark, Douglas; Stosic, Dorothy; Bachmann, Alex; Wilkinson, William; Steffke, Richard

    2004-01-01

    A report discusses the continuing development of a scalable multiprocessor computing system for hard real-time applications aboard a spacecraft. "Hard realtime applications" signifies applications, like real-time radar signal processing, in which the data to be processed are generated at "hundreds" of pulses per second, each pulse "requiring" millions of arithmetic operations. In these applications, the digital processors must be tightly integrated with analog instrumentation (e.g., radar equipment), and data input/output must be synchronized with analog instrumentation, controlled to within fractions of a microsecond. The scalable multiprocessor is a cluster of identical commercial-off-the-shelf generic DSP (digital-signal-processing) computers plus generic interface circuits, including analog-to-digital converters, all controlled by software. The processors are computers interconnected by high-speed serial links. Performance can be increased by adding hardware modules and correspondingly modifying the software. Work is distributed among the processors in a parallel or pipeline fashion by means of a flexible master/slave control and timing scheme. Each processor operates under its own local clock; synchronization is achieved by broadcasting master time signals to all the processors, which compute offsets between the master clock and their local clocks.

  11. High-performance, scalable optical network-on-chip architectures

    NASA Astrophysics Data System (ADS)

    Tan, Xianfang

    The rapid advance of technology enables a large number of processing cores to be integrated into a single chip which is called a Chip Multiprocessor (CMP) or a Multiprocessor System-on-Chip (MPSoC) design. The on-chip interconnection network, which is the communication infrastructure for these processing cores, plays a central role in a many-core system. With the continuously increasing complexity of many-core systems, traditional metallic wired electronic networks-on-chip (NoC) became a bottleneck because of the unbearable latency in data transmission and extremely high energy consumption on chip. Optical networks-on-chip (ONoC) has been proposed as a promising alternative paradigm for electronic NoC with the benefits of optical signaling communication such as extremely high bandwidth, negligible latency, and low power consumption. This dissertation focus on the design of high-performance and scalable ONoC architectures and the contributions are highlighted as follow: 1. A micro-ring resonator (MRR)-based Generic Wavelength-routed Optical Router (GWOR) is proposed. A method for developing any sized GWOR is introduced. GWOR is a scalable non-blocking ONoC architecture with simple structure, low cost and high power efficiency compared to existing ONoC designs. 2. To expand the bandwidth and improve the fault tolerance of the GWOR, a redundant GWOR architecture is designed by cascading different type of GWORs into one network. 3. The redundant GWOR built with MRR-based comb switches is proposed. Comb switches can expand the bandwidth while keep the topology of GWOR unchanged by replacing the general MRRs with comb switches. 4. A butterfly fat tree (BFT)-based hybrid optoelectronic NoC (HONoC) architecture is developed in which GWORs are used for global communication and electronic routers are used for local communication. The proposed HONoC uses less numbers of electronic routers and links than its counterpart of electronic BFT-based NoC. It takes the advantages of

  12. Highly flexible electronics from scalable vertical thin film transistors.

    PubMed

    Liu, Yuan; Zhou, Hailong; Cheng, Rui; Yu, Woojong; Huang, Yu; Duan, Xiangfeng

    2014-03-12

    Flexible thin-film transistors (TFTs) are of central importance for diverse electronic and particularly macroelectronic applications. The current TFTs using organic or inorganic thin film semiconductors are usually limited by either poor electrical performance or insufficient mechanical flexibility. Here, we report a new design of highly flexible vertical TFTs (VTFTs) with superior electrical performance and mechanical robustness. By using the graphene as a work-function tunable contact for amorphous indium gallium zinc oxide (IGZO) thin film, the vertical current flow across the graphene-IGZO junction can be effectively modulated by an external gate potential to enable VTFTs with a highest on-off ratio exceeding 10(5). The unique vertical transistor architecture can readily enable ultrashort channel devices with very high delivering current and exceptional mechanical flexibility. With large area graphene and IGZO thin film available, our strategy is intrinsically scalable for large scale integration of VTFT arrays and logic circuits, opening up a new pathway to highly flexible macroelectronics. PMID:24502192

  13. Highly scalable and robust rule learner: performance evaluation and comparison.

    PubMed

    Kurgan, Lukasz A; Cios, Krzysztof J; Dick, Scott

    2006-02-01

    Business intelligence and bioinformatics applications increasingly require the mining of datasets consisting of millions of data points, or crafting real-time enterprise-level decision support systems for large corporations and drug companies. In all cases, there needs to be an underlying data mining system, and this mining system must be highly scalable. To this end, we describe a new rule learner called DataSqueezer. The learner belongs to the family of inductive supervised rule extraction algorithms. DataSqueezer is a simple, greedy, rule builder that generates a set of production rules from labeled input data. In spite of its relative simplicity, DataSqueezer is a very effective learner. The rules generated by the algorithm are compact, comprehensible, and have accuracy comparable to rules generated by other state-of-the-art rule extraction algorithms. The main advantages of DataSqueezer are very high efficiency, and missing data resistance. DataSqueezer exhibits log-linear asymptotic complexity with the number of training examples, and it is faster than other state-of-the-art rule learners. The learner is also robust to large quantities of missing data, as verified by extensive experimental comparison with the other learners. DataSqueezer is thus well suited to modern data mining and business intelligence tasks, which commonly involve huge datasets with a large fraction of missing data. PMID:16468565

  14. Highly scalable parallel processing of extracellular recordings of Multielectrode Arrays.

    PubMed

    Gehring, Tiago V; Vasilaki, Eleni; Giugliano, Michele

    2015-01-01

    Technological advances of Multielectrode Arrays (MEAs) used for multisite, parallel electrophysiological recordings, lead to an ever increasing amount of raw data being generated. Arrays with hundreds up to a few thousands of electrodes are slowly seeing widespread use and the expectation is that more sophisticated arrays will become available in the near future. In order to process the large data volumes resulting from MEA recordings there is a pressing need for new software tools able to process many data channels in parallel. Here we present a new tool for processing MEA data recordings that makes use of new programming paradigms and recent technology developments to unleash the power of modern highly parallel hardware, such as multi-core CPUs with vector instruction sets or GPGPUs. Our tool builds on and complements existing MEA data analysis packages. It shows high scalability and can be used to speed up some performance critical pre-processing steps such as data filtering and spike detection, helping to make the analysis of larger data sets tractable. PMID:26737215

  15. High Performance Storage System Scalability: Architecture, Implementation, and Experience

    SciTech Connect

    Watson, R W

    2005-01-05

    The High Performance Storage System (HPSS) provides scalable hierarchical storage management (HSM), archive, and file system services. Its design, implementation and current dominant use are focused on HSM and archive services. It is also a general-purpose, global, shared, parallel file system, potentially useful in other application domains. When HPSS design and implementation began over a decade ago, scientific computing power and storage capabilities at a site, such as a DOE national laboratory, was measured in a few 10s of gigaops, data archived in HSMs in a few 10s of terabytes at most, data throughput rates to an HSM in a few megabytes/s, and daily throughput with the HSM in a few gigabytes/day. At that time, the DOE national laboratories and IBM HPSS design team recognized that we were headed for a data storage explosion driven by computing power rising to teraops/petaops requiring data stored in HSMs to rise to petabytes and beyond, data transfer rates with the HSM to rise to gigabytes/s and higher, and daily throughput with a HSM in 10s of terabytes/day. This paper discusses HPSS architectural, implementation and deployment experiences that contributed to its success in meeting the above orders of magnitude scaling targets. We also discuss areas that need additional attention as we continue significant scaling into the future.

  16. Providing scalable system software for high-end simulations

    SciTech Connect

    Greenberg, D.

    1997-12-31

    Detailed, full-system, complex physics simulations have been shown to be feasible on systems containing thousands of processors. In order to manage these computer systems it has been necessary to create scalable system services. In this talk Sandia`s research on scalable systems will be described. The key concepts of low overhead data movement through portals and of flexible services through multi-partition architectures will be illustrated in detail. The talk will conclude with a discussion of how these techniques can be applied outside of the standard monolithic MPP system.

  17. High-Performance Scalable Information Service for the ATLAS Experiment

    NASA Astrophysics Data System (ADS)

    Kolos, S.; Boutsioukis, G.; Hauser, R.

    2012-12-01

    The ATLAS[1] experiment is operated by a highly distributed computing system which is constantly producing a lot of status information which is used to monitor the experiment operational conditions as well as to assess the quality of the physics data being taken. For example the ATLAS High Level Trigger(HLT) algorithms are executed on the online computing farm consisting from about 1500 nodes. Each HLT algorithm is producing few thousands histograms, which have to be integrated over the whole farm and carefully analyzed in order to properly tune the event rejection. In order to handle such non-physics data the Information Service (IS) facility has been developed in the scope of the ATLAS Trigger and Data Acquisition (TDAQ)[2] project. The IS provides a high-performance scalable solution for information exchange in distributed environment. In the course of an ATLAS data taking session the IS handles about a hundred gigabytes of information which is being constantly updated with the update interval varying from a second to a few tens of seconds. IS provides access to any information item on request as well as distributing notification to all the information subscribers. In the latter case IS subscribers receive information within a few milliseconds after it was updated. IS can handle arbitrary types of information, including histograms produced by the HLT applications, and provides C++, Java and Python API. The Information Service is a unique source of information for the majority of the online monitoring analysis and GUI applications used to control and monitor the ATLAS experiment. Information Service provides streaming functionality allowing efficient replication of all or part of the managed information. This functionality is used to duplicate the subset of the ATLAS monitoring data to the CERN public network with a latency of a few milliseconds, allowing efficient real-time monitoring of the data taking from outside the protected ATLAS network. Each information

  18. Scalable software-defined optical networking with high-performance routing and wavelength assignment algorithms.

    PubMed

    Lee, Chankyun; Cao, Xiaoyuan; Yoshikane, Noboru; Tsuritani, Takehiro; Rhee, June-Koo Kevin

    2015-10-19

    The feasibility of software-defined optical networking (SDON) for a practical application critically depends on scalability of centralized control performance. The paper, highly scalable routing and wavelength assignment (RWA) algorithms are investigated on an OpenFlow-based SDON testbed for proof-of-concept demonstration. Efficient RWA algorithms are proposed to achieve high performance in achieving network capacity with reduced computation cost, which is a significant attribute in a scalable centralized-control SDON. The proposed heuristic RWA algorithms differ in the orders of request processes and in the procedures of routing table updates. Combined in a shortest-path-based routing algorithm, a hottest-request-first processing policy that considers demand intensity and end-to-end distance information offers both the highest throughput of networks and acceptable computation scalability. We further investigate trade-off relationship between network throughput and computation complexity in routing table update procedure by a simulation study. PMID:26480397

  19. Layered Low-Density Generator Matrix Codes for Super High Definition Scalable Video Coding System

    NASA Astrophysics Data System (ADS)

    Tonomura, Yoshihide; Shirai, Daisuke; Nakachi, Takayuki; Fujii, Tatsuya; Kiya, Hitoshi

    In this paper, we introduce layered low-density generator matrix (Layered-LDGM) codes for super high definition (SHD) scalable video systems. The layered-LDGM codes maintain the correspondence relationship of each layer from the encoder side to the decoder side. This resulting structure supports partial decoding. Furthermore, the proposed layered-LDGM codes create highly efficient forward error correcting (FEC) data by considering the relationship between each scalable component. Therefore, the proposed layered-LDGM codes raise the probability of restoring the important components. Simulations show that the proposed layered-LDGM codes offer better error resiliency than the existing method which creates FEC data for each scalable component independently. The proposed layered-LDGM codes support partial decoding and raise the probability of restoring the base component. These characteristics are very suitable for scalable video coding systems.

  20. Developing highly scalable fluid solvers for enabling multiphysics simulation.

    SciTech Connect

    Clausen, Jonathan

    2013-03-01

    We performed an investigation into explicit algorithms for the simulation of incompressible flows using methods with a finite, but small amount of compressibility added. Such methods include the artificial compressibility method and the lattice-Boltzmann method. The impetus for investigating such techniques stems from the increasing use of parallel computation at all levels (processors, clusters, and graphics processing units). Explicit algorithms have the potential to leverage these resources. In our investigation, a new form of artificial compressibility was derived. This method, referred to as the Entropically Damped Artificial Compressibility (EDAC) method, demonstrated superior results to traditional artificial compressibility methods by damping the numerical acoustic waves associated with these methods. Performance nearing that of the lattice- Boltzmann technique was observed, without the requirement of recasting the problem in terms of particle distribution functions; continuum variables may be used. Several example problems were investigated using a finite-di erence and finite-element discretizations of the EDAC equations. Example problems included lid-driven cavity flow, a convecting Taylor-Green vortex, a doubly periodic shear layer, freely decaying turbulence, and flow over a square cylinder. Additionally, a scalability study was performed using in excess of one million processing cores. Explicit methods were found to have desirable scaling properties; however, some robustness and general applicability issues remained.

  1. A highly scalable, interoperable clinical decision support service

    PubMed Central

    Goldberg, Howard S; Paterno, Marilyn D; Rocha, Beatriz H; Schaeffer, Molly; Wright, Adam; Erickson, Jessica L; Middleton, Blackford

    2014-01-01

    Objective To create a clinical decision support (CDS) system that is shareable across healthcare delivery systems and settings over large geographic regions. Materials and methods The enterprise clinical rules service (ECRS) realizes nine design principles through a series of enterprise java beans and leverages off-the-shelf rules management systems in order to provide consistent, maintainable, and scalable decision support in a variety of settings. Results The ECRS is deployed at Partners HealthCare System (PHS) and is in use for a series of trials by members of the CDS consortium, including internally developed systems at PHS, the Regenstrief Institute, and vendor-based systems deployed at locations in Oregon and New Jersey. Performance measures indicate that the ECRS provides sub-second response time when measured apart from services required to retrieve data and assemble the continuity of care document used as input. Discussion We consider related work, design decisions, comparisons with emerging national standards, and discuss uses and limitations of the ECRS. Conclusions ECRS design, implementation, and use in CDS consortium trials indicate that it provides the flexibility and modularity needed for broad use and performs adequately. Future work will investigate additional CDS patterns, alternative methods of data passing, and further optimizations in ECRS performance. PMID:23828174

  2. Air-stable ink for scalable, high-throughput layer deposition

    DOEpatents

    Weil, Benjamin D; Connor, Stephen T; Cui, Yi

    2014-02-11

    A method for producing and depositing air-stable, easily decomposable, vulcanized ink on any of a wide range of substrates is disclosed. The ink enables high-volume production of optoelectronic and/or electronic devices using scalable production methods, such as roll-to-roll transfer, fast rolling processes, and the like.

  3. Scalable High Performance Computing: Direct and Large-Eddy Turbulent Flow Simulations Using Massively Parallel Computers

    NASA Technical Reports Server (NTRS)

    Morgan, Philip E.

    2004-01-01

    This final report contains reports of research related to the tasks "Scalable High Performance Computing: Direct and Lark-Eddy Turbulent FLow Simulations Using Massively Parallel Computers" and "Devleop High-Performance Time-Domain Computational Electromagnetics Capability for RCS Prediction, Wave Propagation in Dispersive Media, and Dual-Use Applications. The discussion of Scalable High Performance Computing reports on three objectives: validate, access scalability, and apply two parallel flow solvers for three-dimensional Navier-Stokes flows; develop and validate a high-order parallel solver for Direct Numerical Simulations (DNS) and Large Eddy Simulation (LES) problems; and Investigate and develop a high-order Reynolds averaged Navier-Stokes turbulence model. The discussion of High-Performance Time-Domain Computational Electromagnetics reports on five objectives: enhancement of an electromagnetics code (CHARGE) to be able to effectively model antenna problems; utilize lessons learned in high-order/spectral solution of swirling 3D jets to apply to solving electromagnetics project; transition a high-order fluids code, FDL3DI, to be able to solve Maxwell's Equations using compact-differencing; develop and demonstrate improved radiation absorbing boundary conditions for high-order CEM; and extend high-order CEM solver to address variable material properties. The report also contains a review of work done by the systems engineer.

  4. LED light engine concept with ultra-high scalable luminance

    NASA Astrophysics Data System (ADS)

    Hoelen, Christoph; de Boer, Dick; Bruls, Dominique; van der Eyden, Joost; Koole, Rolf; Li, Yun; Mirsadeghi, Mo; Vanbroekhoven, Vincent; Van den Bergh, John-John; Van de Voorde, Patrick

    2016-03-01

    Although LEDs have been introduced successfully in many general lighting applications during the past decade, high brightness light source applications are still suffering from the limited luminance of LEDs. High power LEDs are generally limited in luminance to ca 100 Mnit (108 lm/m2sr) or less, while dedicated devices for projection may achieve luminance values up to ca 300 Mnit with phosphor converted green. In particular for high luminous flux applications with limited étendue, like in front projection systems, only very modest luminous flux values in the beam can be achieved with LEDs compared to systems based on discharge lamps. In this paper we introduce a light engine concept based on a light converter rod pumped with blue LEDs that breaks through the étendue and brightness limits of LEDs, enabling LED light source luminance values that are more than 4 times higher than what can be achieved with LEDs so far. In LED front projection systems, green LEDs are the main limiting factor. With our green light emitting modules, peak luminance values well above 1.2 Gnit have been achieved, enabling doubling of the screen brightness of LED based DLP projection systems, and even more when this technology is applied to other colors as well. This light source concept, introduced as the ColorSpark High Lumen Density (HLD) LED technology, enables a breakthrough in the performance of LED-based light engines not only for projection, where >2700 ANSI lm was demonstrated, but for a wide variety of high brightness applications.

  5. A highly scalable and high-performance storage architecture for multimedia applications

    NASA Astrophysics Data System (ADS)

    Liu, Zhaobin; Xie, Changsheng; Fu, Xianglin; Cao, Qiang

    2002-12-01

    Due to the excitement of Internet and high bandwidth, there are more and more multimedia applications involving digital industry. However the storage and the real-time of the conventional storage architecture cannot cater for the requirements of continuous media. The most important storage architecture used in past is Direct Attached Storage (DAS) and RAID cabinet, and recently, both Network Attached Storage (NAS) and Storage Area Networks (SAN) are the alterative storage network topology. But as for the multimedia characters, there need more storage capacity and more simultaneous streams. In this paper, we have introduced a novel concept 'Unified Storage Network' (USN) to build efficient SAN over IP, to bridge the gap of NAS and SAN, furthermore to resolve the scalability problem of storage for multimedia applications.

  6. Building and managing high performance, scalable, commodity mass storage systems

    NASA Technical Reports Server (NTRS)

    Lekashman, John

    1998-01-01

    The NAS Systems Division has recently embarked on a significant new way of handling the mass storage problem. One of the basic goals of this new development are to build systems at very large capacity and high performance, yet have the advantages of commodity products. The central design philosophy is to build storage systems the way the Internet was built. Competitive, survivable, expandable, and wide open. The thrust of this paper is to describe the motivation for this effort, what we mean by commodity mass storage, what the implications are for a facility that performs such an action, and where we think it will lead.

  7. Scalable Nearest Neighbor Algorithms for High Dimensional Data.

    PubMed

    Muja, Marius; Lowe, David G

    2014-11-01

    For many computer vision and machine learning problems, large training sets are key for good performance. However, the most computationally expensive part of many computer vision and machine learning algorithms consists of finding nearest neighbor matches to high dimensional vectors that represent the training data. We propose new algorithms for approximate nearest neighbor matching and evaluate and compare them with previous algorithms. For matching high dimensional features, we find two algorithms to be the most efficient: the randomized k-d forest and a new algorithm proposed in this paper, the priority search k-means tree. We also propose a new algorithm for matching binary features by searching multiple hierarchical clustering trees and show it outperforms methods typically used in the literature. We show that the optimal nearest neighbor algorithm and its parameters depend on the data set characteristics and describe an automated configuration procedure for finding the best algorithm to search a particular data set. In order to scale to very large data sets that would otherwise not fit in the memory of a single machine, we propose a distributed nearest neighbor matching framework that can be used with any of the algorithms described in the paper. All this research has been released as an open source library called fast library for approximate nearest neighbors (FLANN), which has been incorporated into OpenCV and is now one of the most popular libraries for nearest neighbor matching. PMID:26353063

  8. Highly scalable digital front end architectures for digital printing

    NASA Astrophysics Data System (ADS)

    Staas, David

    2011-01-01

    HP's digital printing presses consume a tremendous amount of data. The architectures of the Digital Front Ends (DFEs) that feed these large, very fast presses have evolved from basic, single-RIP (Raster Image Processor) systems to multirack, distributed systems that can take a PDF file and deliver data in excess of 3 Gigapixels per second to keep the presses printing at 2000+ pages per minute. This paper highlights some of the more interesting parallelism features of our DFE architectures. The high-performance architecture developed over the last 5+ years can scale up to HP's largest digital press, out to multiple mid-range presses, and down into a very low-cost single box deployment for low-end devices as appropriate. Principles of parallelism pervade every aspect of the architecture, from the lowest-level elements of jobs to parallel imaging pipelines that feed multiple presses. From cores to threads to arrays to network teams to distributed machines, we use a systematic approach to move bottlenecks. The ultimate goals of these efforts are: to take the best advantage of the prevailing hardware options at our disposal; to reduce power consumption and cooling requirements; and to ultimately reduce the cost of the solution to our customers.

  9. Scalable, high performance, enzymatic cathodes based on nanoimprint lithography.

    PubMed

    Pankratov, Dmitry; Sundberg, Richard; Sotres, Javier; Suyatin, Dmitry B; Maximov, Ivan; Shleev, Sergey; Montelius, Lars

    2015-01-01

    Here we detail high performance, enzymatic electrodes for oxygen bio-electroreduction, which can be easily and reproducibly fabricated with industry-scale throughput. Planar and nanostructured electrodes were built on biocompatible, flexible polymer sheets, while nanoimprint lithography was used for electrode nanostructuring. To the best of our knowledge, this is one of the first reports concerning the usage of nanoimprint lithography for amperometric bioelectronic devices. The enzyme (Myrothecium verrucaria bilirubin oxidase) was immobilised on planar (control) and artificially nanostructured, gold electrodes by direct physical adsorption. The detailed electrochemical investigation of bioelectrodes was performed and the following parameters were obtained: open circuit voltage of approximately 0.75 V, and maximum bio-electrocatalytic current densities of 18 µA/cm(2) and 58 µA/cm(2) in air-saturated buffers versus 48 µA/cm(2) and 186 µA/cm(2) in oxygen-saturated buffers for planar and nanostructured electrodes, respectively. The half-deactivation times of planar and nanostructured biocathodes were measured to be 2 h and 14 h, respectively. The comparison of standard heterogeneous and bio-electrocatalytic rate constants showed that the improved bio-electrocatalytic performance of the nanostructured biocathodes compared to planar biodevices is due to the increased surface area of the nanostructured electrodes, whereas their improved operational stability is attributed to stabilisation of the enzyme inside nanocavities. PMID:26199841

  10. Analysis of scalability of high-performance 3D image processing platform for virtual colonoscopy

    NASA Astrophysics Data System (ADS)

    Yoshida, Hiroyuki; Wu, Yin; Cai, Wenli

    2014-03-01

    One of the key challenges in three-dimensional (3D) medical imaging is to enable the fast turn-around time, which is often required for interactive or real-time response. This inevitably requires not only high computational power but also high memory bandwidth due to the massive amount of data that need to be processed. For this purpose, we previously developed a software platform for high-performance 3D medical image processing, called HPC 3D-MIP platform, which employs increasingly available and affordable commodity computing systems such as the multicore, cluster, and cloud computing systems. To achieve scalable high-performance computing, the platform employed size-adaptive, distributable block volumes as a core data structure for efficient parallelization of a wide range of 3D-MIP algorithms, supported task scheduling for efficient load distribution and balancing, and consisted of a layered parallel software libraries that allow image processing applications to share the common functionalities. We evaluated the performance of the HPC 3D-MIP platform by applying it to computationally intensive processes in virtual colonoscopy. Experimental results showed a 12-fold performance improvement on a workstation with 12-core CPUs over the original sequential implementation of the processes, indicating the efficiency of the platform. Analysis of performance scalability based on the Amdahl's law for symmetric multicore chips showed the potential of a high performance scalability of the HPC 3DMIP platform when a larger number of cores is available.

  11. Analysis of scalability of high-performance 3D image processing platform for virtual colonoscopy.

    PubMed

    Yoshida, Hiroyuki; Wu, Yin; Cai, Wenli

    2014-03-19

    One of the key challenges in three-dimensional (3D) medical imaging is to enable the fast turn-around time, which is often required for interactive or real-time response. This inevitably requires not only high computational power but also high memory bandwidth due to the massive amount of data that need to be processed. For this purpose, we previously developed a software platform for high-performance 3D medical image processing, called HPC 3D-MIP platform, which employs increasingly available and affordable commodity computing systems such as the multicore, cluster, and cloud computing systems. To achieve scalable high-performance computing, the platform employed size-adaptive, distributable block volumes as a core data structure for efficient parallelization of a wide range of 3D-MIP algorithms, supported task scheduling for efficient load distribution and balancing, and consisted of a layered parallel software libraries that allow image processing applications to share the common functionalities. We evaluated the performance of the HPC 3D-MIP platform by applying it to computationally intensive processes in virtual colonoscopy. Experimental results showed a 12-fold performance improvement on a workstation with 12-core CPUs over the original sequential implementation of the processes, indicating the efficiency of the platform. Analysis of performance scalability based on the Amdahl's law for symmetric multicore chips showed the potential of a high performance scalability of the HPC 3D-MIP platform when a larger number of cores is available. PMID:24910506

  12. Analysis of scalability of high-performance 3D image processing platform for virtual colonoscopy

    PubMed Central

    Yoshida, Hiroyuki; Wu, Yin; Cai, Wenli

    2014-01-01

    One of the key challenges in three-dimensional (3D) medical imaging is to enable the fast turn-around time, which is often required for interactive or real-time response. This inevitably requires not only high computational power but also high memory bandwidth due to the massive amount of data that need to be processed. For this purpose, we previously developed a software platform for high-performance 3D medical image processing, called HPC 3D-MIP platform, which employs increasingly available and affordable commodity computing systems such as the multicore, cluster, and cloud computing systems. To achieve scalable high-performance computing, the platform employed size-adaptive, distributable block volumes as a core data structure for efficient parallelization of a wide range of 3D-MIP algorithms, supported task scheduling for efficient load distribution and balancing, and consisted of a layered parallel software libraries that allow image processing applications to share the common functionalities. We evaluated the performance of the HPC 3D-MIP platform by applying it to computationally intensive processes in virtual colonoscopy. Experimental results showed a 12-fold performance improvement on a workstation with 12-core CPUs over the original sequential implementation of the processes, indicating the efficiency of the platform. Analysis of performance scalability based on the Amdahl’s law for symmetric multicore chips showed the potential of a high performance scalability of the HPC 3D-MIP platform when a larger number of cores is available. PMID:24910506

  13. Scalable Light Module for Low-Cost, High-Efficiency Light- Emitting Diode Luminaires

    SciTech Connect

    Tarsa, Eric

    2015-08-31

    During this two-year program Cree developed a scalable, modular optical architecture for low-cost, high-efficacy light emitting diode (LED) luminaires. Stated simply, the goal of this architecture was to efficiently and cost-effectively convey light from LEDs (point sources) to broad luminaire surfaces (area sources). By simultaneously developing warm-white LED components and low-cost, scalable optical elements, a high system optical efficiency resulted. To meet program goals, Cree evaluated novel approaches to improve LED component efficacy at high color quality while not sacrificing LED optical efficiency relative to conventional packages. Meanwhile, efficiently coupling light from LEDs into modular optical elements, followed by optimally distributing and extracting this light, were challenges that were addressed via novel optical design coupled with frequent experimental evaluations. Minimizing luminaire bill of materials and assembly costs were two guiding principles for all design work, in the effort to achieve luminaires with significantly lower normalized cost ($/klm) than existing LED fixtures. Chief project accomplishments included the achievement of >150 lm/W warm-white LEDs having primary optics compatible with low-cost modular optical elements. In addition, a prototype Light Module optical efficiency of over 90% was measured, demonstrating the potential of this scalable architecture for ultra-high-efficacy LED luminaires. Since the project ended, Cree has continued to evaluate optical element fabrication and assembly methods in an effort to rapidly transfer this scalable, cost-effective technology to Cree production development groups. The Light Module concept is likely to make a strong contribution to the development of new cost-effective, high-efficacy luminaries, thereby accelerating widespread adoption of energy-saving SSL in the U.S.

  14. Volume-scalable high-brightness three-dimensional visible light source

    SciTech Connect

    Subramania, Ganapathi; Fischer, Arthur J; Wang, George T; Li, Qiming

    2014-02-18

    A volume-scalable, high-brightness, electrically driven visible light source comprises a three-dimensional photonic crystal (3DPC) comprising one or more direct bandgap semiconductors. The improved light emission performance of the invention is achieved based on the enhancement of radiative emission of light emitters placed inside a 3DPC due to the strong modification of the photonic density-of-states engendered by the 3DPC.

  15. Palacios and Kitten : high performance operating systems for scalable virtualized and native supercomputing.

    SciTech Connect

    Widener, Patrick; Jaconette, Steven; Bridges, Patrick G.; Xia, Lei; Dinda, Peter; Cui, Zheng.; Lange, John; Hudson, Trammell B.; Levenhagen, Michael J.; Pedretti, Kevin Thomas Tauke; Brightwell, Ronald Brian

    2009-09-01

    Palacios and Kitten are new open source tools that enable applications, whether ported or not, to achieve scalable high performance on large machines. They provide a thin layer over the hardware to support both full-featured virtualized environments and native code bases. Kitten is an OS under development at Sandia that implements a lightweight kernel architecture to provide predictable behavior and increased flexibility on large machines, while also providing Linux binary compatibility. Palacios is a VMM that is under development at Northwestern University and the University of New Mexico. Palacios, which can be embedded into Kitten and other OSes, supports existing, unmodified applications and operating systems by using virtualization that leverages hardware technologies. We describe the design and implementation of both Kitten and Palacios. Our benchmarks show that they provide near native, scalable performance. Palacios and Kitten provide an incremental path to using supercomputer resources that is not performance-compromised.

  16. Scalable Clustering of High-Dimensional Data Technique Using SPCM with Ant Colony Optimization Intelligence.

    PubMed

    Srinivasan, Thenmozhi; Palanisamy, Balasubramanie

    2015-01-01

    Clusters of high-dimensional data techniques are emerging, according to data noisy and poor quality challenges. This paper has been developed to cluster data using high-dimensional similarity based PCM (SPCM), with ant colony optimization intelligence which is effective in clustering nonspatial data without getting knowledge about cluster number from the user. The PCM becomes similarity based by using mountain method with it. Though this is efficient clustering, it is checked for optimization using ant colony algorithm with swarm intelligence. Thus the scalable clustering technique is obtained and the evaluation results are checked with synthetic datasets. PMID:26495413

  17. Scalable Clustering of High-Dimensional Data Technique Using SPCM with Ant Colony Optimization Intelligence

    PubMed Central

    Srinivasan, Thenmozhi; Palanisamy, Balasubramanie

    2015-01-01

    Clusters of high-dimensional data techniques are emerging, according to data noisy and poor quality challenges. This paper has been developed to cluster data using high-dimensional similarity based PCM (SPCM), with ant colony optimization intelligence which is effective in clustering nonspatial data without getting knowledge about cluster number from the user. The PCM becomes similarity based by using mountain method with it. Though this is efficient clustering, it is checked for optimization using ant colony algorithm with swarm intelligence. Thus the scalable clustering technique is obtained and the evaluation results are checked with synthetic datasets. PMID:26495413

  18. Scalable fabrication of high-quality, ultra-thin single crystal diamond membrane windows

    NASA Astrophysics Data System (ADS)

    Piracha, Afaq Habib; Ganesan, Kumaravelu; Lau, Desmond W. M.; Stacey, Alastair; McGuinness, Liam P.; Tomljenovic-Hanic, Snjezana; Prawer, Steven

    2016-03-01

    High quality, ultra-thin single crystal diamond (SCD) membranes that have a thickness in the sub-micron range are of extreme importance as a materials platform for photonics, quantum sensing, nano/micro electro-mechanical systems (N/MEMS) and other diverse applications. However, the scalable fabrication of such thin SCD membranes is a challenging process. In this paper, we demonstrate a new method which enables high quality, large size (~4 × 4 mm) and low surface roughness, low strain, ultra-thin SCD membranes which can be fabricated without deformations such as breakage, bowing or bending. These membranes are easy to handle making them particularly suitable for fabrication of optical and mechanical devices. We demonstrate arrays of single crystal diamond membrane windows (SCDMW), each up to 1 × 1 mm in dimension and as thin as ~300 nm, supported by a diamond frame as thick as ~150 μm. The fabrication method is robust, reproducible, scalable and cost effective. Microwave plasma chemical vapour deposition is used for in situ creation of single nitrogen-vacancy (NV) centers into the thin SCDMW. We have also developed SCD drum head mechanical resonator composed of our fully clamped and freely suspended membranes.High quality, ultra-thin single crystal diamond (SCD) membranes that have a thickness in the sub-micron range are of extreme importance as a materials platform for photonics, quantum sensing, nano/micro electro-mechanical systems (N/MEMS) and other diverse applications. However, the scalable fabrication of such thin SCD membranes is a challenging process. In this paper, we demonstrate a new method which enables high quality, large size (~4 × 4 mm) and low surface roughness, low strain, ultra-thin SCD membranes which can be fabricated without deformations such as breakage, bowing or bending. These membranes are easy to handle making them particularly suitable for fabrication of optical and mechanical devices. We demonstrate arrays of single crystal diamond

  19. A scalable, high resolution strain sensing matrix suitable for tactile transduction.

    PubMed

    Scibelli, Anthony E; Krans, Jacob L

    2016-02-01

    The integration of tactile information, such as contact area, displacement magnitude, velocity, and acceleration, is paramount to the optimization of robotics in human-centric environments. Cost effective embeddable sensors with scalable receptive field size and strain sensitivity are not readily commercially available and would benefit investigations of in situ tissue mechanics. We describe the design and performance of a scalable sensor matrix that transduces fine parameters of strain and is made of combinable "modules". The sensors transduce static and dynamic strains of both uniaxial and multi-dimensional nature. Modules consist of three silicon wafers placed on top of and three on the bottom of a hexagonal collar, wafers are thus positioned 120° to one another to facilitate force vector extrapolation. Analog signals from each sensor can be easily compared to neighboring sensor output to determine mechanical phenomena such as slip or shear. The smallest of our prototype multiunit matrices consisted of seven hexes in a honeycomb orientation of 4.1mm diameter (containing 42 silicon gauges). Unamplified, unshielded output from this embodiment (3 Vexc button cell) yielded 1 mV from 5 μm displacement. Transduction linearity was high (R>0.99 nearest displacement) and exhibited nominal hysteresis. Modules may be placed upon or embedded into a multitude of materials and the size of individual hexagons may be scaled for favorable stiffness to strain ratio and to scale receptive field. Given the scalability of matrix size and resolution, we believe the sensor matrices could benefit the fields of prosthetics, robotics, and physiologic investigation of tissue mechanics. PMID:26710986

  20. Lilith: A Java framework for the development of scalable tools for high performance distributed computing platforms

    SciTech Connect

    Evensky, D.A.; Gentile, A.C.; Armstrong, R.C.

    1998-03-19

    Increasingly, high performance computing constitutes the use of very large heterogeneous clusters of machines. The use and maintenance of such clusters are subject to complexities of communication between the machines in a time efficient and secure manner. Lilith is a general purpose tool that provides a highly scalable, secure, and easy distribution of user code across a heterogeneous computing platform. By handling the details of code distribution and communication, such a framework allows for the rapid development of tools for the use and management of large distributed systems. Lilith is written in Java, taking advantage of Java`s unique features of loading and distributing code dynamically, its platform independence, its thread support, and its provision of graphical components to facilitate easy to use resultant tools. The authors describe the use of Lilith in a tool developed for the maintenance of the large distributed cluster at their institution and present details of the Lilith architecture and user API for the general user development of scalable tools.

  1. Scalable fabrication of high-quality, ultra-thin single crystal diamond membrane windows.

    PubMed

    Piracha, Afaq Habib; Ganesan, Kumaravelu; Lau, Desmond W M; Stacey, Alastair; McGuinness, Liam P; Tomljenovic-Hanic, Snjezana; Prawer, Steven

    2016-03-28

    High quality, ultra-thin single crystal diamond (SCD) membranes that have a thickness in the sub-micron range are of extreme importance as a materials platform for photonics, quantum sensing, nano/micro electro-mechanical systems (N/MEMS) and other diverse applications. However, the scalable fabrication of such thin SCD membranes is a challenging process. In this paper, we demonstrate a new method which enables high quality, large size (∼4 × 4 mm) and low surface roughness, low strain, ultra-thin SCD membranes which can be fabricated without deformations such as breakage, bowing or bending. These membranes are easy to handle making them particularly suitable for fabrication of optical and mechanical devices. We demonstrate arrays of single crystal diamond membrane windows (SCDMW), each up to 1 × 1 mm in dimension and as thin as ∼300 nm, supported by a diamond frame as thick as ∼150 μm. The fabrication method is robust, reproducible, scalable and cost effective. Microwave plasma chemical vapour deposition is used for in situ creation of single nitrogen-vacancy (NV) centers into the thin SCDMW. We have also developed SCD drum head mechanical resonator composed of our fully clamped and freely suspended membranes. PMID:26956525

  2. The Open Connectome Project Data Cluster: Scalable Analysis and Vision for High-Throughput Neuroscience.

    PubMed

    Burns, Randal; Roncal, William Gray; Kleissas, Dean; Lillaney, Kunal; Manavalan, Priya; Perlman, Eric; Berger, Daniel R; Bock, Davi D; Chung, Kwanghun; Grosenick, Logan; Kasthuri, Narayanan; Weiler, Nicholas C; Deisseroth, Karl; Kazhdan, Michael; Lichtman, Jeff; Reid, R Clay; Smith, Stephen J; Szalay, Alexander S; Vogelstein, Joshua T; Vogelstein, R Jacob

    2013-01-01

    We describe a scalable database cluster for the spatial analysis and annotation of high-throughput brain imaging data, initially for 3-d electron microscopy image stacks, but for time-series and multi-channel data as well. The system was designed primarily for workloads that build connectomes- neural connectivity maps of the brain-using the parallel execution of computer vision algorithms on high-performance compute clusters. These services and open-science data sets are publicly available at openconnecto.me. The system design inherits much from NoSQL scale-out and data-intensive computing architectures. We distribute data to cluster nodes by partitioning a spatial index. We direct I/O to different systems-reads to parallel disk arrays and writes to solid-state storage-to avoid I/O interference and maximize throughput. All programming interfaces are RESTful Web services, which are simple and stateless, improving scalability and usability. We include a performance evaluation of the production system, highlighting the effec-tiveness of spatial data organization. PMID:24401992

  3. Thermally efficient and highly scalable In2Se3 nanowire phase change memory

    NASA Astrophysics Data System (ADS)

    Jin, Bo; Kang, Daegun; Kim, Jungsik; Meyyappan, M.; Lee, Jeong-Soo

    2013-04-01

    The electrical characteristics of nonvolatile In2Se3 nanowire phase change memory are reported. Size-dependent memory switching behavior was observed in nanowires of varying diameters and the reduction in set/reset threshold voltage was as low as 3.45 V/6.25 V for a 60 nm nanowire, which is promising for highly scalable nanowire memory applications. Also, size-dependent thermal resistance of In2Se3 nanowire memory cells was estimated with values as high as 5.86×1013 and 1.04×106 K/W for a 60 nm nanowire memory cell in amorphous and crystalline phases, respectively. Such high thermal resistances are beneficial for improvement of thermal efficiency and thus reduction in programming power consumption based on Fourier's law. The evaluation of thermal resistance provides an avenue to develop thermally efficient memory cell architecture.

  4. Inter-Layer Prediction of Color in High Dynamic Range Image Scalable Compression.

    PubMed

    Le Pendu, Mikael; Guillemot, Christine; Thoreau, Dominique

    2016-08-01

    This paper presents a color inter-layer prediction (ILP) method for scalable coding of high dynamic range (HDR) video content with a low dynamic range (LDR) base layer. Relying on the assumption of hue preservation between the colors of an HDR image and its LDR tone mapped version, we derived equations for predicting the chromatic components of the HDR layer given the decoded LDR layer. Two color representations are studied. In a first encoding scheme, the HDR image is represented in the classical Y'CbCr format. In addition, a second scheme is proposed using a colorspace based on the CIE u'v' uniform chromaticity scale diagram. In each case, different prediction equations are derived based on a color model ensuring the hue preservation. Our experiments highlight several advantages of using a CIE u'v'-based colorspace for the compression of HDR content, especially in a scalable context. In addition, our ILP scheme using this color representation improves on the state-of-the-art ILP method, which directly predicts the HDR layer u'v' components by computing the LDR layers u'v' values of each pixel. PMID:27244738

  5. Scalable Growth of High Mobility Dirac Semimetal Cd3As2 Microbelts.

    PubMed

    Chen, Zhi-Gang; Zhang, Cheng; Zou, Yichao; Zhang, Enze; Yang, Lei; Hong, Min; Xiu, Faxian; Zou, Jin

    2015-09-01

    Three-dimensional (3D) Dirac semimetals are 3D analogues of graphene, which display Dirac points with linear dispersion in k-space, stabilized by crystal symmetry. Cd3As2 has been predicted to be 3D Dirac semimetals and was subsequently demonstrated by angle-resolved photoemission spectroscopy. As unveiled by transport measurements, several exotic phases, such as Weyl semimetals, topological insulators, and topological superconductors, can be deduced by breaking time reversal or inversion symmetry. Here, we reported a facile and scalable chemical vapor deposition method to fabricate high-quality Dirac semimetal Cd3As2 microbelts; they have shown ultrahigh mobility up to 1.15 × 10(5) cm(2) V(-1) s(-1) and pronounced Shubnikov-de Haas oscillations. Such extraordinary features are attributed to the suppression of electron backscattering. This research opens a new avenue for the scalable fabrication of Cd3As2 materials toward exciting electronic applications of 3D Dirac semimetals. PMID:26305792

  6. Simple, Scalable Proteomic Imaging for High-Dimensional Profiling of Intact Systems.

    PubMed

    Murray, Evan; Cho, Jae Hun; Goodwin, Daniel; Ku, Taeyun; Swaney, Justin; Kim, Sung-Yon; Choi, Heejin; Park, Young-Gyun; Park, Jeong-Yoon; Hubbert, Austin; McCue, Margaret; Vassallo, Sara; Bakh, Naveed; Frosch, Matthew P; Wedeen, Van J; Seung, H Sebastian; Chung, Kwanghun

    2015-12-01

    Combined measurement of diverse molecular and anatomical traits that span multiple levels remains a major challenge in biology. Here, we introduce a simple method that enables proteomic imaging for scalable, integrated, high-dimensional phenotyping of both animal tissues and human clinical samples. This method, termed SWITCH, uniformly secures tissue architecture, native biomolecules, and antigenicity across an entire system by synchronizing the tissue preservation reaction. The heat- and chemical-resistant nature of the resulting framework permits multiple rounds (>20) of relabeling. We have performed 22 rounds of labeling of a single tissue with precise co-registration of multiple datasets. Furthermore, SWITCH synchronizes labeling reactions to improve probe penetration depth and uniformity of staining. With SWITCH, we performed combinatorial protein expression profiling of the human cortex and also interrogated the geometric structure of the fiber pathways in mouse brains. Such integrated high-dimensional information may accelerate our understanding of biological systems at multiple levels. PMID:26638076

  7. A scalable silicon photonic chip-scale optical switch for high performance computing systems.

    PubMed

    Yu, Runxiang; Cheung, Stanley; Li, Yuliang; Okamoto, Katsunari; Proietti, Roberto; Yin, Yawei; Yoo, S J B

    2013-12-30

    This paper discusses the architecture and provides performance studies of a silicon photonic chip-scale optical switch for scalable interconnect network in high performance computing systems. The proposed switch exploits optical wavelength parallelism and wavelength routing characteristics of an Arrayed Waveguide Grating Router (AWGR) to allow contention resolution in the wavelength domain. Simulation results from a cycle-accurate network simulator indicate that, even with only two transmitter/receiver pairs per node, the switch exhibits lower end-to-end latency and higher throughput at high (>90%) input loads compared with electronic switches. On the device integration level, we propose to integrate all the components (ring modulators, photodetectors and AWGR) on a CMOS-compatible silicon photonic platform to ensure a compact, energy efficient and cost-effective device. We successfully demonstrate proof-of-concept routing functions on an 8 × 8 prototype fabricated using foundry services provided by OpSIS-IME. PMID:24514859

  8. Muster: Massively Scalable Clustering

    Energy Science and Technology Software Center (ESTSC)

    2010-05-20

    Muster is a framework for scalable cluster analysis. It includes implementations of classic K-Medoids partitioning algorithms, as well as infrastructure for making these algorithms run scalably on very large systems. In particular, Muster contains algorithms such as CAPEK (described in reference 1) that are capable of clustering highly distributed data sets in-place on a hundred thousand or more processes.

  9. Scalable High Performance Message Passing over InfiniBand for Open MPI

    SciTech Connect

    Friedley, A; Hoefler, T; Leininger, M L; Lumsdaine, A

    2007-10-24

    InfiniBand (IB) is a popular network technology for modern high-performance computing systems. MPI implementations traditionally support IB using a reliable, connection-oriented (RC) transport. However, per-process resource usage that grows linearly with the number of processes, makes this approach prohibitive for large-scale systems. IB provides an alternative in the form of a connectionless unreliable datagram transport (UD), which allows for near-constant resource usage and initialization overhead as the process count increases. This paper describes a UD-based implementation for IB in Open MPI as a scalable alternative to existing RC-based schemes. We use the software reliability capabilities of Open MPI to provide the guaranteed delivery semantics required by MPI. Results show that UD not only requires fewer resources at scale, but also allows for shorter MPI startup times. A connectionless model also improves performance for applications that tend to send small messages to many different processes.

  10. Highly Efficient and Scalable Separation of Semiconducting Carbon Nanotubes via Weak Field Centrifugation.

    PubMed

    Reis, Wieland G; Weitz, R Thomas; Kettner, Michel; Kraus, Alexander; Schwab, Matthias Georg; Tomović, Željko; Krupke, Ralph; Mikhael, Jules

    2016-01-01

    The identification of scalable processes that transfer random mixtures of single-walled carbon nanotubes (SWCNTs) into fractions featuring a high content of semiconducting species is crucial for future application of SWCNTs in high-performance electronics. Herein we demonstrate a highly efficient and simple separation method that relies on selective interactions between tailor-made amphiphilic polymers and semiconducting SWCNTs in the presence of low viscosity separation media. High purity individualized semiconducting SWCNTs or even self-organized semiconducting sheets are separated from an as-produced SWCNT dispersion via a single weak field centrifugation run. Absorption and Raman spectroscopy are applied to verify the high purity of the obtained SWCNTs. Furthermore SWCNT - network field-effect transistors were fabricated, which exhibit high ON/OFF ratios (10(5)) and field-effect mobilities (17 cm(2)/Vs). In addition to demonstrating the feasibility of high purity separation by a novel low complexity process, our method can be readily transferred to large scale production. PMID:27188435

  11. Highly Efficient and Scalable Separation of Semiconducting Carbon Nanotubes via Weak Field Centrifugation

    NASA Astrophysics Data System (ADS)

    Reis, Wieland G.; Weitz, R. Thomas; Kettner, Michel; Kraus, Alexander; Schwab, Matthias Georg; Tomović, Željko; Krupke, Ralph; Mikhael, Jules

    2016-05-01

    The identification of scalable processes that transfer random mixtures of single-walled carbon nanotubes (SWCNTs) into fractions featuring a high content of semiconducting species is crucial for future application of SWCNTs in high-performance electronics. Herein we demonstrate a highly efficient and simple separation method that relies on selective interactions between tailor-made amphiphilic polymers and semiconducting SWCNTs in the presence of low viscosity separation media. High purity individualized semiconducting SWCNTs or even self-organized semiconducting sheets are separated from an as-produced SWCNT dispersion via a single weak field centrifugation run. Absorption and Raman spectroscopy are applied to verify the high purity of the obtained SWCNTs. Furthermore SWCNT - network field-effect transistors were fabricated, which exhibit high ON/OFF ratios (105) and field-effect mobilities (17 cm2/Vs). In addition to demonstrating the feasibility of high purity separation by a novel low complexity process, our method can be readily transferred to large scale production.

  12. Highly Efficient and Scalable Separation of Semiconducting Carbon Nanotubes via Weak Field Centrifugation

    PubMed Central

    Reis, Wieland G.; Weitz, R. Thomas; Kettner, Michel; Kraus, Alexander; Schwab, Matthias Georg; Tomović, Željko; Krupke, Ralph; Mikhael, Jules

    2016-01-01

    The identification of scalable processes that transfer random mixtures of single-walled carbon nanotubes (SWCNTs) into fractions featuring a high content of semiconducting species is crucial for future application of SWCNTs in high-performance electronics. Herein we demonstrate a highly efficient and simple separation method that relies on selective interactions between tailor-made amphiphilic polymers and semiconducting SWCNTs in the presence of low viscosity separation media. High purity individualized semiconducting SWCNTs or even self-organized semiconducting sheets are separated from an as-produced SWCNT dispersion via a single weak field centrifugation run. Absorption and Raman spectroscopy are applied to verify the high purity of the obtained SWCNTs. Furthermore SWCNT - network field-effect transistors were fabricated, which exhibit high ON/OFF ratios (105) and field-effect mobilities (17 cm2/Vs). In addition to demonstrating the feasibility of high purity separation by a novel low complexity process, our method can be readily transferred to large scale production. PMID:27188435

  13. Scalable Sub-micron Patterning of Organic Materials Toward High Density Soft Electronics

    PubMed Central

    Kim, Jaekyun; Kim, Myung-Gil; Kim, Jaehyun; Jo, Sangho; Kang, Jingu; Jo, Jeong-Wan; Lee, Woobin; Hwang, Chahwan; Moon, Juhyuk; Yang, Lin; Kim, Yun-Hi; Noh, Yong-Young; Yun Jaung, Jae; Kim, Yong-Hoon; Kyu Park, Sung

    2015-01-01

    The success of silicon based high density integrated circuits ignited explosive expansion of microelectronics. Although the inorganic semiconductors have shown superior carrier mobilities for conventional high speed switching devices, the emergence of unconventional applications, such as flexible electronics, highly sensitive photosensors, large area sensor array, and tailored optoelectronics, brought intensive research on next generation electronic materials. The rationally designed multifunctional soft electronic materials, organic and carbon-based semiconductors, are demonstrated with low-cost solution process, exceptional mechanical stability, and on-demand optoelectronic properties. Unfortunately, the industrial implementation of the soft electronic materials has been hindered due to lack of scalable fine-patterning methods. In this report, we demonstrated facile general route for high throughput sub-micron patterning of soft materials, using spatially selective deep-ultraviolet irradiation. For organic and carbon-based materials, the highly energetic photons (e.g. deep-ultraviolet rays) enable direct photo-conversion from conducting/semiconducting to insulating state through molecular dissociation and disordering with spatial resolution down to a sub-μm-scale. The successful demonstration of organic semiconductor circuitry promise our result proliferate industrial adoption of soft materials for next generation electronics. PMID:26411932

  14. Scalable sub-micron patterning of organic materials toward high density soft electronics

    SciTech Connect

    Kim, Jaekyun; Kim, Myung -Gil; Kim, Jaehyun; Jo, Sangho; Kang, Jingu; Jo, Jeong -Wan; Lee, Woobin; Hwang, Chahwan; Moon, Juhyuk; Yang, Lin; Kim, Yun -Hi; Noh, Yong -Young; Yun Jaung, Jae; Kim, Yong -Hoon; Kyu Park, Sung

    2015-09-28

    The success of silicon based high density integrated circuits ignited explosive expansion of microelectronics. Although the inorganic semiconductors have shown superior carrier mobilities for conventional high speed switching devices, the emergence of unconventional applications, such as flexible electronics, highly sensitive photosensors, large area sensor array, and tailored optoelectronics, brought intensive research on next generation electronic materials. The rationally designed multifunctional soft electronic materials, organic and carbon-based semiconductors, are demonstrated with low-cost solution process, exceptional mechanical stability, and on-demand optoelectronic properties. Unfortunately, the industrial implementation of the soft electronic materials has been hindered due to lack of scalable fine-patterning methods. In this report, we demonstrated facile general route for high throughput sub-micron patterning of soft materials, using spatially selective deep-ultraviolet irradiation. For organic and carbon-based materials, the highly energetic photons (e.g. deep-ultraviolet rays) enable direct photo-conversion from conducting/semiconducting to insulating state through molecular dissociation and disordering with spatial resolution down to a sub-μm-scale. As a result, the successful demonstration of organic semiconductor circuitry promise our result proliferate industrial adoption of soft materials for next generation electronics.

  15. Scalable sub-micron patterning of organic materials toward high density soft electronics

    DOE PAGESBeta

    Kim, Jaekyun; Kim, Myung -Gil; Kim, Jaehyun; Jo, Sangho; Kang, Jingu; Jo, Jeong -Wan; Lee, Woobin; Hwang, Chahwan; Moon, Juhyuk; Yang, Lin; et al

    2015-09-28

    The success of silicon based high density integrated circuits ignited explosive expansion of microelectronics. Although the inorganic semiconductors have shown superior carrier mobilities for conventional high speed switching devices, the emergence of unconventional applications, such as flexible electronics, highly sensitive photosensors, large area sensor array, and tailored optoelectronics, brought intensive research on next generation electronic materials. The rationally designed multifunctional soft electronic materials, organic and carbon-based semiconductors, are demonstrated with low-cost solution process, exceptional mechanical stability, and on-demand optoelectronic properties. Unfortunately, the industrial implementation of the soft electronic materials has been hindered due to lack of scalable fine-patterning methods. Inmore » this report, we demonstrated facile general route for high throughput sub-micron patterning of soft materials, using spatially selective deep-ultraviolet irradiation. For organic and carbon-based materials, the highly energetic photons (e.g. deep-ultraviolet rays) enable direct photo-conversion from conducting/semiconducting to insulating state through molecular dissociation and disordering with spatial resolution down to a sub-μm-scale. As a result, the successful demonstration of organic semiconductor circuitry promise our result proliferate industrial adoption of soft materials for next generation electronics.« less

  16. Scalable Sub-micron Patterning of Organic Materials Toward High Density Soft Electronics

    NASA Astrophysics Data System (ADS)

    Kim, Jaekyun; Kim, Myung-Gil; Kim, Jaehyun; Jo, Sangho; Kang, Jingu; Jo, Jeong-Wan; Lee, Woobin; Hwang, Chahwan; Moon, Juhyuk; Yang, Lin; Kim, Yun-Hi; Noh, Yong-Young; Yun Jaung, Jae; Kim, Yong-Hoon; Kyu Park, Sung

    2015-09-01

    The success of silicon based high density integrated circuits ignited explosive expansion of microelectronics. Although the inorganic semiconductors have shown superior carrier mobilities for conventional high speed switching devices, the emergence of unconventional applications, such as flexible electronics, highly sensitive photosensors, large area sensor array, and tailored optoelectronics, brought intensive research on next generation electronic materials. The rationally designed multifunctional soft electronic materials, organic and carbon-based semiconductors, are demonstrated with low-cost solution process, exceptional mechanical stability, and on-demand optoelectronic properties. Unfortunately, the industrial implementation of the soft electronic materials has been hindered due to lack of scalable fine-patterning methods. In this report, we demonstrated facile general route for high throughput sub-micron patterning of soft materials, using spatially selective deep-ultraviolet irradiation. For organic and carbon-based materials, the highly energetic photons (e.g. deep-ultraviolet rays) enable direct photo-conversion from conducting/semiconducting to insulating state through molecular dissociation and disordering with spatial resolution down to a sub-μm-scale. The successful demonstration of organic semiconductor circuitry promise our result proliferate industrial adoption of soft materials for next generation electronics.

  17. Scalable Sub-micron Patterning of Organic Materials Toward High Density Soft Electronics.

    PubMed

    Kim, Jaekyun; Kim, Myung-Gil; Kim, Jaehyun; Jo, Sangho; Kang, Jingu; Jo, Jeong-Wan; Lee, Woobin; Hwang, Chahwan; Moon, Juhyuk; Yang, Lin; Kim, Yun-Hi; Noh, Yong-Young; Jaung, Jae Yun; Kim, Yong-Hoon; Park, Sung Kyu

    2015-01-01

    The success of silicon based high density integrated circuits ignited explosive expansion of microelectronics. Although the inorganic semiconductors have shown superior carrier mobilities for conventional high speed switching devices, the emergence of unconventional applications, such as flexible electronics, highly sensitive photosensors, large area sensor array, and tailored optoelectronics, brought intensive research on next generation electronic materials. The rationally designed multifunctional soft electronic materials, organic and carbon-based semiconductors, are demonstrated with low-cost solution process, exceptional mechanical stability, and on-demand optoelectronic properties. Unfortunately, the industrial implementation of the soft electronic materials has been hindered due to lack of scalable fine-patterning methods. In this report, we demonstrated facile general route for high throughput sub-micron patterning of soft materials, using spatially selective deep-ultraviolet irradiation. For organic and carbon-based materials, the highly energetic photons (e.g. deep-ultraviolet rays) enable direct photo-conversion from conducting/semiconducting to insulating state through molecular dissociation and disordering with spatial resolution down to a sub-μm-scale. The successful demonstration of organic semiconductor circuitry promise our result proliferate industrial adoption of soft materials for next generation electronics. PMID:26411932

  18. Scalable Functionalized Graphene Nano-platelets as Tunable Cathodes for High-performance Lithium Rechargeable Batteries

    PubMed Central

    Kim, Haegyeom; Lim, Hee-Dae; Kim, Sung-Wook; Hong, Jihyun; Seo, Dong-Hwa; Kim, Dae-chul; Jeon, Seokwoo; Park, Sungjin; Kang, Kisuk

    2013-01-01

    High-performance and cost-effective rechargeable batteries are key to the success of electric vehicles and large-scale energy storage systems. Extensive research has focused on the development of (i) new high-energy electrodes that can store more lithium or (ii) high-power nano-structured electrodes hybridized with carbonaceous materials. However, the current status of lithium batteries based on redox reactions of heavy transition metals still remains far below the demands required for the proposed applications. Herein, we present a novel approach using tunable functional groups on graphene nano-platelets as redox centers. The electrode can deliver high capacity of ~250 mAh g−1, power of ~20 kW kg−1 in an acceptable cathode voltage range, and provide excellent cyclability up to thousands of repeated charge/discharge cycles. The simple, mass-scalable synthetic route for the functionalized graphene nano-platelets proposed in this work suggests that the graphene cathode can be a promising new class of electrode. PMID:23514953

  19. Scalable, High-performance 3D Imaging Software Platform: System Architecture and Application to Virtual Colonoscopy

    PubMed Central

    Yoshida, Hiroyuki; Wu, Yin; Cai, Wenli; Brett, Bevin

    2013-01-01

    One of the key challenges in three-dimensional (3D) medical imaging is to enable the fast turn-around time, which is often required for interactive or real-time response. This inevitably requires not only high computational power but also high memory bandwidth due to the massive amount of data that need to be processed. In this work, we have developed a software platform that is designed to support high-performance 3D medical image processing for a wide range of applications using increasingly available and affordable commodity computing systems: multi-core, clusters, and cloud computing systems. To achieve scalable, high-performance computing, our platform (1) employs size-adaptive, distributable block volumes as a core data structure for efficient parallelization of a wide range of 3D image processing algorithms; (2) supports task scheduling for efficient load distribution and balancing; and (3) consists of a layered parallel software libraries that allow a wide range of medical applications to share the same functionalities. We evaluated the performance of our platform by applying it to an electronic cleansing system in virtual colonoscopy, with initial experimental results showing a 10 times performance improvement on an 8-core workstation over the original sequential implementation of the system. PMID:23366803

  20. XGet: a highly scalable and efficient file transfer tool for clusters

    SciTech Connect

    Greenberg, Hugh; Ionkov, Latchesar; Minnich, Ronald

    2008-01-01

    As clusters rapidly grow in size, transferring files between nodes can no longer be solved by the traditional transfer utilities due to their inherent lack of scalability. In this paper, we describe a new file transfer utility called XGet, which was designed to address the scalability problem of standard tools. We compared XGet against four transfer tools: Bittorrent, Rsync, TFTP, and Udpcast and our results show that XGet's performance is superior to the these utilities in many cases.

  1. Concept and experimental implementation of a scalable high power and highly homogeneous laser line generator for industrial applications

    NASA Astrophysics Data System (ADS)

    Brodner, M.; Bayer, A.; Meinschien, J.

    2011-03-01

    High power diode laser line generators are nowadays industrial standard for applications like plastic processing, vision inspection and drying. With increased beam quality, especially peak intensity and homogeneity, they also enable new applications like hardening, annealing or cutting of various materials. All of these applications have in common that simultaneous processing is limited by the scalability of the generated line length without changing process relevant parameters of the line like working distance, peak intensity, homogeneity and depth of focus. Therefore, a patent pending beam shaping concept is presented that enables the interconnection of an arbitrary number of nearly free selectable laser sources to generate scalable laser lines with outstanding beam parameters. System design, experimental setup and results of a laser line generator are shown. It is based on a stitching concept consisting of ten fibre coupled high power diode lasers, which generates a 200mm long and 2mm wide laser line with a homogeneity level of 97% p-v over a depth of focus of +/- 5 mm with an overall output power of up to 4.2 kW. The concept is discussed regarding industrial applications and the options for even higher beam quality, especially the capability of generating lines with increased power densities up to several kW/cm² and a line length of several meters.

  2. Prodigious Effects of Concentration Intensification on Nanoparticle Synthesis: A High-Quality, Scalable Approach.

    PubMed

    Williamson, Curtis B; Nevers, Douglas R; Hanrath, Tobias; Robinson, Richard D

    2015-12-23

    Realizing the promise of nanoparticle-based technologies demands more efficient, robust synthesis methods (i.e., process intensification) that consistently produce large quantities of high-quality nanoparticles (NPs). We explored NP synthesis via the heat-up method in a regime of previously unexplored high concentrations near the solubility limit of the precursors. We discovered that in this highly concentrated and viscous regime the NP synthesis parameters are less sensitive to experimental variability and thereby provide a robust, scalable, and size-focusing NP synthesis. Specifically, we synthesize high-quality metal sulfide NPs (<7% relative standard deviation for Cu2-xS and CdS), and demonstrate a 10-1000-fold increase in Cu2-xS NP production (>200 g) relative to the current field of large-scale (0.1-5 g yields) and laboratory-scale (<0.1 g) efforts. Compared to conventional synthesis methods (hot injection with dilute precursor concentration) characterized by rapid growth and low yield, our highly concentrated NP system supplies remarkably controlled growth rates and a 10-fold increase in NP volumetric production capacity (86 g/L). The controlled growth, high yield, and robust nature of highly concentrated solutions can facilitate large-scale nanomanufacturing of NPs by relaxing the synthesis requirements to achieve monodisperse products. Mechanistically, our investigation of the thermal and rheological properties and growth rates reveals that this high concentration regime has reduced mass diffusion (a 5-fold increase in solution viscosity), is stable to thermal perturbations (∼64% increase in heat capacity), and is resistant to Ostwald ripening. PMID:26592380

  3. Scalable parallel programming for high performance seismic simulation on petascale heterogeneous supercomputers

    NASA Astrophysics Data System (ADS)

    Zhou, Jun

    The 1994 Northridge earthquake in Los Angeles, California, killed 57 people, injured over 8,700 and caused an estimated $20 billion in damage. Petascale simulations are needed in California and elsewhere to provide society with a better understanding of the rupture and wave dynamics of the largest earthquakes at shaking frequencies required to engineer safe structures. As the heterogeneous supercomputing infrastructures are becoming more common, numerical developments in earthquake system research are particularly challenged by the dependence on the accelerator elements to enable "the Big One" simulations with higher frequency and finer resolution. Reducing time to solution and power consumption are two primary focus area today for the enabling technology of fault rupture dynamics and seismic wave propagation in realistic 3D models of the crust's heterogeneous structure. This dissertation presents scalable parallel programming techniques for high performance seismic simulation running on petascale heterogeneous supercomputers. A real world earthquake simulation code, AWP-ODC, one of the most advanced earthquake codes to date, was chosen as the base code in this research, and the testbed is based on Titan at Oak Ridge National Laboraratory, the world's largest hetergeneous supercomputer. The research work is primarily related to architecture study, computation performance tuning and software system scalability. An earthquake simulation workflow has also been developed to support the efficient production sets of simulations. The highlights of the technical development are an aggressive performance optimization focusing on data locality and a notable data communication model that hides the data communication latency. This development results in the optimal computation efficiency and throughput for the 13-point stencil code on heterogeneous systems, which can be extended to general high-order stencil codes. Started from scratch, the hybrid CPU/GPU version of AWP

  4. Scalable Fabrication of Metal Oxide Functional Materials and Their Applications in High-Temperature Optical Sensing

    NASA Astrophysics Data System (ADS)

    Yan, Aidong; Poole, Zsolt L.; Chen, Rongzhang; Leu, Paul W.; Ohodnicki, Paul; Chen, Kevin P.

    2015-01-01

    We report a scalable manufacturing approach to produce nano-porous metal oxide films and the dopant variants using a block-copolymer template combined with a sol-gel solution processing approach. The refractive index of the film can be tailored to 1.2-2.4 by 3D nanostructuring in the sub-wavelength regime at scales of 20 nm or less. Based on this approach, this paper reports the synthesis of nanoporous palladium (Pd)-doped titanium dioxide (TiO2) film with refractive index matching the optical fiber material, and its importance on D-shaped fiber Bragg grating for hydrogen sensing at extremely high temperature up to 700°C. The sensor is based on evanescent field interaction in hydrogen-sensitive cladding. The flat side of D-shaped fiber grating was etched to remove a residual 4 μm cladding material, and thermally stabilized for high-temperature requirements. The peak intensity change of the fiber Bragg wavelength was observed with different hydrogen concentrations from 0.25 vol.% H2/N2 to 5 vol.% H2/N2. The experimental result shows that the sensor's hydrogen response is reversible and fast. The response time of the hydrogen sensor is <8 s.

  5. High-performance graphene-based supercapacitors made by a scalable blade-coating approach.

    PubMed

    Wang, Bin; Liu, Jinzhang; Mirri, Francesca; Pasquali, Matteo; Motta, Nunzio; Holmes, John W

    2016-04-22

    Graphene oxide (GO) sheets can form liquid crystals (LCs) in their aqueous dispersions that are more viscous with a stronger LC feature. In this work we combine the viscous LC-GO solution with the blade-coating technique to make GO films, for constructing graphene-based supercapacitors in a scalable way. Reduced GO (rGO) films are prepared by wet chemical methods, using either hydrazine (HZ) or hydroiodic acid (HI). Solid-state supercapacitors with rGO films as electrodes and highly conductive carbon nanotube films as current collectors are fabricated and the capacitive properties of different rGO films are compared. It is found that the HZ-rGO film is superior to the HI-rGO film in achieving high capacitance, owing to the 3D structure of graphene sheets in the electrode. Compared to gelled electrolyte, the use of liquid electrolyte (H2SO4) can further increase the capacitance to 265 F per gram (corresponding to 52 mF per cm(2)) of the HZ-rGO film. PMID:26953864

  6. High-performance graphene-based supercapacitors made by a scalable blade-coating approach

    NASA Astrophysics Data System (ADS)

    Wang, Bin; Liu, Jinzhang; Mirri, Francesca; Pasquali, Matteo; Motta, Nunzio; Holmes, John W.

    2016-04-01

    Graphene oxide (GO) sheets can form liquid crystals (LCs) in their aqueous dispersions that are more viscous with a stronger LC feature. In this work we combine the viscous LC-GO solution with the blade-coating technique to make GO films, for constructing graphene-based supercapacitors in a scalable way. Reduced GO (rGO) films are prepared by wet chemical methods, using either hydrazine (HZ) or hydroiodic acid (HI). Solid-state supercapacitors with rGO films as electrodes and highly conductive carbon nanotube films as current collectors are fabricated and the capacitive properties of different rGO films are compared. It is found that the HZ-rGO film is superior to the HI-rGO film in achieving high capacitance, owing to the 3D structure of graphene sheets in the electrode. Compared to gelled electrolyte, the use of liquid electrolyte (H2SO4) can further increase the capacitance to 265 F per gram (corresponding to 52 mF per cm2) of the HZ-rGO film.

  7. Very High Resolution Mapping of Tree Cover Using Scalable Deep Learning Architectures

    NASA Astrophysics Data System (ADS)

    ganguly, sangram; basu, saikat; nemani, ramakrishna; mukhopadhyay, supratik; michaelis, andrew; votava, petr; saatchi, sassan

    2016-04-01

    Several studies to date have provided an extensive knowledge base for estimating forest aboveground biomass (AGB) and recent advances in space-based modeling of the 3-D canopy structure, combined with canopy reflectance measured by passive optical sensors and radar backscatter, are providing improved satellite-derived AGB density mapping for large scale carbon monitoring applications. A key limitation in forest AGB estimation from remote sensing, however, is the large uncertainty in forest cover estimates from the coarse-to-medium resolution satellite-derived land cover maps (present resolution is limited to 30-m of the USGS NLCD Program). As part of our NASA Carbon Monitoring System Phase II activities, we have demonstrated that uncertainties in forest cover estimates at the Landsat scale result in high uncertainties in AGB estimation, predominantly in heterogeneous forest and urban landscapes. We have successfully tested an approach using scalable deep learning architectures (Feature-enhanced Deep Belief Networks and Semantic Segmentation using Convolutional Neural Networks) and High-Performance Computing with NAIP air-borne imagery data for mapping tree cover at 1-m over California and Maryland. Our first high resolution satellite training label dataset from the NAIP data can be found here at http://csc.lsu.edu/~saikat/deepsat/ . In a comparison with high resolution LiDAR data available over selected regions in the two states, we found our results to be promising both in terms of accuracy as well as our ability to scale nationally. In this project, we propose to estimate very high resolution forest cover for the continental US at spatial resolution of 1-m in support of reducing uncertainties in the AGB estimation. The proposed work will substantially contribute to filling the gaps in ongoing carbon monitoring research and help quantifying the errors and uncertainties in related carbon products.

  8. WESTPA: an interoperable, highly scalable software package for weighted ensemble simulation and analysis.

    PubMed

    Zwier, Matthew C; Adelman, Joshua L; Kaus, Joseph W; Pratt, Adam J; Wong, Kim F; Rego, Nicholas B; Suárez, Ernesto; Lettieri, Steven; Wang, David W; Grabe, Michael; Zuckerman, Daniel M; Chong, Lillian T

    2015-02-10

    The weighted ensemble (WE) path sampling approach orchestrates an ensemble of parallel calculations with intermittent communication to enhance the sampling of rare events, such as molecular associations or conformational changes in proteins or peptides. Trajectories are replicated and pruned in a way that focuses computational effort on underexplored regions of configuration space while maintaining rigorous kinetics. To enable the simulation of rare events at any scale (e.g., atomistic, cellular), we have developed an open-source, interoperable, and highly scalable software package for the execution and analysis of WE simulations: WESTPA (The Weighted Ensemble Simulation Toolkit with Parallelization and Analysis). WESTPA scales to thousands of CPU cores and includes a suite of analysis tools that have been implemented in a massively parallel fashion. The software has been designed to interface conveniently with any dynamics engine and has already been used with a variety of molecular dynamics (e.g., GROMACS, NAMD, OpenMM, AMBER) and cell-modeling packages (e.g., BioNetGen, MCell). WESTPA has been in production use for over a year, and its utility has been demonstrated for a broad set of problems, ranging from atomically detailed host–guest associations to nonspatial chemical kinetics of cellular signaling networks. The following describes the design and features of WESTPA, including the facilities it provides for running WE simulations and storing and analyzing WE simulation data, as well as examples of input and output. PMID:26392815

  9. Highly scalable non-volatile and ultra-low-power phase-change nanowire memory.

    PubMed

    Lee, Se-Ho; Jung, Yeonwoong; Agarwal, Ritesh

    2007-10-01

    The search for a universal memory storage device that combines rapid read and write speeds, high storage density and non-volatility is driving the exploration of new materials in nanostructured form. Phase-change materials, which can be reversibly switched between amorphous and crystalline states, are promising in this respect, but top-down processing of these materials into nanostructures often damages their useful properties. Self-assembled nanowire-based phase-change material memory devices offer an attractive solution owing to their sub-lithographic sizes and unique geometry, coupled with the facile etch-free processes with which they can be fabricated. Here, we explore the effects of nanoscaling on the memory-storage capability of self-assembled Ge2Sb2Te5 nanowires, an important phase-change material. Our measurements of write-current amplitude, switching speed, endurance and data retention time in these devices show that such nanowires are promising building blocks for non-volatile scalable memory and may represent the ultimate size limit in exploring current-induced phase transition in nanoscale systems. PMID:18654387

  10. ScalaTrace: Scalable Compression and Replay of Communication Traces for High Performance Computing

    SciTech Connect

    Noeth, M; Ratn, P; Mueller, F; Schulz, M; de Supinski, B R

    2008-05-16

    Characterizing the communication behavior of large-scale applications is a difficult and costly task due to code/system complexity and long execution times. While many tools to study this behavior have been developed, these approaches either aggregate information in a lossy way through high-level statistics or produce huge trace files that are hard to handle. We contribute an approach that provides orders of magnitude smaller, if not near-constant size, communication traces regardless of the number of nodes while preserving structural information. We introduce intra- and inter-node compression techniques of MPI events that are capable of extracting an application's communication structure. We further present a replay mechanism for the traces generated by our approach and discuss results of our implementation for BlueGene/L. Given this novel capability, we discuss its impact on communication tuning and beyond. To the best of our knowledge, such a concise representation of MPI traces in a scalable manner combined with deterministic MPI call replay are without any precedent.

  11. WESTPA: An interoperable, highly scalable software package for weighted ensemble simulation and analysis

    PubMed Central

    Zwier, Matthew C.; Adelman, Joshua L.; Kaus, Joseph W.; Pratt, Adam J.; Wong, Kim F.; Rego, Nicholas B.; Suárez, Ernesto; Lettieri, Steven; Wang, David W.; Grabe, Michael; Zuckerman, Daniel M.; Chong, Lillian T.

    2015-01-01

    The weighted ensemble (WE) path sampling approach orchestrates an ensemble of parallel calculations with intermittent communication to enhance the sampling of rare events, such as molecular associations or conformational changes in proteins or peptides. Trajectories are replicated and pruned in a way that focuses computational effort on under-explored regions of configuration space while maintaining rigorous kinetics. To enable the simulation of rare events at any scale (e.g. atomistic, cellular), we have developed an open-source, interoperable, and highly scalable software package for the execution and analysis of WE simulations: WESTPA (The Weighted Ensemble Simulation Toolkit with Parallelization and Analysis). WESTPA scales to thousands of CPU cores and includes a suite of analysis tools that have been implemented in a massively parallel fashion. The software has been designed to interface conveniently with any dynamics engine and has already been used with a variety of molecular dynamics (e.g. GROMACS, NAMD, OpenMM, AMBER) and cell-modeling packages (e.g. BioNetGen, MCell). WESTPA has been in production use for over a year, and its utility has been demonstrated for a broad set of problems, ranging from atomically detailed host-guest associations to non-spatial chemical kinetics of cellular signaling networks. The following describes the design and features of WESTPA, including the facilities it provides for running WE simulations, storing and analyzing WE simulation data, as well as examples of input and output. PMID:26392815

  12. Technical Report: Toward a Scalable Algorithm to Compute High-Dimensional Integrals of Arbitrary Functions

    SciTech Connect

    Snyder, Abigail C.; Jiao, Yu

    2010-10-01

    Neutron experiments at the Spallation Neutron Source (SNS) at Oak Ridge National Laboratory (ORNL) frequently generate large amounts of data (on the order of 106-1012 data points). Hence, traditional data analysis tools run on a single CPU take too long to be practical and scientists are unable to efficiently analyze all data generated by experiments. Our goal is to develop a scalable algorithm to efficiently compute high-dimensional integrals of arbitrary functions. This algorithm can then be used to integrate the four-dimensional integrals that arise as part of modeling intensity from the experiments at the SNS. Here, three different one-dimensional numerical integration solvers from the GNU Scientific Library were modified and implemented to solve four-dimensional integrals. The results of these solvers on a final integrand provided by scientists at the SNS can be compared to the results of other methods, such as quasi-Monte Carlo methods, computing the same integral. A parallelized version of the most efficient method can allow scientists the opportunity to more effectively analyze all experimental data.

  13. Scalable Clean Exfoliation of High-Quality Few-Layer Black Phosphorus for a Flexible Lithium Ion Battery.

    PubMed

    Chen, Long; Zhou, Guangmin; Liu, Zhibo; Ma, Xiaomeng; Chen, Jing; Zhang, Zhiyong; Ma, Xiuliang; Li, Feng; Cheng, Hui-Ming; Ren, Wencai

    2016-01-20

    Few-layer black phosphorus (BP) nanosheets that are clean and of high quality, are efficiently produced by exfoliating bulk BP crystals, which are prepared by a scalable gas-phase catalytic transformation method in water. They are stable enough in water for further processing and applications. As an example, these BP nanosheets are combined with graphene to give high-performance flexible lithium-ion batteries. PMID:26584241

  14. High-Performance Optical 3R Regeneration for Scalable Fiber Transmission System Applications

    NASA Astrophysics Data System (ADS)

    Zhu, Zuqing; Funabashi, Masaki; Pan, Zhong; Paraschis, Loukas; Harris, David L.; Ben Yoo, S. J.

    2007-02-01

    This paper proposes and demonstrates optical 3R regeneration techniques for high-performance and scalable 10-Gb/s transmission systems. The 3R structures rely on monolithically integrated all-active semiconductor optical amplifier-based Mach Zehnder interferometers (SOA-MZIs) for signal reshaping and optical narrowband filtering using a Fabry Pérot filter (FPF) for all-optical clock recovery. The experimental results indicate very stable operation and superior cascadability of the proposed optical 3R structure, allowing error-free and low-penalty 10-Gb/s [pseudorandom bit sequence (PRBS) 223 - 1] return-to-zero (RZ) transmission through a record distance of 1 250 000 km using 10 000 optical 3R stages. Clock-enhancement techniques using a SOA-MZI are then proposed to accommodate the clock performance degradations that arise from dispersion uncompensated transmission. Leveraging such clock-enhancement techniques, we experimentally demonstrate error-free 125 000-km RZ dispersion uncompensated transmission at 10 Gb/s (PRBS 223 - 1) using 1000 stages of optical 3R regenerators spaced by 125-km large-effective-area fiber spans. To evaluate the proposed optical 3R structures in a relatively realistic environment and to investigate the tradeoff between the cascadability and the spacing of the optical 3R, a fiber recirculation loop is set up with 264- and 462-km deployed fiber. The field-trial experiment achieves error-free 10-Gb/s RZ transmission using PRBS 223} - 1 through 264 000-km deployed fiber across 1000 stages of optical 3R regenerators spaced by 264-km spans.

  15. Scalable High Performance Image Registration Framework by Unsupervised Deep Feature Representations Learning

    PubMed Central

    Wu, Guorong; Kim, Minjeong; Wang, Qian; Munsell, Brent C.

    2015-01-01

    Feature selection is a critical step in deformable image registration. In particular, selecting the most discriminative features that accurately and concisely describe complex morphological patterns in image patches improves correspondence detection, which in turn improves image registration accuracy. Furthermore, since more and more imaging modalities are being invented to better identify morphological changes in medical imaging data,, the development of deformable image registration method that scales well to new image modalities or new image applications with little to no human intervention would have a significant impact on the medical image analysis community. To address these concerns, a learning-based image registration framework is proposed that uses deep learning to discover compact and highly discriminative features upon observed imaging data. Specifically, the proposed feature selection method uses a convolutional stacked auto-encoder to identify intrinsic deep feature representations in image patches. Since deep learning is an unsupervised learning method, no ground truth label knowledge is required. This makes the proposed feature selection method more flexible to new imaging modalities since feature representations can be directly learned from the observed imaging data in a very short amount of time. Using the LONI and ADNI imaging datasets, image registration performance was compared to two existing state-of-the-art deformable image registration methods that use handcrafted features. To demonstrate the scalability of the proposed image registration framework image registration experiments were conducted on 7.0-tesla brain MR images. In all experiments, the results showed the new image registration framework consistently demonstrated more accurate registration results when compared to state-of-the-art. PMID:26552069

  16. Implementation of scalable video coding deblocking filter from high-level SystemC description

    NASA Astrophysics Data System (ADS)

    Carballo, Pedro P.; Espino, Omar; Neris, Romén.; Hernández-Fernández, Pedro; Szydzik, Tomasz M.; Núñez, Antonio

    2013-05-01

    This paper describes key concepts in the design and implementation of a deblocking filter (DF) for a H.264/SVC video decoder. The DF supports QCIF and CIF video formats with temporal and spatial scalability. The design flow starts from a SystemC functional model and has been refined using high-level synthesis methodology to RTL microarchitecture. The process is guided with performance measurements (latency, cycle time, power, resource utilization) with the objective of assuring the quality of results of the final system. The functional model of the DF is created in an incremental way from the AVC DF model using OpenSVC source code as reference. The design flow continues with the logic synthesis and the implementation on the FPGA using various strategies. The final implementation is chosen among the implementations that meet the timing constraints. The DF is capable to run at 100 MHz, and macroblocks are processed in 6,500 clock cycles for a throughput of 130 fps for QCIF format and 37 fps for CIF format. The proposed architecture for the complete H.264/SVC decoder is composed of an OMAP 3530 SOC (ARM Cortex-A8 GPP + DSP) and the FPGA Virtex-5 acting as a coprocessor for DF implementation. The DF is connected to the OMAP SOC using the GPMC interface. A validation platform has been developed using the embedded PowerPC processor in the FPGA, composing a SoC that integrates the frame generation and visualization in a TFT screen. The FPGA implements both the DF core and a GPMC slave core. Both cores are connected to the PowerPC440 embedded processor using LocalLink interfaces. The FPGA also contains a local memory capable of storing information necessary to filter a complete frame and to store a decoded picture frame. The complete system is implemented in a Virtex5 FX70T device.

  17. Scalable High-Performance Image Registration Framework by Unsupervised Deep Feature Representations Learning.

    PubMed

    Wu, Guorong; Kim, Minjeong; Wang, Qian; Munsell, Brent C; Shen, Dinggang

    2016-07-01

    Feature selection is a critical step in deformable image registration. In particular, selecting the most discriminative features that accurately and concisely describe complex morphological patterns in image patches improves correspondence detection, which in turn improves image registration accuracy. Furthermore, since more and more imaging modalities are being invented to better identify morphological changes in medical imaging data, the development of deformable image registration method that scales well to new image modalities or new image applications with little to no human intervention would have a significant impact on the medical image analysis community. To address these concerns, a learning-based image registration framework is proposed that uses deep learning to discover compact and highly discriminative features upon observed imaging data. Specifically, the proposed feature selection method uses a convolutional stacked autoencoder to identify intrinsic deep feature representations in image patches. Since deep learning is an unsupervised learning method, no ground truth label knowledge is required. This makes the proposed feature selection method more flexible to new imaging modalities since feature representations can be directly learned from the observed imaging data in a very short amount of time. Using the LONI and ADNI imaging datasets, image registration performance was compared to two existing state-of-the-art deformable image registration methods that use handcrafted features. To demonstrate the scalability of the proposed image registration framework, image registration experiments were conducted on 7.0-T brain MR images. In all experiments, the results showed that the new image registration framework consistently demonstrated more accurate registration results when compared to state of the art. PMID:26552069

  18. Personalised Prescription of Scalable High Intensity Interval Training to Inactive Female Adults of Different Ages

    PubMed Central

    Mair, Jacqueline L.

    2016-01-01

    Stepping is a convenient form of scalable high-intensity interval training (HIIT) that may lead to health benefits. However, the accurate personalised prescription of stepping is hampered by a lack of evidence on optimal stepping cadences and step heights for various populations. This study examined the acute physiological responses to stepping exercise at various heights and cadences in young (n = 14) and middle-aged (n = 14) females in order to develop an equation that facilitates prescription of stepping at targeted intensities. Participants completed a step test protocol consisting of randomised three-minute bouts at different step cadences (80, 90, 100, 110 steps·min-1) and step heights (17, 25, 30, 34 cm). Aerobic demand and heart rate values were measured throughout. Resting metabolic rate was measured in order to develop female specific metabolic equivalents (METs) for stepping. Results revealed significant differences between age groups for METs and heart rate reserve, and within-group differences for METs, heart rate, and metabolic cost, at different step heights and cadences. At a given step height and cadence, middle-aged females were required to work at an intensity on average 1.9 ± 0.26 METs greater than the younger females. A prescriptive equation was developed to assess energy cost in METs using multilevel regression analysis with factors of step height, step cadence and age. Considering recent evidence supporting accumulated bouts of HIIT exercise for health benefits, this equation, which allows HIIT to be personally prescribed to inactive and sedentary women, has potential impact as a public health exercise prescription tool. PMID:26848956

  19. Personalised Prescription of Scalable High Intensity Interval Training to Inactive Female Adults of Different Ages.

    PubMed

    Mair, Jacqueline L; Nevill, Alan M; De Vito, Giuseppe; Boreham, Colin A

    2016-01-01

    Stepping is a convenient form of scalable high-intensity interval training (HIIT) that may lead to health benefits. However, the accurate personalised prescription of stepping is hampered by a lack of evidence on optimal stepping cadences and step heights for various populations. This study examined the acute physiological responses to stepping exercise at various heights and cadences in young (n = 14) and middle-aged (n = 14) females in order to develop an equation that facilitates prescription of stepping at targeted intensities. Participants completed a step test protocol consisting of randomised three-minute bouts at different step cadences (80, 90, 100, 110 steps·min-1) and step heights (17, 25, 30, 34 cm). Aerobic demand and heart rate values were measured throughout. Resting metabolic rate was measured in order to develop female specific metabolic equivalents (METs) for stepping. Results revealed significant differences between age groups for METs and heart rate reserve, and within-group differences for METs, heart rate, and metabolic cost, at different step heights and cadences. At a given step height and cadence, middle-aged females were required to work at an intensity on average 1.9 ± 0.26 METs greater than the younger females. A prescriptive equation was developed to assess energy cost in METs using multilevel regression analysis with factors of step height, step cadence and age. Considering recent evidence supporting accumulated bouts of HIIT exercise for health benefits, this equation, which allows HIIT to be personally prescribed to inactive and sedentary women, has potential impact as a public health exercise prescription tool. PMID:26848956

  20. Ultra-High Performance, High-Temperature Superconducting Wires via Cost-effective, Scalable, Co-evaporation Process

    SciTech Connect

    Kim, Dr. Hosup; Oh, Sang-Soo; Ha, HS; Youm, D; Moon, SH; Kim, JH; Heo, YU; Dou, SX; Wee, Sung Hun; Goyal, Amit

    2014-01-01

    Long-length, high-temperature superconducting (HTS) wires capable of carrying high critical current, Ic, are required for a wide range of applications. Here, we report extremely high performance HTS wires based on 5 m thick SmBa2Cu3O7- (SmBCO) single layer films on textured metallic templates. SmBCO layer wires over 20 meters long were deposited by a cost-effective, scalable co-evaporation process using a batch-type drum in a dual chamber. All deposition parameters influencing the composition, phase, and texture of the films were optimized via a unique combinatorial method that is broadly applicable for co-evaporation of other promising complex materials containing several cations. Thick SmBCO layers deposited under optimized conditions exhibit excellent cube-on-cube epitaxy. Such excellent structural epitaxy over the entire thickness results in exceptionally high Ic performance, with average Ic over 1000 A/cm for the entire 22 meter long wire and maximum Ic over 1,500 A/cm for a short 12 cm long tape. The Ic values reported in this work are the highest values ever reported from any lengths of cuprate-based HTS wire or conductor.

  1. Ultra-High Performance, High-Temperature Superconducting Wires via Cost-effective, Scalable, Co-evaporation Process

    PubMed Central

    Kim, Ho-Sup; Oh, Sang-Soo; Ha, Hong-Soo; Youm, Dojun; Moon, Seung-Hyun; Kim, Jung Ho; Dou, Shi Xue; Heo, Yoon-Uk; Wee, Sung-Hun; Goyal, Amit

    2014-01-01

    Long-length, high-temperature superconducting (HTS) wires capable of carrying high critical current, Ic, are required for a wide range of applications. Here, we report extremely high performance HTS wires based on 5 μm thick SmBa2Cu3O7 − δ (SmBCO) single layer films on textured metallic templates. SmBCO layer wires over 20 meters long were deposited by a cost-effective, scalable co-evaporation process using a batch-type drum in a dual chamber. All deposition parameters influencing the composition, phase, and texture of the films were optimized via a unique combinatorial method that is broadly applicable for co-evaporation of other promising complex materials containing several cations. Thick SmBCO layers deposited under optimized conditions exhibit excellent cube-on-cube epitaxy. Such excellent structural epitaxy over the entire thickness results in exceptionally high Ic performance, with average Ic over 1,000 A/cm-width for the entire 22 meter long wire and maximum Ic over 1,500 A/cm-width for a short 12 cm long tape. The Ic values reported in this work are the highest values ever reported from any lengths of cuprate-based HTS wire or conductor. PMID:24752189

  2. The construction of a FBG-based hierarchical AOFSN with high reliability and scalability

    NASA Astrophysics Data System (ADS)

    Peng, Li-mei; Yang, Won-Hyuk; Li, Xin-wan; Kim, Young-Chon

    2008-11-01

    To improve the reliability and scalability that are very important for large-scale all optical fiber sensor networks (AOFSN), three-level hierarchical sensor network architectures are proposed. The first two levels consist of active interrogation and RNs, respectively. The third level called sensor subnet (SSN) consists of passive FBGs and a few switches. As AOFSN is mainly multiplexed by wired and passive FBGs, the routing algorithm for scanning sensors is determined by the virtual topology of SSN due to the passivity. Therefore, the research concentrates on the construction of SSN and aims at proposing regular and unicursal virtual topology to realize reliable and scalable routing schemes. Two regular types of SSNs are proposed. Each type consists of several sensor cells (SC), square-based SC (SSC) or pentagon-based SC (PSC) and is scaled several times from the SCs. The virtual topologies maintain the self-similar square- or pentagon-like architecture so as to gain simple routing. Finally, the switch architecture of RN is proposed for the reliability of the first two levels; and then, the reliability and scalability of SSN are discussed in view of how much link failures can be tolerant, and how each SC is scaled to maintain the self-similarity, respectively.

  3. Thickness scalability of large volume cadmium zinc telluride high resolution radiation detectors

    NASA Astrophysics Data System (ADS)

    Awadalla, S. A.; Chen, H.; Mackenzie, J.; Lu, P.; Iniewski, K.; Marthandam, P.; Redden, R.; Bindley, G.; He, Z.; Zhang, F.

    2009-06-01

    This work focuses on the thickness scalability of traveling heater method (THM) grown CdZnTe crystals to produce large volume detectors with optimized spectroscopic performance. To meet this challenge, we have tuned both our THM growth process, to grow 75 mm diameter ingots, and our postgrowth annealing process. We have increased the thickness of our sliced wafers from 6 to 12 and 18 mm allowing the production of 10 and 15 mm thick detectors. As the detectors' thickness is scaled up, the energy resolution of both types, as pseudo-Frisch grid and pixelated monolithic detectors showed no degradation indicating improved materials uniformity and transport properties.

  4. A scalable strategy for high-throughput GFP tagging of endogenous human proteins.

    PubMed

    Leonetti, Manuel D; Sekine, Sayaka; Kamiyama, Daichi; Weissman, Jonathan S; Huang, Bo

    2016-06-21

    A central challenge of the postgenomic era is to comprehensively characterize the cellular role of the ∼20,000 proteins encoded in the human genome. To systematically study protein function in a native cellular background, libraries of human cell lines expressing proteins tagged with a functional sequence at their endogenous loci would be very valuable. Here, using electroporation of Cas9 nuclease/single-guide RNA ribonucleoproteins and taking advantage of a split-GFP system, we describe a scalable method for the robust, scarless, and specific tagging of endogenous human genes with GFP. Our approach requires no molecular cloning and allows a large number of cell lines to be processed in parallel. We demonstrate the scalability of our method by targeting 48 human genes and show that the resulting GFP fluorescence correlates with protein expression levels. We next present how our protocols can be easily adapted for the tagging of a given target with GFP repeats, critically enabling the study of low-abundance proteins. Finally, we show that our GFP tagging approach allows the biochemical isolation of native protein complexes for proteomic studies. Taken together, our results pave the way for the large-scale generation of endogenously tagged human cell lines for the proteome-wide analysis of protein localization and interaction networks in a native cellular context. PMID:27274053

  5. Three-dimensional Finite Element Formulation and Scalable Domain Decomposition for High Fidelity Rotor Dynamic Analysis

    NASA Technical Reports Server (NTRS)

    Datta, Anubhav; Johnson, Wayne R.

    2009-01-01

    This paper has two objectives. The first objective is to formulate a 3-dimensional Finite Element Model for the dynamic analysis of helicopter rotor blades. The second objective is to implement and analyze a dual-primal iterative substructuring based Krylov solver, that is parallel and scalable, for the solution of the 3-D FEM analysis. The numerical and parallel scalability of the solver is studied using two prototype problems - one for ideal hover (symmetric) and one for a transient forward flight (non-symmetric) - both carried out on up to 48 processors. In both hover and forward flight conditions, a perfect linear speed-up is observed, for a given problem size, up to the point of substructure optimality. Substructure optimality and the linear parallel speed-up range are both shown to depend on the problem size as well as on the selection of the coarse problem. With a larger problem size, linear speed-up is restored up to the new substructure optimality. The solver also scales with problem size - even though this conclusion is premature given the small prototype grids considered in this study.

  6. Enabling a Highly-Scalable Global Address Space Model for Petascale Computing

    SciTech Connect

    Apra, Edoardo; Vetter, Jeffrey S; Yu, Weikuan

    2010-01-01

    Over the past decade, the trajectory to the petascale has been built on increased complexity and scale of the underlying parallel architectures. Meanwhile, software de- velopers have struggled to provide tools that maintain the productivity of computational science teams using these new systems. In this regard, Global Address Space (GAS) programming models provide a straightforward and easy to use addressing model, which can lead to improved produc- tivity. However, the scalability of GAS depends directly on the design and implementation of the runtime system on the target petascale distributed-memory architecture. In this paper, we describe the design, implementation, and optimization of the Aggregate Remote Memory Copy Interface (ARMCI) runtime library on the Cray XT5 2.3 PetaFLOPs computer at Oak Ridge National Laboratory. We optimized our implementation with the flow intimation technique that we have introduced in this paper. Our optimized ARMCI implementation improves scalability of both the Global Arrays (GA) programming model and a real-world chemistry application NWChem from small jobs up through 180,000 cores.

  7. Scalable coherent interface

    SciTech Connect

    Alnaes, K.; Kristiansen, E.H. ); Gustavson, D.B. ); James, D.V. )

    1990-01-01

    The Scalable Coherent Interface (IEEE P1596) is establishing an interface standard for very high performance multiprocessors, supporting a cache-coherent-memory model scalable to systems with up to 64K nodes. This Scalable Coherent Interface (SCI) will supply a peak bandwidth per node of 1 GigaByte/second. The SCI standard should facilitate assembly of processor, memory, I/O and bus bridge cards from multiple vendors into massively parallel systems with throughput far above what is possible today. The SCI standard encompasses two levels of interface, a physical level and a logical level. The physical level specifies electrical, mechanical and thermal characteristics of connectors and cards that meet the standard. The logical level describes the address space, data transfer protocols, cache coherence mechanisms, synchronization primitives and error recovery. In this paper we address logical level issues such as packet formats, packet transmission, transaction handshake, flow control, and cache coherence. 11 refs., 10 figs.

  8. Predicting High-Throughput Screening Results With Scalable Literature-Based Discovery Methods

    PubMed Central

    Cohen, T; Widdows, D; Stephan, C; Zinner, R; Kim, J; Rindflesch, T; Davies, P

    2014-01-01

    The identification of new therapeutic uses for existing agents has been proposed as a means to mitigate the escalating cost of drug development. A common approach to such repurposing involves screening libraries of agents for activities against cell lines. In silico methods using knowledge from the biomedical literature have been proposed to constrain the costs of screening by identifying agents that are likely to be effective a priori. However, results obtained with these methods are seldom evaluated empirically. Conversely, screening experiments have been criticized for their inability to reveal the biological basis of their results. In this paper, we evaluate the ability of a scalable literature-based approach, discovery-by-analogy, to identify a small number of active agents within a large library screened for activity against prostate cancer cells. The methods used permit retrieval of the knowledge used to infer their predictions, providing a plausible biological basis for predicted activity. PMID:25295575

  9. Fast generation of a high-quality computer-generated hologram using a scalable and flexible PC cluster.

    PubMed

    Song, Joongseok; Kim, Changseob; Park, Hanhoon; Park, Jong-Il

    2016-05-01

    In order to efficiently generate a high-quality computer-generated hologram (HQ-CGH), which requires that both a three-dimensional object image and its computer-generated hologram (CGH) are in high-definition resolution, we implement a fast CGH generation system using a scalable and flexible personal computer (PC) cluster. From experimental results obtained in generating a HQ-CGH with a CGH resolution of 1536×1536 and 2,155,898 light sources using a PC cluster comprising a server PC and nine client PCs, it is verified that the proposed system is approximately 4.7 times faster than a single PC with two high-performance GPUs. PMID:27140388

  10. Real-time high-resolution downsampling algorithm on many-core processor for spatially scalable video coding

    NASA Astrophysics Data System (ADS)

    Buhari, Adamu Muhammad; Ling, Huo-Chong; Baskaran, Vishnu Monn; Wong, KokSheik

    2015-01-01

    The progression toward spatially scalable video coding (SVC) solutions for ubiquitous endpoint systems introduces challenges to sustain real-time frame rates in downsampling high-resolution videos into multiple layers. In addressing these challenges, we put forward a hardware accelerated downsampling algorithm on a parallel computing platform. First, we investigate the principal architecture of a serial downsampling algorithm in the Joint-Scalable-Video-Model reference software to identify the performance limitations for spatially SVC. Then, a parallel multicore-based downsampling algorithm is studied as a benchmark. Experimental results for this algorithm using an 8-core processor exhibit performance speedup of 5.25× against the serial algorithm in downsampling a quantum extended graphics array at 1536p video resolution into three lower resolution layers (i.e., Full-HD at 1080p, HD at 720p, and Quarter-HD at 540p). However, the achieved speedup here does not translate into the minimum required frame rate of 15 frames per second (fps) for real-time video processing. To improve the speedup, a many-core based downsampling algorithm using the compute unified device architecture parallel computing platform is proposed. The proposed algorithm increases the performance speedup to 26.14× against the serial algorithm. Crucially, the proposed algorithm exceeds the target frame rate of 15 fps, which in turn is advantageous to the overall performance of the video encoding process.

  11. High power impulse magnetron sputtering and related discharges: scalable plasma sources for plasma-based ion implantation and deposition

    SciTech Connect

    Anders, Andre

    2009-09-01

    High power impulse magnetron sputtering (HIPIMS) and related self-sputtering techniques are reviewed from a viewpoint of plasma-based ion implantation and deposition (PBII&D). HIPIMS combines the classical, scalable sputtering technology with pulsed power, which is an elegant way of ionizing the sputtered atoms. Related approaches, such as sustained self-sputtering, are also considered. The resulting intense flux of ions to the substrate consists of a mixture of metal and gas ions when using a process gas, or of metal ions only when using `gasless? or pure self-sputtering. In many respects, processing with HIPIMS plasmas is similar to processing with filtered cathodic arc plasmas, though the former is easier to scale to large areas. Both ion implantation and etching (high bias voltage, without deposition) and thin film deposition (low bias, or bias of low duty cycle) have been demonstrated.

  12. Integrated Scalable Parallel Firewall and Intrusion Detection System for High-Speed Networks

    SciTech Connect

    Fulp, Errin W; Anderson, Robert E; Ahn, David K

    2009-08-31

    This project developed a new scalable network firewall and Intrusion Protection System (IPS) that can manage increasing traffic loads, higher network speeds, and strict Quality of Service (QoS) requirements. This new approach provides a strong foundation for next-generation network security technologies and products that address growing and unmet needs in the government and corporate sectors by delivering Optimal Network Security. Controlling access is an essential task for securing networks that are vital to private industry, government agencies, and the military. This access can be granted or denied based on the packet header or payload contents. For example, a simple network firewall enforces a security policy by inspecting and filtering the packet headers. As a complement to the firewall, an Intrusion Detection System (IDS) inspects the packet payload for known threat signatures; for example, virus or worm. Similar to a firewall policy, IDS policies consist of multiple rules that specify an action for matching packets. Each rule can specify different items, such as the signature contents and the signature location within the payload. When the firewall and IDS are merged into one device, the resulting system is referred to as an Intrusion Protection System (IPS), which provides both packet header and payload inspections. Having both types of inspections is very desirable and more manageable in a single device.

  13. A highly scalable parallel computation strategy and optimized implementation for Fresnel Seismic Tomography

    NASA Astrophysics Data System (ADS)

    Gao, Yongan; Zhao, Changhai; Li, Chuang; Yan, Haihua; Zhao, Liang

    2013-03-01

    Fresnel Seismic Tomography which uses a huge amount of seismic data is an efficient methodology of researching three-dimensional structure of earth. However, in practical application, it confronts with two key challenges of enormous data volume and huge computation. It is difficult to accomplish computation tasks under normal operating environment and computation strategies. In this paper, a Job-By-Application parallel computation strategy, which uses MPI (Message Passing Interface) and Pthread hybrid programming models based on the cluster, is designed to implement Fresnel seismic tomography, this method can solve the problem of allocating tasks dynamically, improve the load balancing and scalability of the system effectively; and we adopted the cached I/O strategy to accommodate the limited memory resources. Experimental results demonstrated that the program implemented on these strategies could completed the actual job within the idea time, the running of the program was stable, achieved load balancing, showed a good speedup and could adapt to the hardware environment of insufficient memory.

  14. Churchill: an ultra-fast, deterministic, highly scalable and balanced parallelization strategy for the discovery of human genetic variation in clinical and population-scale genomics.

    PubMed

    Kelly, Benjamin J; Fitch, James R; Hu, Yangqiu; Corsmeier, Donald J; Zhong, Huachun; Wetzel, Amy N; Nordquist, Russell D; Newsom, David L; White, Peter

    2015-01-01

    While advances in genome sequencing technology make population-scale genomics a possibility, current approaches for analysis of these data rely upon parallelization strategies that have limited scalability, complex implementation and lack reproducibility. Churchill, a balanced regional parallelization strategy, overcomes these challenges, fully automating the multiple steps required to go from raw sequencing reads to variant discovery. Through implementation of novel deterministic parallelization techniques, Churchill allows computationally efficient analysis of a high-depth whole genome sample in less than two hours. The method is highly scalable, enabling full analysis of the 1000 Genomes raw sequence dataset in a week using cloud resources. http://churchill.nchri.org/. PMID:25600152

  15. The microwave-to-flow paradigm: translating high-temperature batch microwave chemistry to scalable continuous-flow processes.

    PubMed

    Glasnov, Toma N; Kappe, C Oliver

    2011-10-17

    The popularity of dedicated microwave reactors in many academic and industrial laboratories has produced a plethora of synthetic protocols that are based on this enabling technology. In the majority of examples, transformations that require several hours when performed using conventional heating under reflux conditions reach completion in a few minutes or even seconds in sealed-vessel, autoclave-type, microwave reactors. However, one severe drawback of microwave chemistry is the difficulty in scaling this technology to a production-scale level. This Concept article demonstrates that this limitation can be overcome by translating batch microwave chemistry to scalable continuous-flow processes. For this purpose, conventionally heated micro- or mesofluidic flow devices fitted with a back-pressure regulator are employed, in which the high temperatures and pressures attainable in a sealed-vessel microwave chemistry batch experiment can be mimicked. PMID:21932289

  16. SciSpark: Highly Interactive and Scalable Model Evaluation and Climate Metrics

    NASA Astrophysics Data System (ADS)

    Wilson, B. D.; Mattmann, C. A.; Waliser, D. E.; Kim, J.; Loikith, P.; Lee, H.; McGibbney, L. J.; Whitehall, K. D.

    2014-12-01

    Remote sensing data and climate model output are multi-dimensional arrays of massive sizes locked away in heterogeneous file formats (HDF5/4, NetCDF 3/4) and metadata models (HDF-EOS, CF) making it difficult to perform multi-stage, iterative science processing since each stage requires writing and reading data to and from disk. We are developing a lightning fast Big Data technology called SciSpark based on ApacheTM Spark. Spark implements the map-reduce paradigm for parallel computing on a cluster, but emphasizes in-memory computation, "spilling" to disk only as needed, and so outperforms the disk-based ApacheTM Hadoop by 100x in memory and by 10x on disk, and makes iterative algorithms feasible. SciSpark will enable scalable model evaluation by executing large-scale comparisons of A-Train satellite observations to model grids on a cluster of 100 to 1000 compute nodes. This 2nd generation capability for NASA's Regional Climate Model Evaluation System (RCMES) will compute simple climate metrics at interactive speeds, and extend to quite sophisticated iterative algorithms such as machine-learning (ML) based clustering of temperature PDFs, and even graph-based algorithms for searching for Mesocale Convective Complexes. The goals of SciSpark are to: (1) Decrease the time to compute comparison statistics and plots from minutes to seconds; (2) Allow for interactive exploration of time-series properties over seasons and years; (3) Decrease the time for satellite data ingestion into RCMES to hours; (4) Allow for Level-2 comparisons with higher-order statistics or PDF's in minutes to hours; and (5) Move RCMES into a near real time decision-making platform. We will report on: the architecture and design of SciSpark, our efforts to integrate climate science algorithms in Python and Scala, parallel ingest and partitioning (sharding) of A-Train satellite observations from HDF files and model grids from netCDF files, first parallel runs to compute comparison statistics and PDF

  17. SciSpark: Highly Interactive and Scalable Model Evaluation and Climate Metrics

    NASA Astrophysics Data System (ADS)

    Wilson, B. D.; Palamuttam, R. S.; Mogrovejo, R. M.; Whitehall, K. D.; Mattmann, C. A.; Verma, R.; Waliser, D. E.; Lee, H.

    2015-12-01

    Remote sensing data and climate model output are multi-dimensional arrays of massive sizes locked away in heterogeneous file formats (HDF5/4, NetCDF 3/4) and metadata models (HDF-EOS, CF) making it difficult to perform multi-stage, iterative science processing since each stage requires writing and reading data to and from disk. We are developing a lightning fast Big Data technology called SciSpark based on ApacheTM Spark under a NASA AIST grant (PI Mattmann). Spark implements the map-reduce paradigm for parallel computing on a cluster, but emphasizes in-memory computation, "spilling" to disk only as needed, and so outperforms the disk-based ApacheTM Hadoop by 100x in memory and by 10x on disk. SciSpark will enable scalable model evaluation by executing large-scale comparisons of A-Train satellite observations to model grids on a cluster of 10 to 1000 compute nodes. This 2nd generation capability for NASA's Regional Climate Model Evaluation System (RCMES) will compute simple climate metrics at interactive speeds, and extend to quite sophisticated iterative algorithms such as machine-learning based clustering of temperature PDFs, and even graph-based algorithms for searching for Mesocale Convective Complexes. We have implemented a parallel data ingest capability in which the user specifies desired variables (arrays) as several time-sorted lists of URL's (i.e. using OPeNDAP model.nc?varname, or local files). The specified variables are partitioned by time/space and then each Spark node pulls its bundle of arrays into memory to begin a computation pipeline. We also investigated the performance of several N-dim. array libraries (scala breeze, java jblas & netlib-java, and ND4J). We are currently developing science codes using ND4J and studying memory behavior on the JVM. On the pyspark side, many of our science codes already use the numpy and SciPy ecosystems. The talk will cover: the architecture of SciSpark, the design of the scientific RDD (sRDD) data structure, our

  18. Scalability of a Low-Cost Multi-Teraflop Linux Cluster for High-End Classical Atomistic and Quantum Mechanical Simulations

    NASA Technical Reports Server (NTRS)

    Kikuchi, Hideaki; Kalia, Rajiv K.; Nakano, Aiichiro; Vashishta, Priya; Shimojo, Fuyuki; Saini, Subhash

    2003-01-01

    Scalability of a low-cost, Intel Xeon-based, multi-Teraflop Linux cluster is tested for two high-end scientific applications: Classical atomistic simulation based on the molecular dynamics method and quantum mechanical calculation based on the density functional theory. These scalable parallel applications use space-time multiresolution algorithms and feature computational-space decomposition, wavelet-based adaptive load balancing, and spacefilling-curve-based data compression for scalable I/O. Comparative performance tests are performed on a 1,024-processor Linux cluster and a conventional higher-end parallel supercomputer, 1,184-processor IBM SP4. The results show that the performance of the Linux cluster is comparable to that of the SP4. We also study various effects, such as the sharing of memory and L2 cache among processors, on the performance.

  19. High-flux ionic diodes, ionic transistors and ionic amplifiers based on external ion concentration polarization by an ion exchange membrane: a new scalable ionic circuit platform.

    PubMed

    Sun, Gongchen; Senapati, Satyajyoti; Chang, Hsueh-Chia

    2016-04-01

    A microfluidic ion exchange membrane hybrid chip is fabricated using polymer-based, lithography-free methods to achieve ionic diode, transistor and amplifier functionalities with the same four-terminal design. The high ionic flux (>100 μA) feature of the chip can enable a scalable integrated ionic circuit platform for micro-total-analytical systems. PMID:26960551

  20. Scalable Work Stealing

    SciTech Connect

    Dinan, James S.; Larkins, D. B.; Sadayappan, Ponnuswamy; Krishnamoorthy, Sriram; Nieplocha, Jaroslaw

    2009-11-14

    Irregular and dynamic parallel applications pose significant challenges to achieving scalable performance on large-scale multicore clusters. These applications often require ongoing, dynamic load balancing in order to maintain efficiency. While effective at small scale, centralized load balancing schemes quickly become a bottleneck on large-scale clusters. Work stealing is a popular approach to distributed dynamic load balancing; however its performance on large-scale clusters is not well understood. Prior work on work stealing has largely focused on shared memory machines. In this work we investigate the design and scalability of work stealing on modern distributed memory systems. We demonstrate high efficiency and low overhead when scaling to 8,192 processors for three benchmark codes: a producer-consumer benchmark, the unbalanced tree search benchmark, and a multiresolution analysis kernel.

  1. Cactus and Visapult: A case study of ultra-high performance distributed visualization using connectionless protocols

    SciTech Connect

    Shalf, John; Bethel, E. Wes

    2002-05-07

    This past decade has seen rapid growth in the size, resolution, and complexity of Grand Challenge simulation codes. Many such problems still require interactive visualization tools to make sense of multi-terabyte data stores. Visapult is a parallel volume rendering tool that employs distributed components, latency tolerant algorithms, and high performance network I/O for effective remote visualization of massive datasets. In this paper we discuss using connectionless protocols to accelerate Visapult network I/O and interfacing Visapult to the Cactus General Relativity code to enable scalable remote monitoring and steering capabilities. With these modifications, network utilization has moved from 25 percent of line-rate using tuned multi-streamed TCP to sustaining 88 percent of line rate using the new UDP-based transport protocol.

  2. A peripheral component interconnect express-based scalable and highly integrated pulsed spectrometer for solution state dynamic nuclear polarization

    NASA Astrophysics Data System (ADS)

    He, Yugui; Feng, Jiwen; Zhang, Zhi; Wang, Chao; Wang, Dong; Chen, Fang; Liu, Maili; Liu, Chaoyang

    2015-08-01

    High sensitivity, high data rates, fast pulses, and accurate synchronization all represent challenges for modern nuclear magnetic resonance spectrometers, which make any expansion or adaptation of these devices to new techniques and experiments difficult. Here, we present a Peripheral Component Interconnect Express (PCIe)-based highly integrated distributed digital architecture pulsed spectrometer that is implemented with electron and nucleus double resonances and is scalable specifically for broad dynamic nuclear polarization (DNP) enhancement applications, including DNP-magnetic resonance spectroscopy/imaging (DNP-MRS/MRI). The distributed modularized architecture can implement more transceiver channels flexibly to meet a variety of MRS/MRI instrumentation needs. The proposed PCIe bus with high data rates can significantly improve data transmission efficiency and communication reliability and allow precise control of pulse sequences. An external high speed double data rate memory chip is used to store acquired data and pulse sequence elements, which greatly accelerates the execution of the pulse sequence, reduces the TR (time of repetition) interval, and improves the accuracy of TR in imaging sequences. Using clock phase-shift technology, we can produce digital pulses accurately with high timing resolution of 1 ns and narrow widths of 4 ns to control the microwave pulses required by pulsed DNP and ensure overall system synchronization. The proposed spectrometer is proved to be both feasible and reliable by observation of a maximum signal enhancement factor of approximately -170 for 1H, and a high quality water image was successfully obtained by DNP-enhanced spin-echo 1H MRI at 0.35 T.

  3. A peripheral component interconnect express-based scalable and highly integrated pulsed spectrometer for solution state dynamic nuclear polarization

    SciTech Connect

    He, Yugui; Liu, Chaoyang; Feng, Jiwen; Wang, Dong; Chen, Fang; Liu, Maili; Zhang, Zhi; Wang, Chao

    2015-08-15

    High sensitivity, high data rates, fast pulses, and accurate synchronization all represent challenges for modern nuclear magnetic resonance spectrometers, which make any expansion or adaptation of these devices to new techniques and experiments difficult. Here, we present a Peripheral Component Interconnect Express (PCIe)-based highly integrated distributed digital architecture pulsed spectrometer that is implemented with electron and nucleus double resonances and is scalable specifically for broad dynamic nuclear polarization (DNP) enhancement applications, including DNP-magnetic resonance spectroscopy/imaging (DNP-MRS/MRI). The distributed modularized architecture can implement more transceiver channels flexibly to meet a variety of MRS/MRI instrumentation needs. The proposed PCIe bus with high data rates can significantly improve data transmission efficiency and communication reliability and allow precise control of pulse sequences. An external high speed double data rate memory chip is used to store acquired data and pulse sequence elements, which greatly accelerates the execution of the pulse sequence, reduces the TR (time of repetition) interval, and improves the accuracy of TR in imaging sequences. Using clock phase-shift technology, we can produce digital pulses accurately with high timing resolution of 1 ns and narrow widths of 4 ns to control the microwave pulses required by pulsed DNP and ensure overall system synchronization. The proposed spectrometer is proved to be both feasible and reliable by observation of a maximum signal enhancement factor of approximately −170 for {sup 1}H, and a high quality water image was successfully obtained by DNP-enhanced spin-echo {sup 1}H MRI at 0.35 T.

  4. Highly Disordered Array of Silicon Nanowires: an Effective and Scalable Approach for Performing and Flexible Electrochemical Biosensors.

    PubMed

    Maiolo, Luca; Polese, Davide; Pecora, Alessandro; Fortunato, Guglielmo; Shacham-Diamand, Yosi; Convertino, Annalisa

    2016-03-01

    The direct integration of disordered arranged and randomly oriented silicon nanowires (SiNWs) into ultraflexible and transferable electronic circuits for electrochemical biosensing applications is proposed. The working electrode (WE) of a three-electrode impedance device, fabricated on a polyimide (PI) film, is modified with SiNWs covered by a thin Au layer and functionalized to bind the sensing element. The biosensing behavior is investigated through the ligand-receptor binding of biotin-avidin system. Impedance measurements show a very efficient detection of the avidin over a broad range of concentrations from hundreds of micromolar down to the picomolar values. The impedance response is modeled through a simple equivalent circuit, which takes into account the unique WE morphology and its modification with successive layers of biomolecules. This approach of exploiting highly disordered SiNW ensemble in biosensing proves to be very promising for the following three main reasons: first, the system morphology allows high sensing performance; second, these nanostructures can be built via scalable and transferable fabrication methodology allowing an easy integration on non-conventional substrates; third, reliable modeling of the sensing response can be developed by considering the morphological and surface characteristics over an ensemble of disordered NWs rather than over individual NWs. PMID:26717420

  5. High Yield and Scalable Fabrication of Nano/Bio Hybrid Graphene Field Effect Transistors for Cancer Biomarker Detection

    NASA Astrophysics Data System (ADS)

    Ducos, Pedro; Diaz, Madeline; Robinson, Matthew; Johnson, A. T. Charlie

    2015-03-01

    Graphene field effect transistors (GFETs) hold tremendous promise for use as biosensor transduction elements due to graphene's high mobility, low noise and all-surface structure with every atom exposed to the environment. We developed a GFET array fabrication based on two approaches, pre-patterned transfer and post-transfer photolithography. Both approaches are scalable, high yield, and electrically stable. Functional groups for protein immobilization were added to the GFET using various bi-functional pyrene-based linkers. One approach immobilized an azide engineered protein through a ``Staudinger Reaction'' chemistry with NHS-phosphine reacting with a 1-aminopyrene linker. Another approach bound an engineered antibody via 1-pyrene butanoic acid succinimidyl ester, where an amine group of the antibody reacts to the succinimide of the linker. GFETs were studied by Raman spectroscopy, AFM and current-gate voltage (I-Vg) characterization at several steps of the fabrication process. A sensing response was obtained for a breast cancer biomarker (HER2) as a function of target concentration. We have started to design multiplexed sensor arrays by adding several functional groups to GFETs on a single chip. Simultaneous detection with these devices will be discussed.

  6. Scalable shear-exfoliation of high-quality phosphorene nanoflakes with reliable electrochemical cycleability in nano batteries

    NASA Astrophysics Data System (ADS)

    Xu, Feng; Ge, Binghui; Chen, Jing; Nathan, Arokia; Xin, Linhuo L.; Ma, Hongyu; Min, Huihua; Zhu, Chongyang; Xia, Weiwei; Li, Zhengrui; Li, Shengli; Yu, Kaihao; Wu, Lijun; Cui, Yiping; Sun, Litao; Zhu, Yimei

    2016-06-01

    Atomically thin black phosphorus (called phosphorene) holds great promise as an alternative to graphene and other two-dimensional transition-metal dichalcogenides as an anode material for lithium-ion batteries (LIBs). However, bulk black phosphorus (BP) suffers from rapid capacity fading and poor rechargeable performance. This work reports for the first time the use of in situ transmission electron microscopy (TEM) to construct nanoscale phosphorene LIBs. This enables direct visualization of the mechanisms underlying capacity fading in thick multilayer phosphorene through real-time capture of delithiation-induced structural decomposition, which serves to reduce electrical conductivity thus causing irreversibility of the lithiated phases. We further demonstrate that few-layer-thick phosphorene successfully circumvents the structural decomposition and holds superior structural restorability, even when subject to multi-cycle lithiation/delithiation processes and concomitant huge volume expansion. This finding provides breakthrough insights into thickness-dependent lithium diffusion kinetics in phosphorene. More importantly, a scalable liquid-phase shear exfoliation route has been developed to produce high-quality ultrathin phosphorene using simple means such as a high-speed shear mixer or even a household kitchen blender with the shear rate threshold of ˜1.25 × 104 s-1. The results reported here will pave the way for industrial-scale applications of rechargeable phosphorene LIBs.

  7. Scalable High Throughput Selection From Phage-displayed Synthetic Antibody Libraries

    PubMed Central

    Miersch, Shane; Li, Zhijian; Hanna, Rachel; McLaughlin, Megan E.; Hornsby, Michael; Matsuguchi, Tet; Paduch, Marcin; Sääf, Annika; Wells, Jim; Koide, Shohei; Kossiakoff, Anthony; Sidhu, Sachdev S.

    2015-01-01

    The demand for antibodies that fulfill the needs of both basic and clinical research applications is high and will dramatically increase in the future. However, it is apparent that traditional monoclonal technologies are not alone up to this task. This has led to the development of alternate methods to satisfy the demand for high quality and renewable affinity reagents to all accessible elements of the proteome. Toward this end, high throughput methods for conducting selections from phage-displayed synthetic antibody libraries have been devised for applications involving diverse antigens and optimized for rapid throughput and success. Herein, a protocol is described in detail that illustrates with video demonstration the parallel selection of Fab-phage clones from high diversity libraries against hundreds of targets using either a manual 96 channel liquid handler or automated robotics system. Using this protocol, a single user can generate hundreds of antigens, select antibodies to them in parallel and validate antibody binding within 6-8 weeks. Highlighted are: i) a viable antigen format, ii) pre-selection antigen characterization, iii) critical steps that influence the selection of specific and high affinity clones, and iv) ways of monitoring selection effectiveness and early stage antibody clone characterization. With this approach, we have obtained synthetic antibody fragments (Fabs) to many target classes including single-pass membrane receptors, secreted protein hormones, and multi-domain intracellular proteins. These fragments are readily converted to full-length antibodies and have been validated to exhibit high affinity and specificity. Further, they have been demonstrated to be functional in a variety of standard immunoassays including Western blotting, ELISA, cellular immunofluorescence, immunoprecipitation and related assays. This methodology will accelerate antibody discovery and ultimately bring us closer to realizing the goal of generating renewable

  8. Multicatalytic colloids with highly scalable, adjustable, and stable functionalities in organic and aqueous media

    NASA Astrophysics Data System (ADS)

    Kim, Donghee; Cheong, Sanghyuk; Ahn, Yun Gyong; Ryu, Sook Won; Kim, Jai-Kyeong; Cho, Jinhan

    2016-03-01

    Despite a large number of developments of noble metal (or metal oxide) NP-based catalysts, it has been a great challenge to prepare high-performance recyclable catalysts with integrated functionalities that can be used in various solvent media. Here, we report on layer-by-layer (LbL) assembled multicatalysts with high catalytic performance, showing high dispersion and recycling stability in organic and aqueous media. The remarkable advantages of our approach are as follows. (i) Various metal or metal oxide NPs with desired catalytic performance can be easily incorporated into multilayered shells, forming densely packed arrays that allow one colloid to be used as a multicatalyst with highly integrated and controllable catalytic properties. (ii) Additionally, the dispersion stability of catalytic colloids in a desired solvent can be determined by the type of ultrathin outermost layer coating each colloid. (iii) Lastly, the covalent bonding between inorganic NPs and dendrimers within multilayer shells enhances the recycling stability of multicatalytic colloids. The resulting core-shell colloids including OA-Fe3O4 NPs, TOABr-Pd NPs, and OA-TiO2 NPs exhibited excellent performance in the oxidation of 3,3',5,5'-tetramethylbenzidine (TMB) and photocatalysis in aqueous media and in the Sonogashira coupling reaction (99% yield) in organic media. Given that the catalytic properties of recyclable colloids reported to date have entirely depended on the functionality of a single catalytic NP layer deposited onto colloids in selective solvent media, our approach provides a basis for the design and exploitation of high-performance recyclable colloids with integrated multicatalytic properties and high dispersion stability in a variety of solvents.Despite a large number of developments of noble metal (or metal oxide) NP-based catalysts, it has been a great challenge to prepare high-performance recyclable catalysts with integrated functionalities that can be used in various solvent

  9. Multicatalytic colloids with highly scalable, adjustable, and stable functionalities in organic and aqueous media.

    PubMed

    Kim, Donghee; Cheong, Sanghyuk; Ahn, Yun Gyong; Ryu, Sook Won; Kim, Jai-Kyeong; Cho, Jinhan

    2016-04-01

    Despite a large number of developments of noble metal (or metal oxide) NP-based catalysts, it has been a great challenge to prepare high-performance recyclable catalysts with integrated functionalities that can be used in various solvent media. Here, we report on layer-by-layer (LbL) assembled multicatalysts with high catalytic performance, showing high dispersion and recycling stability in organic and aqueous media. The remarkable advantages of our approach are as follows. (i) Various metal or metal oxide NPs with desired catalytic performance can be easily incorporated into multilayered shells, forming densely packed arrays that allow one colloid to be used as a multicatalyst with highly integrated and controllable catalytic properties. (ii) Additionally, the dispersion stability of catalytic colloids in a desired solvent can be determined by the type of ultrathin outermost layer coating each colloid. (iii) Lastly, the covalent bonding between inorganic NPs and dendrimers within multilayer shells enhances the recycling stability of multicatalytic colloids. The resulting core-shell colloids including OA-Fe3O4 NPs, TOABr-Pd NPs, and OA-TiO2 NPs exhibited excellent performance in the oxidation of 3,3',5,5'-tetramethylbenzidine (TMB) and photocatalysis in aqueous media and in the Sonogashira coupling reaction (99% yield) in organic media. Given that the catalytic properties of recyclable colloids reported to date have entirely depended on the functionality of a single catalytic NP layer deposited onto colloids in selective solvent media, our approach provides a basis for the design and exploitation of high-performance recyclable colloids with integrated multicatalytic properties and high dispersion stability in a variety of solvents. PMID:26524289

  10. Scalable preparation of porous micron-SnO2/C composites as high performance anode material for lithium ion battery

    NASA Astrophysics Data System (ADS)

    Wang, Ming-Shan; Lei, Ming; Wang, Zhi-Qiang; Zhao, Xing; Xu, Jun; Yang, Wei; Huang, Yun; Li, Xing

    2016-03-01

    Nano tin dioxide-carbon (SnO2/C) composites prepared by various carbon materials, such as carbon nanotubes, porous carbon, and graphene, have attracted extensive attention in wide fields. However, undesirable concerns of nanoparticles, including in higher surface area, low tap density, and self-agglomeration, greatly restricted their large-scale practical applications. In this study, novel porous micron-SnO2/C (p-SnO2/C) composites are scalable prepared by a simple hydrothermal approach using glucose as a carbon source and Pluronic F127 as a pore forming agent/soft template. The SnO2 nanoparticles were homogeneously dispersed in micron carbon spheres by assembly with F127/glucose. The continuous three-dimensional porous carbon networks have effectively provided strain relaxation for SnO2 volume expansion/shrinkage during lithium insertion/extraction. In addition, the carbon matrix could largely minimize the direct exposure of SnO2 to the electrolyte, thus ensure formation of stable solid electrolyte interface films. Moreover, the porous structure could also create efficient channels for the fast transport of lithium ions. As a consequence, the p-SnO2/C composites exhibit stable cycle performance, such as a high capacity retention of over 96% for 100 cycles at a current density of 200 mA g-1 and a long cycle life up to 800 times at a higher current density of 1000 mA g-1.

  11. Printed microelectrodes for scalable, high-areal-capacity lithium-sulfur batteries.

    PubMed

    Milroy, Craig; Manthiram, Arumugam

    2016-03-10

    Printed microelectrodes for lithium-sulfur cathodes are produced with aqueous inks and a one-step printing process. The cathodes exhibit high areal capacities of ≥5 mA h cm(-2) for 50 cycles and withstand ≥500 cycles. This performance meets energy-storage benchmarks for powering microdevices, and presents a strategic option for future microbatteries. PMID:26833188

  12. Large enhancement of quantum dot fluorescence by highly scalable nanoporous gold.

    PubMed

    Zhang, Ling; Song, Yunke; Fujita, Takeshi; Zhang, Ye; Chen, Mingwei; Wang, Tza-Huei

    2014-02-26

    Dealloyed nanoporous gold (NPG) dramatically enhances quantum dot (QD) fluorescence by amplifying near-field excitation and increasing the radiative decay rate. Originating from plasmonic coupling, the fluorescence enhancement is highly dependent upon the nanopore size of the NPG. In contrast to other nanoengineered metallic structures, NPG exhibits fluorescence enhancement of QDs over a large substrate surface. PMID:24339211

  13. Scalable Fabrication of Electrospun Nanofibrous Membranes Functionalized with Citric Acid for High-Performance Protein Adsorption.

    PubMed

    Fu, Qiuxia; Wang, Xueqin; Si, Yang; Liu, Lifang; Yu, Jianyong; Ding, Bin

    2016-05-11

    Fabricating protein adsorbents with high adsorption capacity and appreciable throughput is extremely important and highly desired for the separation and purification of protein products in the biomedical and pharmaceutical industries, yet still remains a great challenge. Herein, we demonstrate the synthesis of a novel protein adsorbent by in situ functionalizing eletrospun ethylene-vinyl alcohol (EVOH) nanofibrous membranes (NFM) with critic acid (CCA). Taking advantage of the merits of large specific surface area, highly tortuous open-porous structure, abundant active carboxyl groups introduced by CCA, superior chemical stability, and robust mechanical strength, the obtained CCA-grafted EVOH NFM (EVOH-CCA NFM) present an excellent integrated protein (take lysozyme as the model protein) adsorption performance with a high capacity of 284 mg g(-1), short equilibrium time of 6 h, ease of elution, and good reusability. Meanwhile, the adsorption performance of EVOH-CCA NFM can be optimized by regulating buffer pH, ionic strength, and initial concentration of protein solutions. More importantly, a dynamic binding efficiency of 250 mg g(-1) can be achieved driven solely by the gravity of protein solution, which matches well with the demands of the high yield and energy conservation in the actual protein purification process. Furthermore, the resultant EVOH-CCA NFM also possess unique selectivity for positively charged proteins which was confirmed by the method of sodium dodecyl sulfate polyacrylamide gel electrophoresis. Significantly, the successful synthesis of such intriguing and economic EVOH-CCA NFM may provide a promising candidate for the next generation of protein adsorbents for rapid, massive, and cost-effective separation and purification of proteins. PMID:27111287

  14. Scalable synthesis of Fe₃O₄ nanoparticles anchored on graphene as a high-performance anode for lithium ion batteries

    SciTech Connect

    Dong, Yu Cheng; Ma, Ru Guang; Jun Hu, Ming; Cheng, Hua; Tsang, Chun Kwan; Yang, Qing Dan; Yang Li, Yang; Zapien, Juan Antonio

    2013-05-01

    We report a scalable strategy to synthesize Fe₃O₄/graphene nanocomposites as a high-performance anode material for lithium ion batteries. In this study, ferric citrate is used as precursor to prepare Fe₃O₄ nanoparticles without introducing additional reducing agent; furthermore and show that such Fe₃O₄ nanoparticles can be anchored on graphene sheets which attributed to multifunctional group effect of citrate. Electrochemical characterization of the Fe₃O₄/graphene nanocomposites exhibit large reversible capacity (~1347 mA h g⁻¹ at a current density of 0.2 C up to 100 cycles, and subsequent capacity of ~619 mA h g⁻¹ at a current density of 2 C up to 200 cycles), as well as high coulombic efficiency (~97%), excellent rate capability, and good cyclic stability. High resolution transmission electron microscopy confirms that Fe₃O₄ nanoparticles, with a size of ~4–16 nm are densely anchored on thin graphene sheets, resulting in large synergetic effects between Fe₃O₄ nanoparticles and graphene sheets with high electrochemical performance. - Graphical abstract: The reduction of Fe³⁺ to Fe²⁺ and the deposition of Fe₃O₄ on graphene sheets occur simultaneously using citrate function as reductant and anchor agent in this reaction process. Highlights: • Fe₃O₄/graphene composites are synthesized directly from graphene and C₆H₅FeO₇. • The citrate function as reductant and anchor agent in this reaction process. • The resulting Fe₃O₄ particles (~4–16 nm) are densely anchored on graphene sheets. • The prepared Fe₃O₄/graphene composites exhibit excellent electrochemical performance.

  15. High-performance hollow sulfur nanostructured battery cathode through a scalable, room temperature, one-step, bottom-up approach

    PubMed Central

    Li, Weiyang; Zheng, Guangyuan; Yang, Yuan; Seh, Zhi Wei; Liu, Nian; Cui, Yi

    2013-01-01

    Sulfur is an exciting cathode material with high specific capacity of 1,673 mAh/g, more than five times the theoretical limits of its transition metal oxides counterpart. However, successful applications of sulfur cathode have been impeded by rapid capacity fading caused by multiple mechanisms, including large volume expansion during lithiation, dissolution of intermediate polysulfides, and low ionic/electronic conductivity. Tackling the sulfur cathode problems requires a multifaceted approach, which can simultaneously address the challenges mentioned above. Herein, we present a scalable, room temperature, one-step, bottom-up approach to fabricate monodisperse polymer (polyvinylpyrrolidone)-encapsulated hollow sulfur nanospheres for sulfur cathode, allowing unprecedented control over electrode design from nanoscale to macroscale. We demonstrate high specific discharge capacities at different current rates (1,179, 1,018, and 990 mAh/g at C/10, C/5, and C/2, respectively) and excellent capacity retention of 77.6% (at C/5) and 73.4% (at C/2) after 300 and 500 cycles, respectively. Over a long-term cycling of 1,000 cycles at C/2, a capacity decay as low as 0.046% per cycle and an average coulombic efficiency of 98.5% was achieved. In addition, a simple modification on the sulfur nanosphere surface with a layer of conducting polymer, poly(3,4-ethylenedioxythiophene), allows the sulfur cathode to achieve excellent high-rate capability, showing a high reversible capacity of 849 and 610 mAh/g at 2C and 4C, respectively. PMID:23589875

  16. Scalable Computational Methods for the Analysis of High-Throughput Biological Data

    SciTech Connect

    Langston, Michael A

    2012-09-06

    This primary focus of this research project is elucidating genetic regulatory mechanisms that control an organism's responses to low-dose ionizing radiation. Although low doses (at most ten centigrays) are not lethal to humans, they elicit a highly complex physiological response, with the ultimate outcome in terms of risk to human health unknown. The tools of molecular biology and computational science will be harnessed to study coordinated changes in gene expression that orchestrate the mechanisms a cell uses to manage the radiation stimulus. High performance implementations of novel algorithms that exploit the principles of fixed-parameter tractability will be used to extract gene sets suggestive of co-regulation. Genomic mining will be performed to scrutinize, winnow and highlight the most promising gene sets for more detailed investigation. The overall goal is to increase our understanding of the health risks associated with exposures to low levels of radiation.

  17. Fast, scalable generation of high-quality protein multiple sequence alignments using Clustal Omega

    PubMed Central

    Sievers, Fabian; Wilm, Andreas; Dineen, David; Gibson, Toby J; Karplus, Kevin; Li, Weizhong; Lopez, Rodrigo; McWilliam, Hamish; Remmert, Michael; Söding, Johannes; Thompson, Julie D; Higgins, Desmond G

    2011-01-01

    Multiple sequence alignments are fundamental to many sequence analysis methods. Most alignments are computed using the progressive alignment heuristic. These methods are starting to become a bottleneck in some analysis pipelines when faced with data sets of the size of many thousands of sequences. Some methods allow computation of larger data sets while sacrificing quality, and others produce high-quality alignments, but scale badly with the number of sequences. In this paper, we describe a new program called Clustal Omega, which can align virtually any number of protein sequences quickly and that delivers accurate alignments. The accuracy of the package on smaller test cases is similar to that of the high-quality aligners. On larger data sets, Clustal Omega outperforms other packages in terms of execution time and quality. Clustal Omega also has powerful features for adding sequences to and exploiting information in existing alignments, making use of the vast amount of precomputed information in public databases like Pfam. PMID:21988835

  18. Lightweight, flexible, high-performance carbon nanotube cables made by scalable flow coating

    DOE PAGESBeta

    Mirri, Francesca; Orloff, Nathan D.; Forser, Aaron M.; Ashkar, Rana; Headrick, Robert J.; Bengio, E. Amram; Long, Christian J.; Choi, April; Luo, Yimin; Hight Walker, Angela R.; et al

    2016-01-21

    Coaxial cables for data transmission are ubiquitous in telecommunications, aerospace, automotive, and robotics industries. Yet, the metals used to make commercial cables are unsuitably heavy and stiff. These undesirable traits are particularly problematic in aerospace applications, where weight is at a premium and flexibility is necessary to conform with the distributed layout of electronic components in satellites and aircraft. The cable outer conductor (OC) is usually the heaviest component of modern data cables; therefore, exchanging the conventional metallic OC for lower weight materials with comparable transmission characteristics is highly desirable. Carbon nanotubes (CNTs) have recently been proposed to replace themore » metal components in coaxial cables; however, signal attenuation was too high in prototypes produced so far. Here, we fabricate the OC of coaxial data cables by directly coating a solution of CNTs in chlorosulfonic acid (CSA) onto the cable inner dielectric. This coating has an electrical conductivity that is approximately 2 orders of magnitude greater than the best CNT OC reported in the literature to date. In conclusion, this high conductivity makes CNT coaxial cables an attractive alternative to commercial cables with a metal (tin-coated copper) OC, providing comparable cable attenuation and mechanical durability with a 97% lower component mass.« less

  19. Generation of Scalable, Metallic High-Aspect Ratio Nanocomposites in a Biological Liquid Medium.

    PubMed

    Cotton Kelly, Kinsey; Wasserman, Jessica R; Deodhar, Sneha; Huckaby, Justin; DeCoster, Mark A

    2015-01-01

    The goal of this protocol is to describe the synthesis of two novel biocomposites with high-aspect ratio structures. The biocomposites consist of copper and cystine, with either copper nanoparticles (CNPs) or copper sulfate contributing the metallic component. Synthesis is carried out in liquid under biological conditions (37 °C) and the self-assembled composites form after 24 hr. Once formed, these composites are highly stable in both liquid media and in a dried form. The composites scale from the nano- to micro- range in length, and from a few microns to 25 nm in diameter. Field emission scanning electron microscopy with energy dispersive X-ray spectroscopy (EDX) demonstrated that sulfur was present in the NP-derived linear structures, while it was absent from the starting CNP material, thus confirming cystine as the source of sulfur in the final nanocomposites. During synthesis of these linear nano- and micro-composites, a diverse range of lengths of structures is formed in the synthesis vessel. Sonication of the liquid mixture after synthesis was demonstrated to assist in controlling average size of the structures by diminishing the average length with increased time of sonication. Since the formed structures are highly stable, do not agglomerate, and are formed in liquid phase, centrifugation may also be used to assist in concentrating and segregating formed composites. PMID:26274773

  20. Homogenous 96-Plex PEA Immunoassay Exhibiting High Sensitivity, Specificity, and Excellent Scalability

    PubMed Central

    Holmquist, Göran; Björkesten, Johan; Bucht Thorsen, Stine; Ekman, Daniel; Eriksson, Anna; Rennel Dickens, Emma; Ohlsson, Sandra; Edfeldt, Gabriella; Andersson, Ann-Catrin; Lindstedt, Patrik; Stenvang, Jan; Gullberg, Mats; Fredriksson, Simon

    2014-01-01

    Medical research is developing an ever greater need for comprehensive high-quality data generation to realize the promises of personalized health care based on molecular biomarkers. The nucleic acid proximity-based methods proximity ligation and proximity extension assays have, with their dual reporters, shown potential to relieve the shortcomings of antibodies and their inherent cross-reactivity in multiplex protein quantification applications. The aim of the present study was to develop a robust 96-plex immunoassay based on the proximity extension assay (PEA) for improved high throughput detection of protein biomarkers. This was enabled by: (1) a modified design leading to a reduced number of pipetting steps compared to the existing PEA protocol, as well as improved intra-assay precision; (2) a new enzymatic system that uses a hyper-thermostabile enzyme, Pwo, for uniting the two probes allowing for room temperature addition of all reagents and improved the sensitivity; (3) introduction of an inter-plate control and a new normalization procedure leading to improved inter-assay precision (reproducibility). The multiplex proximity extension assay was found to perform well in complex samples, such as serum and plasma, and also in xenografted mice and resuspended dried blood spots, consuming only 1 µL sample per test. All-in-all, the development of the current multiplex technique is a step toward robust high throughput protein marker discovery and research. PMID:24755770

  1. Lightweight, Flexible, High-Performance Carbon Nanotube Cables Made by Scalable Flow Coating.

    PubMed

    Mirri, Francesca; Orloff, Nathan D; Forster, Aaron M; Ashkar, Rana; Headrick, Robert J; Bengio, E Amram; Long, Christian J; Choi, April; Luo, Yimin; Walker, Angela R Hight; Butler, Paul; Migler, Kalman B; Pasquali, Matteo

    2016-02-01

    Coaxial cables for data transmission are ubiquitous in telecommunications, aerospace, automotive, and robotics industries. Yet, the metals used to make commercial cables are unsuitably heavy and stiff. These undesirable traits are particularly problematic in aerospace applications, where weight is at a premium and flexibility is necessary to conform with the distributed layout of electronic components in satellites and aircraft. The cable outer conductor (OC) is usually the heaviest component of modern data cables; therefore, exchanging the conventional metallic OC for lower weight materials with comparable transmission characteristics is highly desirable. Carbon nanotubes (CNTs) have recently been proposed to replace the metal components in coaxial cables; however, signal attenuation was too high in prototypes produced so far. Here, we fabricate the OC of coaxial data cables by directly coating a solution of CNTs in chlorosulfonic acid (CSA) onto the cable inner dielectric. This coating has an electrical conductivity that is approximately 2 orders of magnitude greater than the best CNT OC reported in the literature to date. This high conductivity makes CNT coaxial cables an attractive alternative to commercial cables with a metal (tin-coated copper) OC, providing comparable cable attenuation and mechanical durability with a 97% lower component mass. PMID:26791337

  2. Complexity in scalable computing.

    SciTech Connect

    Rouson, Damian W. I.

    2008-12-01

    The rich history of scalable computing research owes much to a rapid rise in computing platform scale in terms of size and speed. As platforms evolve, so must algorithms and the software expressions of those algorithms. Unbridled growth in scale inevitably leads to complexity. This special issue grapples with two facets of this complexity: scalable execution and scalable development. The former results from efficient programming of novel hardware with increasing numbers of processing units (e.g., cores, processors, threads or processes). The latter results from efficient development of robust, flexible software with increasing numbers of programming units (e.g., procedures, classes, components or developers). The progression in the above two parenthetical lists goes from the lowest levels of abstraction (hardware) to the highest (people). This issue's theme encompasses this entire spectrum. The lead author of each article resides in the Scalable Computing Research and Development Department at Sandia National Laboratories in Livermore, CA. Their co-authors hail from other parts of Sandia, other national laboratories and academia. Their research sponsors include several programs within the Department of Energy's Office of Advanced Scientific Computing Research and its National Nuclear Security Administration, along with Sandia's Laboratory Directed Research and Development program and the Office of Naval Research. The breadth of interests of these authors and their customers reflects in the breadth of applications this issue covers. This article demonstrates how to obtain scalable execution on the increasingly dominant high-performance computing platform: a Linux cluster with multicore chips. The authors describe how deep memory hierarchies necessitate reducing communication overhead by using threads to exploit shared register and cache memory. On a matrix-matrix multiplication problem, they achieve up to 96% parallel efficiency with a three-part strategy: intra

  3. Bottom-up, hard template and scalable approaches toward designing nanostructured Li2S for high performance lithium sulfur batteries

    NASA Astrophysics Data System (ADS)

    Chen, Lin; Liu, Yuzi; Dietz-Rago, Nancy; Shaw, Leon L.

    2015-10-01

    Li2S with a high theoretical capacity of 1166 mA h g-1 and the capability to pair with lithium free anodes has drawn much attention for lithium sulfur (Li-S) battery applications. However, the fast battery decay and the low capacity retention due to dissolution of intermediate polysulfides in electrolytes limit its development. Designing a nanosized and nanostructured host for Li2S through facile techniques is one of the ways to alleviate the dissolution and improve Li-S battery performance; nevertheless, it is technically difficult to synthesize nanosized and nanostructured hosts for Li2S because Li2S is highly sensitive to moisture and oxygen. Herein, a novel technique, i.e., a bottom-up, hard template and scalable method, is proposed to engineer nanoLi2S composites with core-shell structures as cathodes of Li-S batteries. The size of the as-prepared nanostructured Li2S is around 100 nm. With the assistance of FETEM, HRTEM and EFTEM elemental mapping, an excellent core-shell structure has been confirmed and the outside carbon shell has a thickness of 20-50 nm, effectively retarding polysulfide outflow and dissolution. A high initial capacity of 915 mA h g-1 at 0.2 C has been achieved upon electrochemical cycling and the battery still has exceptional capacity retention after prolonged 200 cycles with a limited decay of 0.18% per cycle. Also, at 0.5 C the electrode exhibits 60% capacity retention with a long life of 300 cycles. We attribute these good performances to the nano-architecture constructed by the novel and facile method.Li2S with a high theoretical capacity of 1166 mA h g-1 and the capability to pair with lithium free anodes has drawn much attention for lithium sulfur (Li-S) battery applications. However, the fast battery decay and the low capacity retention due to dissolution of intermediate polysulfides in electrolytes limit its development. Designing a nanosized and nanostructured host for Li2S through facile techniques is one of the ways to alleviate

  4. Versatile, High Quality and Scalable Continuous Flow Production of Metal-Organic Frameworks

    PubMed Central

    Rubio-Martinez, Marta; Batten, Michael P.; Polyzos, Anastasios; Carey, Keri-Constanti; Mardel, James I.; Lim, Kok-Seng; Hill, Matthew R.

    2014-01-01

    Further deployment of Metal-Organic Frameworks in applied settings requires their ready preparation at scale. Expansion of typical batch processes can lead to unsuccessful or low quality synthesis for some systems. Here we report how continuous flow chemistry can be adapted as a versatile route to a range of MOFs, by emulating conditions of lab-scale batch synthesis. This delivers ready synthesis of three different MOFs, with surface areas that closely match theoretical maxima, with production rates of 60 g/h at extremely high space-time yields. PMID:24962145

  5. Scalable Memory Registration for High-Performance Networks Using Helper Threads

    SciTech Connect

    Li, Dong; Cameron, Kirk W.; Nikolopoulos, Dimitrios; de Supinski, Bronis R.; Schulz, Martin

    2011-01-01

    Remote DMA (RDMA) enables high performance networks to reduce data copying between an application and the operating system (OS). However RDMA operations in some high performance networks require communication memory explicitly registered with the network adapter and pinned by the OS. Memory registration and pinning limits the flexibility of the memory system and reduces the amount of memory that user processes can allocate. These issues become more significant on multicore platforms, since registered memory demand grows linearly with the number of processor cores. In this paper we propose a new memory registration/deregistration strategy to reduce registered memory on multicore architectures for HPC applications. We hide the cost of dynamic memory management by offloading all dynamic memory registration and deregistration requests to a dedicated memory management helper thread. We investigate design policies and performance implications of the helper thread approach. We evaluate our framework with the NAS parallel benchmarks, for which our registration scheme significantly reduces the registered memory (23.62% on average and up to 49.39%) and avoids memory registration/deregistration costs for reused communication memory. We show that our system enables the execution of problem sizes that could not complete under existing memory registration strategies.

  6. Bottom-up, hard template and scalable approaches toward designing nanostructured Li2S for high performance lithium sulfur batteries.

    PubMed

    Chen, Lin; Liu, Yuzi; Dietz-Rago, Nancy; Shaw, Leon L

    2015-11-21

    Li2S with a high theoretical capacity of 1166 mA h g(-1) and the capability to pair with lithium free anodes has drawn much attention for lithium sulfur (Li-S) battery applications. However, the fast battery decay and the low capacity retention due to dissolution of intermediate polysulfides in electrolytes limit its development. Designing a nanosized and nanostructured host for Li2S through facile techniques is one of the ways to alleviate the dissolution and improve Li-S battery performance; nevertheless, it is technically difficult to synthesize nanosized and nanostructured hosts for Li2S because Li2S is highly sensitive to moisture and oxygen. Herein, a novel technique, i.e., a bottom-up, hard template and scalable method, is proposed to engineer nanoLi2S composites with core-shell structures as cathodes of Li-S batteries. The size of the as-prepared nanostructured Li2S is around 100 nm. With the assistance of FETEM, HRTEM and EFTEM elemental mapping, an excellent core-shell structure has been confirmed and the outside carbon shell has a thickness of 20-50 nm, effectively retarding polysulfide outflow and dissolution. A high initial capacity of 915 mA h g(-1) at 0.2 C has been achieved upon electrochemical cycling and the battery still has exceptional capacity retention after prolonged 200 cycles with a limited decay of 0.18% per cycle. Also, at 0.5 C the electrode exhibits 60% capacity retention with a long life of 300 cycles. We attribute these good performances to the nano-architecture constructed by the novel and facile method. PMID:26420373

  7. K-Boost: a scalable algorithm for high-quality clustering of microarray gene expression data.

    PubMed

    Geraci, Filippo; Leoncini, Mauro; Montangero, Manuela; Pellegrini, Marco; Renda, M Elena

    2009-06-01

    Microarray technology for profiling gene expression levels is a popular tool in modern biological research. Applications range from tissue classification to the detection of metabolic networks, from drug discovery to time-critical personalized medicine. Given the increase in size and complexity of the data sets produced, their analysis is becoming problematic in terms of time/quality trade-offs. Clustering genes with similar expression profiles is a key initial step for subsequent manipulations and the increasing volumes of data to be analyzed requires methods that are at the same time efficient (completing an analysis in minutes rather than hours) and effective (identifying significant clusters with high biological correlations). In this paper, we propose K-Boost, a clustering algorithm based on a combination of the furthest-point-first (FPF) heuristic for solving the metric k-center problem, a stability-based method for determining the number of clusters, and a k-means-like cluster refinement. K-Boost runs in O (|N| x k) time, where N is the input matrix and k is the number of proposed clusters. Experiments show that this low complexity is usually coupled with a very good quality of the computed clusterings, which we measure using both internal and external criteria. Supporting data can be found as online Supplementary Material at www.liebertonline.com. PMID:19522668

  8. A power scalable PLL frequency synthesizer for high-speed Δ—Σ ADC

    NASA Astrophysics Data System (ADS)

    Siyang, Han; Baoyong, Chi; Xinwang, Zhang; Zhihua, Wang

    2014-08-01

    A 35-130 MHz/300-360 MHz phase-locked loop frequency synthesizer for Δ—Σ analog-to-digital converter (ADC) in 65 nm CMOS is presented. The frequency synthesizer can work in low phase-noise mode (300-360 MHz) or in low-power mode (35-130 MHz) to satisfy the ADC's requirements. To switch between these two modes, a high frequency GHz LC VCO followed by a divided-by-four frequency divider and a low frequency ring VCO followed by a divided-by-two frequency divider are integrated on-chip. The measured results show that the frequency synthesizer achieves a phase-noise of -132 dBc/Hz at 1 MHz offset and an integrated RMS jitter of 1.12 ps with 1.74 mW power consumption from a 1.2 V power supply in low phase-noise mode. In low-power mode, the frequency synthesizer achieves a phase-noise of -112 dBc/Hz at 1 MHz offset and an integrated RMS jitter of 7.23 ps with 0.92 mW power consumption from a 1.2 V power supply.

  9. Towards Scalable Cost-Effective Service and Survivability Provisioning in Ultra High Speed Networks

    SciTech Connect

    Bin Wang

    2006-12-01

    Optical transport networks based on wavelength division multiplexing (WDM) are considered to be the most appropriate choice for future Internet backbone. On the other hand, future DOE networks are expected to have the ability to dynamically provision on-demand survivable services to suit the needs of various high performance scientific applications and remote collaboration. Since a failure in aWDMnetwork such as a cable cut may result in a tremendous amount of data loss, efficient protection of data transport in WDM networks is therefore essential. As the backbone network is moving towards GMPLS/WDM optical networks, the unique requirement to support DOE’s science mission results in challenging issues that are not directly addressed by existing networking techniques and methodologies. The objectives of this project were to develop cost effective protection and restoration mechanisms based on dedicated path, shared path, preconfigured cycle (p-cycle), and so on, to deal with single failure, dual failure, and shared risk link group (SRLG) failure, under different traffic and resource requirement models; to devise efficient service provisioning algorithms that deal with application specific network resource requirements for both unicast and multicast; to study various aspects of traffic grooming in WDM ring and mesh networks to derive cost effective solutions while meeting application resource and QoS requirements; to design various diverse routing and multi-constrained routing algorithms, considering different traffic models and failure models, for protection and restoration, as well as for service provisioning; to propose and study new optical burst switched architectures and mechanisms for effectively supporting dynamic services; and to integrate research with graduate and undergraduate education. All objectives have been successfully met. This report summarizes the major accomplishments of this project. The impact of the project manifests in many aspects: First

  10. SFT: Scalable Fault Tolerance

    SciTech Connect

    Petrini, Fabrizio; Nieplocha, Jarek; Tipparaju, Vinod

    2006-04-15

    In this paper we will present a new technology that we are currently developing within the SFT: Scalable Fault Tolerance FastOS project which seeks to implement fault tolerance at the operating system level. Major design goals include dynamic reallocation of resources to allow continuing execution in the presence of hardware failures, very high scalability, high efficiency (low overhead), and transparency—requiring no changes to user applications. Our technology is based on a global coordination mechanism, that enforces transparent recovery lines in the system, and TICK, a lightweight, incremental checkpointing software architecture implemented as a Linux kernel module. TICK is completely user-transparent and does not require any changes to user code or system libraries; it is highly responsive: an interrupt, such as a timer interrupt, can trigger a checkpoint in as little as 2.5μs; and it supports incremental and full checkpoints with minimal overhead—less than 6% with full checkpointing to disk performed as frequently as once per minute.

  11. Anatomically accurate high resolution modeling of human whole heart electromechanics: A strongly scalable algebraic multigrid solver method for nonlinear deformation

    PubMed Central

    Augustin, Christoph M.; Neic, Aurel; Liebmann, Manfred; Prassl, Anton J.; Niederer, Steven A.; Haase, Gundolf; Plank, Gernot

    2016-01-01

    Electromechanical (EM) models of the heart have been used successfully to study fundamental mechanisms underlying a heart beat in health and disease. However, in all modeling studies reported so far numerous simplifications were made in terms of representing biophysical details of cellular function and its heterogeneity, gross anatomy and tissue microstructure, as well as the bidirectional coupling between electrophysiology (EP) and tissue distension. One limiting factor is the employed spatial discretization methods which are not sufficiently flexible to accommodate complex geometries or resolve heterogeneities, but, even more importantly, the limited efficiency of the prevailing solver techniques which are not sufficiently scalable to deal with the incurring increase in degrees of freedom (DOF) when modeling cardiac electromechanics at high spatio-temporal resolution. This study reports on the development of a novel methodology for solving the nonlinear equation of finite elasticity using human whole organ models of cardiac electromechanics, discretized at a high para-cellular resolution. Three patient-specific, anatomically accurate, whole heart EM models were reconstructed from magnetic resonance (MR) scans at resolutions of 220 μm, 440 μm and 880 μm, yielding meshes of approximately 184.6, 24.4 and 3.7 million tetrahedral elements and 95.9, 13.2 and 2.1 million displacement DOF, respectively. The same mesh was used for discretizing the governing equations of both electrophysiology (EP) and nonlinear elasticity. A novel algebraic multigrid (AMG) preconditioner for an iterative Krylov solver was developed to deal with the resulting computational load. The AMG preconditioner was designed under the primary objective of achieving favorable strong scaling characteristics for both setup and solution runtimes, as this is key for exploiting current high performance computing hardware. Benchmark results using the 220 μm, 440 μm and 880 μm meshes demonstrate

  12. Anatomically accurate high resolution modeling of human whole heart electromechanics: A strongly scalable algebraic multigrid solver method for nonlinear deformation

    NASA Astrophysics Data System (ADS)

    Augustin, Christoph M.; Neic, Aurel; Liebmann, Manfred; Prassl, Anton J.; Niederer, Steven A.; Haase, Gundolf; Plank, Gernot

    2016-01-01

    Electromechanical (EM) models of the heart have been used successfully to study fundamental mechanisms underlying a heart beat in health and disease. However, in all modeling studies reported so far numerous simplifications were made in terms of representing biophysical details of cellular function and its heterogeneity, gross anatomy and tissue microstructure, as well as the bidirectional coupling between electrophysiology (EP) and tissue distension. One limiting factor is the employed spatial discretization methods which are not sufficiently flexible to accommodate complex geometries or resolve heterogeneities, but, even more importantly, the limited efficiency of the prevailing solver techniques which is not sufficiently scalable to deal with the incurring increase in degrees of freedom (DOF) when modeling cardiac electromechanics at high spatio-temporal resolution. This study reports on the development of a novel methodology for solving the nonlinear equation of finite elasticity using human whole organ models of cardiac electromechanics, discretized at a high para-cellular resolution. Three patient-specific, anatomically accurate, whole heart EM models were reconstructed from magnetic resonance (MR) scans at resolutions of 220 μm, 440 μm and 880 μm, yielding meshes of approximately 184.6, 24.4 and 3.7 million tetrahedral elements and 95.9, 13.2 and 2.1 million displacement DOF, respectively. The same mesh was used for discretizing the governing equations of both electrophysiology (EP) and nonlinear elasticity. A novel algebraic multigrid (AMG) preconditioner for an iterative Krylov solver was developed to deal with the resulting computational load. The AMG preconditioner was designed under the primary objective of achieving favorable strong scaling characteristics for both setup and solution runtimes, as this is key for exploiting current high performance computing hardware. Benchmark results using the 220 μm, 440 μm and 880 μm meshes demonstrate

  13. Highly scalable, uniform, and sensitive biosensors based on top-down indium oxide nanoribbons and electronic enzyme-linked immunosorbent assay.

    PubMed

    Aroonyadet, Noppadol; Wang, Xiaoli; Song, Yan; Chen, Haitian; Cote, Richard J; Thompson, Mark E; Datar, Ram H; Zhou, Chongwu

    2015-03-11

    Nanostructure field-effect transistor (FET) biosensors have shown great promise for ultra sensitive biomolecular detection. Top-down assembly of these sensors increases scalability and device uniformity but faces fabrication challenges in achieving the small dimensions needed for sensitivity. We report top-down fabricated indium oxide (In2O3) nanoribbon FET biosensors using highly scalable radio frequency (RF) sputtering to create uniform channel thicknesses ranging from 50 to 10 nm. We combine this scalable sensing platform with amplification from electronic enzyme-linked immunosorbent assay (ELISA) to achieve high sensitivity to target analytes such as streptavidin and human immunodeficiency virus type 1 (HIV-1) p24 proteins. Our approach circumvents Debye screening in ionic solutions and detects p24 protein at 20 fg/mL (about 250 viruses/mL or about 3 orders of magnitude lower than commercial ELISA) with a 35% conduction change in human serum. The In2O3 nanoribbon biosensors have 100% device yield and use a simple 2 mask photolithography process. The electrical properties of 50 In2O3 nanoribbon FETs showed good uniformity in on-state current, on/off current ratio, mobility, and threshold voltage. In addition, the sensors show excellent pH sensitivity over a broad range (pH 4 to 9) as well as over the physiological-related pH range (pH 6.8 to 8.2). With the demonstrated sensitivity, scalability, and uniformity, the In2O3 nanoribbon sensor platform makes great progress toward clinical testing, such as for early diagnosis of acquired immunodeficiency syndrome (AIDS). PMID:25636984

  14. Scalable and High-Throughput Execution of Clinical Quality Measures from Electronic Health Records using MapReduce and the JBoss® Drools Engine

    PubMed Central

    Peterson, Kevin J.; Pathak, Jyotishman

    2014-01-01

    Automated execution of electronic Clinical Quality Measures (eCQMs) from electronic health records (EHRs) on large patient populations remains a significant challenge, and the testability, interoperability, and scalability of measure execution are critical. The High Throughput Phenotyping (HTP; http://phenotypeportal.org) project aligns with these goals by using the standards-based HL7 Health Quality Measures Format (HQMF) and Quality Data Model (QDM) for measure specification, as well as Common Terminology Services 2 (CTS2) for semantic interpretation. The HQMF/QDM representation is automatically transformed into a JBoss® Drools workflow, enabling horizontal scalability via clustering and MapReduce algorithms. Using Project Cypress, automated verification metrics can then be produced. Our results show linear scalability for nine executed 2014 Center for Medicare and Medicaid Services (CMS) eCQMs for eligible professionals and hospitals for >1,000,000 patients, and verified execution correctness of 96.4% based on Project Cypress test data of 58 eCQMs. PMID:25954459

  15. Application of the FETI Method to ASCI Problems: Scalability Results on One Thousand Processors and Discussion of Highly Heterogeneous Problems

    SciTech Connect

    Bhardwaj, M.; Day, D.; Farhat, C.; Lesoinne, M; Pierson, K.; Rixen, D.

    1999-04-01

    We report on the application of the one-level FETI method to the solution of a class of substructural problems associated with the Department of Energy's Accelerated Strategic Computing Initiative (ASCI). We focus on numerical and parallel scalability issues, and on preliminary performance results obtained on the ASCI Option Red supercomputer configured with as many as one thousand processors, for problems with as many as 5 million degrees of freedom.

  16. Scalable Node Monitoring

    SciTech Connect

    Drotar, Alexander P.; Quinn, Erin E.; Sutherland, Landon D.

    2012-07-30

    Project description is: (1) Build a high performance computer; and (2) Create a tool to monitor node applications in Component Based Tool Framework (CBTF) using code from Lightweight Data Metric Service (LDMS). The importance of this project is that: (1) there is a need a scalable, parallel tool to monitor nodes on clusters; and (2) New LDMS plugins need to be able to be easily added to tool. CBTF stands for Component Based Tool Framework. It's scalable and adjusts to different topologies automatically. It uses MRNet (Multicast/Reduction Network) mechanism for information transport. CBTF is flexible and general enough to be used for any tool that needs to do a task on many nodes. Its components are reusable and 'EASILY' added to a new tool. There are three levels of CBTF: (1) frontend node - interacts with users; (2) filter nodes - filters or concatenates information from backend nodes; and (3) backend nodes - where the actual work of the tool is done. LDMS stands for lightweight data metric servies. It's a tool used for monitoring nodes. Ltool is the name of the tool we derived from LDMS. It's dynamically linked and includes the following components: Vmstat, Meminfo, Procinterrupts and more. It works by: Ltool command is run on the frontend node; Ltool collects information from the backend nodes; backend nodes send information to the filter nodes; and filter nodes concatenate information and send to a database on the front end node. Ltool is a useful tool when it comes to monitoring nodes on a cluster because the overhead involved with running the tool is not particularly high and it will automatically scale to any size cluster.

  17. Volume server: A scalable high speed and high capacity magnetic tape archive architecture with concurrent multi-host access

    NASA Technical Reports Server (NTRS)

    Rybczynski, Fred

    1993-01-01

    A major challenge facing data processing centers today is data management. This includes the storage of large volumes of data and access to it. Current media storage for large data volumes is typically off line and frequently off site in warehouses. Access to data archived in this fashion can be subject to long delays, errors in media selection and retrieval, and even loss of data through misplacement or damage to the media. Similarly, designers responsible for architecting systems capable of continuous high-speed recording of large volumes of digital data are faced with the challenge of identifying technologies and configurations that meet their requirements. Past approaches have tended to evaluate the combination of the fastest tape recorders with the highest capacity tape media and then to compromise technology selection as a consequence of cost. This paper discusses an architecture that addresses both of these challenges and proposes a cost effective solution based on robots, high speed helical scan tape drives, and large-capacity media.

  18. Development of a Scalable, High-Throughput-Compatible Assay to Detect Tau Aggregates Using iPSC-Derived Cortical Neurons Maintained in a Three-Dimensional Culture Format.

    PubMed

    Medda, X; Mertens, L; Versweyveld, S; Diels, A; Barnham, L; Bretteville, A; Buist, A; Verheyen, A; Royaux, I; Ebneth, A; Cabrera-Socorro, A

    2016-09-01

    Tau aggregation is the pathological hallmark that best correlates with the progression of Alzheimer's disease (AD). The presence of neurofibrillary tangles (NFTs), formed of hyperphosphorylated tau, leads to neuronal dysfunction and loss, and is directly associated with the cognitive decline observed in AD patients. The limited success in targeting β-amyloid pathologies has reinforced the hypothesis of blocking tau phosphorylation, aggregation, and/or spreading as alternative therapeutic entry points to treat AD. Identification of novel therapies requires disease-relevant and scalable assays capable of reproducing key features of the pathology in an in vitro setting. Here we use induced pluripotent stem cells (iPSCs) as a virtually unlimited source of human cortical neurons to develop a robust and scalable tau aggregation model compatible with high-throughput screening (HTS). We downscaled cell culture conditions to 384-well plate format and used Matrigel to introduce an extra physical protection against cell detachment that reduces shearing stress and better recapitulates pathological conditions. We complemented the assay with AlphaLISA technology for the detection of tau aggregates in a high-throughput-compatible format. The assay is reproducible across users and works with different commercially available iPSC lines, representing a highly translational tool for the identification of novel treatments against tauopathies, including AD. PMID:26984927

  19. A scalable high-energy diode-pumped solid state laser for laser-plasma interaction science and applications

    NASA Astrophysics Data System (ADS)

    De Vido, M.; Ertel, K.; Mason, P. D.; Banerjee, S.; Phillips, P. J.; Butcher, T. J.; Smith, J. M.; Shaikh, W.; Hernandez-Gomes, C.; Greenhalgh, R. J. S.; Collier, J. L.

    2016-05-01

    Laser systems efficiently generating nanosecond pules at kJ energy levels and at multi-Hz repetition rates are required in order to translate laser-plasma interactions into practical applications. We have developed a scalable, actively-cooled diode-pumped solid state laser amplifier design based on a multi-slab ceramic Yb:YAG architecture called DiPOLE (Diode-Pumped Optical Laser for Experiments) capable of meeting such requirements. We demonstrated 10.8 J, 10 Hz operation at 1030 nm using a scaled-down prototype, reaching an optical-to-optical efficiency of 22.5%. Preliminary results from a larger scale version, delivering 100 J pulse energy at 10 Hz, are also presented.

  20. Scalable Synthesis of Few-Layer MoS2 Incorporated into Hierarchical Porous Carbon Nanosheets for High-Performance Li- and Na-Ion Battery Anodes.

    PubMed

    Park, Seung-Keun; Lee, Jeongyeon; Bong, Sungyool; Jang, Byungchul; Seong, Kwang-Dong; Piao, Yuanzhe

    2016-08-01

    It is still a challenging task to develop a facile and scalable process to synthesize porous hybrid materials with high electrochemical performance. Herein, a scalable strategy is developed for the synthesis of few-layer MoS2 incorporated into hierarchical porous carbon (MHPC) nanosheet composites as anode materials for both Li- (LIB) and Na-ion battery (SIB). An inexpensive oleylamine (OA) is introduced to not only serve as a hinder the stacking of MoS2 nanosheets but also to provide a conductive carbon, allowing large scale production. In addition, a SiO2 template is adopted to direct the growth of both carbon and MoS2 nanosheets, resulting in the formation of hierarchical porous structures with interconnected networks. Due to these unique features, the as-obtained MHPC shows substantial reversible capacity and very long cycling performance when used as an anode material for LIBs and SIBs, even at high current density. Indeed, this material delivers reversible capacities of 732 and 280 mA h g(-1) after 300 cycles at 1 A g(-1) in LIBs and SIBs, respectively. The results suggest that these MHPC composites also have tremendous potential for applications in other fields. PMID:27406553

  1. OneBac: Platform for Scalable and High-Titer Production of Adeno-Associated Virus Serotype 1–12 Vectors for Gene Therapy

    PubMed Central

    Mietzsch, Mario; Grasse, Sabrina; Zurawski, Catherine; Weger, Stefan; Bennett, Antonette; Agbandje-McKenna, Mavis; Muzyczka, Nicholas; Zolotukhin, Sergei

    2014-01-01

    Abstract Scalable and genetically stable recombinant adeno-associated virus (rAAV) production systems combined with facile adaptability for an extended repertoire of AAV serotypes are required to keep pace with the rapidly increasing clinical demand. For scalable high-titer production of the full range of rAAV serotypes 1–12, we developed OneBac, consisting of stable insect Sf9 cell lines harboring silent copies of AAV1–12 rep and cap genes induced upon infection with a single baculovirus that also carries the rAAV genome. rAAV burst sizes reach up to 5×105 benzonase-resistant, highly infectious genomic particles per cell, exceeding typical yields of current rAAV production systems. In contrast to recombinant rep/cap baculovirus strains currently employed for large-scale rAAV production, the Sf9rep/cap cell lines are genetically stable, leading to undiminished rAAV burst sizes over serial passages. Thus, OneBac combines full AAV serotype options with the capacity for stable scale-up production, the current bottleneck for the transition of AAV from gene therapy trials to routine clinical treatment. PMID:24299301

  2. Scalable synthesis of interconnected porous silicon/carbon composites by the Rochow reaction as high-performance anodes of lithium ion batteries.

    PubMed

    Zhang, Zailei; Wang, Yanhong; Ren, Wenfeng; Tan, Qiangqiang; Chen, Yunfa; Li, Hong; Zhong, Ziyi; Su, Fabing

    2014-05-12

    Despite the promising application of porous Si-based anodes in future Li ion batteries, the large-scale synthesis of these materials is still a great challenge. A scalable synthesis of porous Si materials is presented by the Rochow reaction, which is commonly used to produce organosilane monomers for synthesizing organosilane products in chemical industry. Commercial Si microparticles reacted with gas CH3 Cl over various Cu-based catalyst particles to substantially create macropores within the unreacted Si accompanying with carbon deposition to generate porous Si/C composites. Taking advantage of the interconnected porous structure and conductive carbon-coated layer after simple post treatment, these composites as anodes exhibit high reversible capacity and long cycle life. It is expected that by integrating the organosilane synthesis process and controlling reaction conditions, the manufacture of porous Si-based anodes on an industrial scale is highly possible. PMID:24700513

  3. Scalable fabrication of high-power graphene micro-supercapacitors for flexible and on-chip energy storage.

    PubMed

    El-Kady, Maher F; Kaner, Richard B

    2013-01-01

    The rapid development of miniaturized electronic devices has increased the demand for compact on-chip energy storage. Microscale supercapacitors have great potential to complement or replace batteries and electrolytic capacitors in a variety of applications. However, conventional micro-fabrication techniques have proven to be cumbersome in building cost-effective micro-devices, thus limiting their widespread application. Here we demonstrate a scalable fabrication of graphene micro-supercapacitors over large areas by direct laser writing on graphite oxide films using a standard LightScribe DVD burner. More than 100 micro-supercapacitors can be produced on a single disc in 30 min or less. The devices are built on flexible substrates for flexible electronics and on-chip uses that can be integrated with MEMS or CMOS in a single chip. Remarkably, miniaturizing the devices to the microscale results in enhanced charge-storage capacity and rate capability. These micro-supercapacitors demonstrate a power density of ~200 W cm-3, which is among the highest values achieved for any supercapacitor. PMID:23403576

  4. Highly flexible, transparent and self-cleanable superhydrophobic films prepared by a facile and scalable nanopyramid formation technique

    NASA Astrophysics Data System (ADS)

    Kong, Jeong-Ho; Kim, Tae-Hyun; Kim, Ji Hoon; Park, Jong-Kweon; Lee, Deug-Woo; Kim, Soo-Hyung; Kim, Jong-Man

    2014-01-01

    A facile and scalable technique to fabricate optically transparent, mechanically flexible and self-cleanable superhydrophobic films for practical solar cell applications is proposed. The superhydrophobic films were fabricated simply by transferring a transparent porous alumina layer, which was prepared using an anodic aluminium oxidation (AAO) technique, onto a polyethylene terephthalate (PET) film with a UV-curable polymer adhesive layer, followed by the subsequent formation of alumina nano pyramids (NPs) through the time-controlled chemical etching of the transferred porous alumina membrane (PAM). It was found experimentally that the proposed functional films can ensure the superhydrophobicity in the Cassie-Baxter wetting mode with superior water-repellent properties through a series of experimental observations including static contact angle (SCA), contact angle hysteresis (CAH), sliding behaviour on the tilted film, and dynamic behaviour of the liquid droplet impacting on the film. In addition to the superior surface wetting properties, an optical transmittance of ~79% at a light wavelength of 550 nm was achieved. Furthermore, there was no significant degradation in both the surface wetting properties and morphology even after 1500-cycles of repetitive bending tests, which indicates that the proposed superhydrophobic film is mechanically robust. Finally, the practicability of the proposed self-cleanable film was proven quantitatively by observing the changes in the power conversion efficiency (PCE) of a photovoltaic device covering the film before and after the cleaning process.

  5. Scalable fabrication of high-power graphene micro-supercapacitors for flexible and on-chip energy storage

    NASA Astrophysics Data System (ADS)

    El-Kady, Maher F.; Kaner, Richard B.

    2013-02-01

    The rapid development of miniaturized electronic devices has increased the demand for compact on-chip energy storage. Microscale supercapacitors have great potential to complement or replace batteries and electrolytic capacitors in a variety of applications. However, conventional micro-fabrication techniques have proven to be cumbersome in building cost-effective micro-devices, thus limiting their widespread application. Here we demonstrate a scalable fabrication of graphene micro-supercapacitors over large areas by direct laser writing on graphite oxide films using a standard LightScribe DVD burner. More than 100 micro-supercapacitors can be produced on a single disc in 30 min or less. The devices are built on flexible substrates for flexible electronics and on-chip uses that can be integrated with MEMS or CMOS in a single chip. Remarkably, miniaturizing the devices to the microscale results in enhanced charge-storage capacity and rate capability. These micro-supercapacitors demonstrate a power density of ~200 W cm-3, which is among the highest values achieved for any supercapacitor.

  6. A Scalable Media Multicasting Scheme

    NASA Astrophysics Data System (ADS)

    Youwei, Zhang

    IP multicast has been proved to be unfeasible for deployment, Application Layer Multicast (ALM) Based on end multicast system is practical and more scalable than IP multicast in Internet. In this paper, an ALM protocol called Scalable multicast for High Definition streaming media (SHD) is proposed in which end to end transmission capability is fully cultivated for HD media transmission without increasing much control overhead. Similar to the transmission style of BiTtorrent, hosts only forward part of data piece according to the available bandwidth that improves the usage of bandwidth greatly. On the other hand, some novel strategies are adopted to overcome the disadvantages of BiTtorrent protocol in streaming media transmission. Data transmission between hosts is implemented in many-one transmission style in Hierarchical architecture in most circumstances. Simulations implemented on Internet-like topology indicate that SHD achieves low link stress, end to end latency and stability.

  7. A Scalable Tools Communication Infrastructure

    SciTech Connect

    Buntinas, Darius; Bosilca, George; Graham, Richard L; Vallee, Geoffroy R; Watson, Gregory R.

    2008-01-01

    The Scalable Tools Communication Infrastructure (STCI) is an open source collaborative effort intended to provide high-performance, scalable, resilient, and portable communications and process control services for a wide variety of user and system tools. STCI is aimed specifically at tools for ultrascale computing and uses a component architecture to simplify tailoring the infrastructure to a wide range of scenarios. This paper describes STCI's design philosophy, the various components that will be used to provide an STCI implementation for a range of ultrascale platforms, and a range of tool types. These include tools supporting parallel run-time environments, such as MPI, parallel application correctness tools and performance analysis tools, as well as system monitoring and management tools.

  8. Sandia Scalable Encryption Software

    SciTech Connect

    Tarman, Thomas D.

    1997-08-13

    Sandia Scalable Encryption Library (SSEL) Version 1.0 is a library of functions that implement Sandia''s scalable encryption algorithm. This algorithm is used to encrypt Asynchronous Transfer Mode (ATM) data traffic, and is capable of operating on an arbitrary number of bits at a time (which permits scaling via parallel implementations), while being interoperable with differently scaled versions of this algorithm. The routines in this library implement 8 bit and 32 bit versions of a non-linear mixer which is compatible with Sandia''s hardware-based ATM encryptor.

  9. Sandia Scalable Encryption Software

    Energy Science and Technology Software Center (ESTSC)

    1997-08-13

    Sandia Scalable Encryption Library (SSEL) Version 1.0 is a library of functions that implement Sandia''s scalable encryption algorithm. This algorithm is used to encrypt Asynchronous Transfer Mode (ATM) data traffic, and is capable of operating on an arbitrary number of bits at a time (which permits scaling via parallel implementations), while being interoperable with differently scaled versions of this algorithm. The routines in this library implement 8 bit and 32 bit versions of a non-linearmore » mixer which is compatible with Sandia''s hardware-based ATM encryptor.« less

  10. High transport Jc in magnetic fields up to 28 T of stainless steel/Ag double sheathed Ba122 tapes fabricated by scalable rolling process

    NASA Astrophysics Data System (ADS)

    Gao, Zhaoshun; Togano, Kazumasa; Matsumoto, Akiyoshi; Kumakura, Hiroaki

    2015-01-01

    The recently discovered iron-based superconductors with very high upper critical field (Hc2) and small anisotropy have been regarded as a potential candidate material for high field applications. However, enhancements of superconducting properties are still needed to boost the successful use of iron-based superconductors in such applications. Here, we propose a new sheath architecture of stainless steel (SS)/Ag double sheath and investigate its influence on the microstructures and Jc-H property. We found that the transport Jc-H curves for rolled and pressed tapes both show extremely small magnetic field dependence and exceed 3 × 104 A cm-2 under 28 T, which are much higher than those of low-temperature superconductors. More interestingly, 12 cm long rolled tape shows very high homogeneity and sustains Jc as high as 7.7 × 104 A cm-2 at 10 T. These are the highest values reported so far for iron-based superconducting wires fabricated by scalable rolling process. The microstructure investigations indicate that such high Jc was achieved by higher density of the core and uniform deformation resulting better texturing. These results indicate that our process is very promising for fabricating long Ba122 wires for high field magnet, i.e. above 20 T.

  11. Complexity scalable motion-compensated temporal filtering

    NASA Astrophysics Data System (ADS)

    Clerckx, Tom; Verdicchio, Fabio; Munteanu, Adrian; Andreopoulos, Yiannis; Devos, Harald; Eeckhaut, Hendrik; Christiaens, Mark; Stroobandt, Dirk; Verkest, Diederik; Schelkens, Peter

    2004-11-01

    Computer networks and the internet have taken an important role in modern society. Together with their development, the need for digital video transmission over these networks has grown. To cope with the user demands and limitations of the network, compression of the video material has become an important issue. Additionally, many video-applications require flexibility in terms of scalability and complexity (e.g. HD/SD-TV, video-surveillance). Current ITU-T and ISO/IEC video compression standards (MPEG-x, H.26-x) lack efficient support for these types of scalability. Wavelet-based compression techniques have been proposed to tackle this problem, of which the Motion Compensated Temporal Filtering (MCTF)-based architectures couple state-of-the-art performance with full (quality, resolution, and frame-rate) scalability. However, a significant drawback of these architectures is their high complexity. The computational and memory complexity of both spatial domain (SD) MCTF and in-band (IB) MCTF video codec instantiations are examined in this study. Comparisons in terms of complexity versus performance are presented for both types of codecs. The paper indicates how complexity scalability can be achieved in such video-codecs, and analyses some of the trade-offs between complexity and coding performance. Finally, guidelines on how to implement a fully scalable video-codec that incorporates quality, temporal, resolution and complexity scalability are proposed.

  12. N- and S-doped high surface area carbon derived from soya chunks as scalable and efficient electrocatalysts for oxygen reduction

    NASA Astrophysics Data System (ADS)

    Rana, Moumita; Arora, Gunjan; Gautam, Ujjal K.

    2015-02-01

    Highly stable, cost-effective electrocatalysts facilitating oxygen reduction are crucial for the commercialization of membrane-based fuel cell and battery technologies. Herein, we demonstrate that protein-rich soya chunks with a high content of N, S and P atoms are an excellent precursor for heteroatom-doped highly graphitized carbon materials. The materials are nanoporous, with a surface area exceeding 1000 m2 g-1, and they are tunable in doping quantities. These materials exhibit highly efficient catalytic performance toward oxygen reduction reaction (ORR) with an onset potential of -0.045 V and a half-wave potential of -0.211 V (versus a saturated calomel electrode) in a basic medium, which is comparable to commercial Pt catalysts and is better than other recently developed metal-free carbon-based catalysts. These exhibit complete methanol tolerance and a performance degradation of merely ˜5% as compared to ˜14% for a commercial Pt/C catalyst after continuous use for 3000 s at the highest reduction current. We found that the fraction of graphitic N increases at a higher graphitization temperature, leading to the near complete reduction of oxygen. It is believed that due to the easy availability of the precursor and the possibility of genetic engineering to homogeneously control the heteroatom distribution, the synthetic strategy is easily scalable, with further improvement in performance.

  13. Facile and Scalable Fabrication of Highly Efficient Lead Iodide Perovskite Thin-Film Solar Cells in Air Using Gas Pump Method.

    PubMed

    Ding, Bin; Gao, Lili; Liang, Lusheng; Chu, Qianqian; Song, Xiaoxuan; Li, Yan; Yang, Guanjun; Fan, Bin; Wang, Mingkui; Li, Chengxin; Li, Changjiu

    2016-08-10

    Control of the perovskite film formation process to produce high-quality organic-inorganic metal halide perovskite thin films with uniform morphology, high surface coverage, and minimum pinholes is of great importance to highly efficient solar cells. Herein, we report on large-area light-absorbing perovskite films fabrication with a new facile and scalable gas pump method. By decreasing the total pressure in the evaporation environment, the gas pump method can significantly enhance the solvent evaporation rate by 8 times faster and thereby produce an extremely dense, uniform, and full-coverage perovskite thin film. The resulting planar perovskite solar cells can achieve an impressive power conversion efficiency up to 19.00% with an average efficiency of 17.38 ± 0.70% for 32 devices with an area of 5 × 2 mm, 13.91% for devices with a large area up to 1.13 cm(2). The perovskite films can be easily fabricated in air conditions with a relative humidity of 45-55%, which definitely has a promising prospect in industrial application of large-area perovskite solar panels. PMID:27428311

  14. Cost-effective scalable synthesis of mesoporous germanium particles via a redox-transmetalation reaction for high-performance energy storage devices.

    PubMed

    Choi, Sinho; Kim, Jieun; Choi, Nam-Soon; Kim, Min Gyu; Park, Soojin

    2015-02-24

    Nanostructured germanium is a promising material for high-performance energy storage devices. However, synthesizing it in a cost-effective and simple manner on a large scale remains a significant challenge. Herein, we report a redox-transmetalation reaction-based route for the large-scale synthesis of mesoporous germanium particles from germanium oxide at temperatures of 420-600 °C. We could confirm that a unique redox-transmetalation reaction occurs between Zn(0) and Ge(4+) at approximately 420 °C using temperature-dependent in situ X-ray absorption fine structure analysis. This reaction has several advantages, which include (i) the successful synthesis of germanium particles at a low temperature (∼450 °C), (ii) the accommodation of large volume changes, owing to the mesoporous structure of the germanium particles, and (iii) the ability to synthesize the particles in a cost-effective and scalable manner, as inexpensive metal oxides are used as the starting materials. The optimized mesoporous germanium anode exhibits a reversible capacity of ∼1400 mA h g(-1) after 300 cycles at a rate of 0.5 C (corresponding to the capacity retention of 99.5%), as well as stable cycling in a full cell containing a LiCoO2 cathode with a high energy density (charge capacity = 286.62 mA h cm(-3)). PMID:25666187

  15. Scalable filter banks

    NASA Astrophysics Data System (ADS)

    Hur, Youngmi; Okoudjou, Kasso A.

    2015-08-01

    A finite frame is said to be scalable if its vectors can be rescaled so that the resulting set of vectors is a tight frame. The theory of scalable frame has been extended to the setting of Laplacian pyramids which are based on (rectangular) paraunitary matrices whose column vectors are Laurent polynomial vectors. This is equivalent to scaling the polyphase matrices of the associated filter banks. Consequently, tight wavelet frames can be constructed by appropriately scaling the columns of these paraunitary matrices by diagonal matrices whose diagonal entries are square magnitude of Laurent polynomials. In this paper we present examples of tight wavelet frames constructed in this manner and discuss some of their properties in comparison to the (non tight) wavelet frames they arise from.

  16. Rad-Hard, Miniaturized, Scalable, High-Voltage Switching Module for Power Applications Rad-Hard, Miniaturized

    NASA Technical Reports Server (NTRS)

    Adell, Philippe C.; Mojarradi, Mohammad; DelCastillo, Linda Y.; Vo, Tuan A.

    2011-01-01

    A paper discusses the successful development of a miniaturized radiation hardened high-voltage switching module operating at 2.5 kV suitable for space application. The high-voltage architecture was designed, fabricated, and tested using a commercial process that uses a unique combination of 0.25 micrometer CMOS (complementary metal oxide semiconductor) transistors and high-voltage lateral DMOS (diffusion metal oxide semiconductor) device with high breakdown voltage (greater than 650 V). The high-voltage requirements are achieved by stacking a number of DMOS devices within one module, while two modules can be placed in series to achieve higher voltages. Besides the high-voltage requirements, a second generation prototype is currently being developed to provide improved switching capabilities (rise time and fall time for full range of target voltages and currents), the ability to scale the output voltage to a desired value with good accuracy (few percent) up to 10 kV, to cover a wide range of high-voltage applications. In addition, to ensure miniaturization, long life, and high reliability, the assemblies will require intensive high-voltage electrostatic modeling (optimized E-field distribution throughout the module) to complete the proposed packaging approach and test the applicability of using advanced materials in a space-like environment (temperature and pressure) to help prevent potential arcing and corona due to high field regions. Finally, a single-event effect evaluation would have to be performed and single-event mitigation methods implemented at the design and system level or developed to ensure complete radiation hardness of the module.

  17. Enhancing Scalability of Sparse Direct Methods

    SciTech Connect

    Li, Xiaoye S.; Demmel, James; Grigori, Laura; Gu, Ming; Xia,Jianlin; Jardin, Steve; Sovinec, Carl; Lee, Lie-Quan

    2007-07-23

    TOPS is providing high-performance, scalable sparse direct solvers, which have had significant impacts on the SciDAC applications, including fusion simulation (CEMM), accelerator modeling (COMPASS), as well as many other mission-critical applications in DOE and elsewhere. Our recent developments have been focusing on new techniques to overcome scalability bottleneck of direct methods, in both time and memory. These include parallelizing symbolic analysis phase and developing linear-complexity sparse factorization methods. The new techniques will make sparse direct methods more widely usable in large 3D simulations on highly-parallel petascale computers.

  18. SWIFT-scalable clustering for automated identification of rare cell populations in large, high-dimensional flow cytometry datasets, part 1: algorithm design.

    PubMed

    Naim, Iftekhar; Datta, Suprakash; Rebhahn, Jonathan; Cavenaugh, James S; Mosmann, Tim R; Sharma, Gaurav

    2014-05-01

    We present a model-based clustering method, SWIFT (Scalable Weighted Iterative Flow-clustering Technique), for digesting high-dimensional large-sized datasets obtained via modern flow cytometry into more compact representations that are well-suited for further automated or manual analysis. Key attributes of the method include the following: (a) the analysis is conducted in the multidimensional space retaining the semantics of the data, (b) an iterative weighted sampling procedure is utilized to maintain modest computational complexity and to retain discrimination of extremely small subpopulations (hundreds of cells from datasets containing tens of millions), and (c) a splitting and merging procedure is incorporated in the algorithm to preserve distinguishability between biologically distinct populations, while still providing a significant compaction relative to the original data. This article presents a detailed algorithmic description of SWIFT, outlining the application-driven motivations for the different design choices, a discussion of computational complexity of the different steps, and results obtained with SWIFT for synthetic data and relatively simple experimental data that allow validation of the desirable attributes. A companion paper (Part 2) highlights the use of SWIFT, in combination with additional computational tools, for more challenging biological problems. PMID:24677621

  19. NWChem: scalable parallel computational chemistry

    SciTech Connect

    van Dam, Hubertus JJ; De Jong, Wibe A.; Bylaska, Eric J.; Govind, Niranjan; Kowalski, Karol; Straatsma, TP; Valiev, Marat

    2011-11-01

    NWChem is a general purpose computational chemistry code specifically designed to run on distributed memory parallel computers. The core functionality of the code focuses on molecular dynamics, Hartree-Fock and density functional theory methods for both plane-wave basis sets as well as Gaussian basis sets, tensor contraction engine based coupled cluster capabilities and combined quantum mechanics/molecular mechanics descriptions. It was realized from the beginning that scalable implementations of these methods required a programming paradigm inherently different from what message passing approaches could offer. In response a global address space library, the Global Array Toolkit, was developed. The programming model it offers is based on using predominantly one-sided communication. This model underpins most of the functionality in NWChem and the power of it is exemplified by the fact that the code scales to tens of thousands of processors. In this paper the core capabilities of NWChem are described as well as their implementation to achieve an efficient computational chemistry code with high parallel scalability. NWChem is a modern, open source, computational chemistry code1 specifically designed for large scale parallel applications2. To meet the challenges of developing efficient, scalable and portable programs of this nature a particular code design was adopted. This code design involved two main features. First of all, the code is build up in a modular fashion so that a large variety of functionality can be integrated easily. Secondly, to facilitate writing complex parallel algorithms the Global Array toolkit was developed. This toolkit allows one to write parallel applications in a shared memory like approach, but offers additional mechanisms to exploit data locality to lower communication overheads. This framework has proven to be very successful in computational chemistry but is applicable to any engineering domain. Within the context created by the features

  20. Sustainable and scalable production of monodisperse and highly uniform colloidal carbonaceous spheres using sodium polyacrylate as the dispersant.

    PubMed

    Gong, Yutong; Xie, Lei; Li, Haoran; Wang, Yong

    2014-10-28

    Monodisperse, uniform colloidal carbonaceous spheres were fabricated by the hydrothermal treatment of glucose with the help of a tiny amount of sodium polyacrylate (PAANa). This synthetic strategy is effective at high glucose concentration and for scale-up experiments. The sphere size can be easily tuned by the reaction time, temperature and glucose concentration. PMID:25199065

  1. Performance evaluation and capacity planning for a scalable and highly available virtualisation infrastructure for the LHCb experiment

    NASA Astrophysics Data System (ADS)

    Bonaccorsi, E.; Neufeld, N.; Sborzacchi, F.

    2014-06-01

    The virtual computing is often run to satisfy different needs: reduce costs, reduce resources, simplify maintenance and the last but not the least adds flexibility. The use of Virtualization in a complex system such as a farm of PCs that control the hardware of an experiment (PLC, power supplies, gas, magnets...) put us in a condition where not only an High Performance requirements need to be carefully considered but also a deep analysis of strategies to achieve a certain level of High Availability. We conducted a performance evaluation on different and comparable storage/network/virtualization platforms. The performance is measured using a series of independent benchmarks, testing the speed and the stability of multiple VMs running heavy-load operations on the I/O of virtualized storage and the virtualized network. The result from the benchmark tests allowed us to study and evaluate how the different VMs workloads interact with the Hardware/Software resource layers.

  2. Scalable production of highly sensitive nanosensors based on graphene functionalized with a designed G protein-coupled receptor.

    PubMed

    Lerner, Mitchell B; Matsunaga, Felipe; Han, Gang Hee; Hong, Sung Ju; Xi, Jin; Crook, Alexander; Perez-Aguilar, Jose Manuel; Park, Yung Woo; Saven, Jeffery G; Liu, Renyu; Johnson, A T Charlie

    2014-05-14

    We have developed a novel, all-electronic biosensor for opioids that consists of an engineered μ-opioid receptor protein, with high binding affinity for opioids, chemically bonded to a graphene field-effect transistor to read out ligand binding. A variant of the receptor protein that provided chemical recognition was computationally redesigned to enhance its solubility and stability in an aqueous environment. A shadow mask process was developed to fabricate arrays of hundreds of graphene transistors with average mobility of ∼1500 cm(2) V(-1) s(-1) and yield exceeding 98%. The biosensor exhibits high sensitivity and selectivity for the target naltrexone, an opioid receptor antagonist, with a detection limit of 10 pg/mL. PMID:24742304

  3. Scalable Production of Highly Sensitive Nanosensors Based on Graphene Functionalized with a Designed G Protein-Coupled Receptor

    PubMed Central

    2015-01-01

    We have developed a novel, all-electronic biosensor for opioids that consists of an engineered μ-opioid receptor protein, with high binding affinity for opioids, chemically bonded to a graphene field-effect transistor to read out ligand binding. A variant of the receptor protein that provided chemical recognition was computationally redesigned to enhance its solubility and stability in an aqueous environment. A shadow mask process was developed to fabricate arrays of hundreds of graphene transistors with average mobility of ∼1500 cm2 V–1 s–1 and yield exceeding 98%. The biosensor exhibits high sensitivity and selectivity for the target naltrexone, an opioid receptor antagonist, with a detection limit of 10 pg/mL. PMID:24742304

  4. Context-adaptive binary arithmetic coding with precise probability estimation and complexity scalability for high-efficiency video coding

    NASA Astrophysics Data System (ADS)

    Karwowski, Damian; Domański, Marek

    2016-01-01

    An improved context-based adaptive binary arithmetic coding (CABAC) is presented. The idea for the improvement is to use a more accurate mechanism for estimation of symbol probabilities in the standard CABAC algorithm. The authors' proposal of such a mechanism is based on the context-tree weighting technique. In the framework of a high-efficiency video coding (HEVC) video encoder, the improved CABAC allows 0.7% to 4.5% bitrate saving compared to the original CABAC algorithm. The application of the proposed algorithm marginally affects the complexity of HEVC video encoder, but the complexity of video decoder increases by 32% to 38%. In order to decrease the complexity of video decoding, a new tool has been proposed for the improved CABAC that enables scaling of the decoder complexity. Experiments show that this tool gives 5% to 7.5% reduction of the decoding time while still maintaining high efficiency in the data compression.

  5. Scalable Production of Highly Sensitive Nanosensors Based on Graphene Functionalized with a Designed G Protein-Coupled Receptor

    NASA Astrophysics Data System (ADS)

    Lerner, Mitchell B.; Matsunaga, Felipe; Han, Gang Hee; Hong, Sung Ju; Xi, Jin; Crook, Alexander; Perez-Aguilar, Jose Manuel; Park, Yung Woo; Saven, Jeffery G.; Liu, Renyu; Johnson, A. T. Charlie

    2014-05-01

    We have developed a novel, all-electronic biosensor for opioids that consists of an engineered mu opioid receptor protein, with high binding affinity for opioids, chemically bonded to a graphene field-effect transistor to read out ligand binding. A variant of the receptor protein that provided chemical recognition was computationally redesigned to enhance its solubility and stability in an aqueous environment. A shadow mask process was developed to fabricate arrays of hundreds of graphene transistors with average mobility of ~1500 cm2 V-1 s-1 and yield exceeding 98%. The biosensor exhibits high sensitivity and selectivity for the target naltrexone, an opioid receptor antagonist, with a detection limit of 10 pg/mL.

  6. Facile and Scalable Preparation of Graphene Oxide-Based Magnetic Hybrids for Fast and Highly Efficient Removal of Organic Dyes

    PubMed Central

    Jiao, Tifeng; Liu, Yazhou; Wu, Yitian; Zhang, Qingrui; Yan, Xuehai; Gao, Faming; Bauer, Adam J. P.; Liu, Jianzhao; Zeng, Tingying; Li, Bingbing

    2015-01-01

    This study reports the facile preparation and the dye removal efficiency of nanohybrids composed of graphene oxide (GO) and Fe3O4 nanoparticles with various geometrical structures. In comparison to previously reported GO/Fe3O4 composites prepared through the one-pot, in situ deposition of Fe3O4 nanoparticles, the GO/Fe3O4 nanohybrids reported here were obtained by taking advantage of the physical affinities between sulfonated GO and Fe3O4 nanoparticles, which allows tuning the dimensions and geometries of Fe3O4 nanoparticles in order to decrease their contact area with GO, while still maintaining the magnetic properties of the nanohybrids for easy separation and adsorbent recycling. Both the as-prepared and regenerated nanohybrids demonstrate a nearly 100% removal rate for methylene blue and an impressively high removal rate for Rhodamine B. This study provides new insights into the facile and controllable industrial scale fabrication of safe and highly efficient GO-based adsorbents for dye or other organic pollutants in a wide range of environmental-related applications. PMID:26220847

  7. Facile and Scalable Preparation of Graphene Oxide-Based Magnetic Hybrids for Fast and Highly Efficient Removal of Organic Dyes

    NASA Astrophysics Data System (ADS)

    Jiao, Tifeng; Liu, Yazhou; Wu, Yitian; Zhang, Qingrui; Yan, Xuehai; Gao, Faming; Bauer, Adam J. P.; Liu, Jianzhao; Zeng, Tingying; Li, Bingbing

    2015-07-01

    This study reports the facile preparation and the dye removal efficiency of nanohybrids composed of graphene oxide (GO) and Fe3O4 nanoparticles with various geometrical structures. In comparison to previously reported GO/Fe3O4 composites prepared through the one-pot, in situ deposition of Fe3O4 nanoparticles, the GO/Fe3O4 nanohybrids reported here were obtained by taking advantage of the physical affinities between sulfonated GO and Fe3O4 nanoparticles, which allows tuning the dimensions and geometries of Fe3O4 nanoparticles in order to decrease their contact area with GO, while still maintaining the magnetic properties of the nanohybrids for easy separation and adsorbent recycling. Both the as-prepared and regenerated nanohybrids demonstrate a nearly 100% removal rate for methylene blue and an impressively high removal rate for Rhodamine B. This study provides new insights into the facile and controllable industrial scale fabrication of safe and highly efficient GO-based adsorbents for dye or other organic pollutants in a wide range of environmental-related applications.

  8. Controlled scalable synthesis of uniform, high-quality monolayer and few-layer MoS2 films.

    PubMed

    Yu, Yifei; Li, Chun; Liu, Yi; Su, Liqin; Zhang, Yong; Cao, Linyou

    2013-01-01

    Two dimensional (2D) materials with a monolayer of atoms represent an ultimate control of material dimension in the vertical direction. Molybdenum sulfide (MoS2) monolayers, with a direct bandgap of 1.8 eV, offer an unprecedented prospect of miniaturizing semiconductor science and technology down to a truly atomic scale. Recent studies have indeed demonstrated the promise of 2D MoS2 in fields including field effect transistors, low power switches, optoelectronics, and spintronics. However, device development with 2D MoS2 has been delayed by the lack of capabilities to produce large-area, uniform, and high-quality MoS2 monolayers. Here we present a self-limiting approach that can grow high quality monolayer and few-layer MoS2 films over an area of centimeters with unprecedented uniformity and controllability. This approach is compatible with the standard fabrication process in semiconductor industry. It paves the way for the development of practical devices with 2D MoS2 and opens up new avenues for fundamental research. PMID:23689610

  9. The Scalable Checkpoint/Restart Library

    Energy Science and Technology Software Center (ESTSC)

    2009-02-23

    The Scalable Checkpoint/Restart (SCR) library provides an interface that codes may use to worite our and read in application-level checkpoints in a scalable fashion. In the current implementation, checkpoint files are cached in local storage (hard disk or RAM disk) on the compute nodes. This technique provides scalable aggregate bandwidth and uses storage resources that are fully dedicated to the job. This approach addresses the two common drawbacks of checkpointing a large-scale application to amore » shared parallel file system, namely, limited bandwidth and file system contention. In fact, on current platforms, SCR scales linearly with the number of compute nodes. It has been benchmarked as high as 720GB/s on 1094 nodes of Atlas, which is nearly two orders of magnitude faster thanthe parallel file system.« less

  10. The Scalable Checkpoint/Restart Library

    SciTech Connect

    Moody, A.

    2009-02-23

    The Scalable Checkpoint/Restart (SCR) library provides an interface that codes may use to worite our and read in application-level checkpoints in a scalable fashion. In the current implementation, checkpoint files are cached in local storage (hard disk or RAM disk) on the compute nodes. This technique provides scalable aggregate bandwidth and uses storage resources that are fully dedicated to the job. This approach addresses the two common drawbacks of checkpointing a large-scale application to a shared parallel file system, namely, limited bandwidth and file system contention. In fact, on current platforms, SCR scales linearly with the number of compute nodes. It has been benchmarked as high as 720GB/s on 1094 nodes of Atlas, which is nearly two orders of magnitude faster thanthe parallel file system.

  11. Scalable rendering on PC clusters

    SciTech Connect

    WYLIE,BRIAN N.; LEWIS,VASILY; SHIRLEY,DAVID NOYES; PAVLAKOS,CONSTANTINE

    2000-04-25

    This case study presents initial results from research targeted at the development of cost-effective scalable visualization and rendering technologies. The implementations of two 3D graphics libraries based on the popular sort-last and sort-middle parallel rendering techniques are discussed. An important goal of these implementations is to provide scalable rendering capability for extremely large datasets (>> 5 million polygons). Applications can use these libraries for either run-time visualization, by linking to an existing parallel simulation, or for traditional post-processing by linking to an interactive display program. The use of parallel, hardware-accelerated rendering on commodity hardware is leveraged to achieve high performance. Current performance results show that, using current hardware (a small 16-node cluster), they can utilize up to 85% of the aggregate graphics performance and achieve rendering rates in excess of 20 million polygons/second using OpenGL{reg_sign} with lighting, Gouraud shading, and individually specified triangles (not t-stripped).

  12. A Scalable Analysis Toolkit

    NASA Technical Reports Server (NTRS)

    Aiken, Alexander

    2001-01-01

    The Scalable Analysis Toolkit (SAT) project aimed to demonstrate that it is feasible and useful to statically detect software bugs in very large systems. The technical focus of the project was on a relatively new class of constraint-based techniques for analysis software, where the desired facts about programs (e.g., the presence of a particular bug) are phrased as constraint problems to be solved. At the beginning of this project, the most successful forms of formal software analysis were limited forms of automatic theorem proving (as exemplified by the analyses used in language type systems and optimizing compilers), semi-automatic theorem proving for full verification, and model checking. With a few notable exceptions these approaches had not been demonstrated to scale to software systems of even 50,000 lines of code. Realistic approaches to large-scale software analysis cannot hope to make every conceivable formal method scale. Thus, the SAT approach is to mix different methods in one application by using coarse and fast but still adequate methods at the largest scales, and reserving the use of more precise but also more expensive methods at smaller scales for critical aspects (that is, aspects critical to the analysis problem under consideration) of a software system. The principled method proposed for combining a heterogeneous collection of formal systems with different scalability characteristics is mixed constraints. This idea had been used previously in small-scale applications with encouraging results: using mostly coarse methods and narrowly targeted precise methods, useful information (meaning the discovery of bugs in real programs) was obtained with excellent scalability.

  13. Depth-specific optogenetic control in vivo with a scalable, high-density μLED neural probe.

    PubMed

    Scharf, Robert; Tsunematsu, Tomomi; McAlinden, Niall; Dawson, Martin D; Sakata, Shuzo; Mathieson, Keith

    2016-01-01

    Controlling neural circuits is a powerful approach to uncover a causal link between neural activity and behaviour. Optogenetics has been widely adopted by the neuroscience community as it offers cell-type-specific perturbation with millisecond precision. However, these studies require light delivery in complex patterns with cellular-scale resolution, while covering a large volume of tissue at depth in vivo. Here we describe a novel high-density silicon-based microscale light-emitting diode (μLED) array, consisting of up to ninety-six 25 μm-diameter μLEDs emitting at a wavelength of 450 nm with a peak irradiance of 400 mW/mm(2). A width of 100 μm, tapering to a 1 μm point, and a 40 μm thickness help minimise tissue damage during insertion. Thermal properties permit a set of optogenetic operating regimes, with ~0.5 °C average temperature increase. We demonstrate depth-dependent activation of mouse neocortical neurons in vivo, offering an inexpensive novel tool for the precise manipulation of neural activity. PMID:27334849

  14. Designing a Scalable Fault Tolerance Model for High Performance Computational Chemistry: A Case Study with Coupled Cluster Perturbative Triples

    SciTech Connect

    van Dam, Hubertus JJ; Vishnu, Abhinav; De Jong, Wibe A.

    2011-01-11

    In the last couple of decades, the massive computational power provided by the most modern supercomputers has resulted in simulation of higher order computational chem- istry methods, previously considered intractable. As the system sizes continue to increase, computational chemistry domain continues to escalate this trend using parallel comput- ing with programming models such as Message Passing Interface (MPI) and Partitioned Global Address Space (PGAS) programming models such as Global Arrays. The ever increasing scale of these supercomputers comes at a cost of reduced mean time between failures, currently in the order of days, and projected to be in the order of hours for up- coming extreme scale systems. While traditional disk based checkpointing methods are ubiquitous for storing intermediate solutions, they suffer from high overhead of writing and recovering from checkpoints. In practice, checkpointing itself often brings the system down. Clearly, methods beyond checkpointing are imperative to handling the aggravating issue of reducing MTBF. In this paper, we address this challenge by designing and im- plementing an efficient fault tolerant version of coupled cluster method with NWChem, using in memory data redundancy. We present the challenges associated with our de- sign including efficient data storage model, maintenance of at least one consistent data copy and the recovery process. Our performance evaluation without faults shows that the current design exhibits negligible overhead. In the presence of a fault, the proposed design incurs negligible overhead in comparison to the state of the art implementation without faults.

  15. Depth-specific optogenetic control in vivo with a scalable, high-density μLED neural probe

    PubMed Central

    Scharf, Robert; Tsunematsu, Tomomi; McAlinden, Niall; Dawson, Martin D.; Sakata, Shuzo; Mathieson, Keith

    2016-01-01

    Controlling neural circuits is a powerful approach to uncover a causal link between neural activity and behaviour. Optogenetics has been widely adopted by the neuroscience community as it offers cell-type-specific perturbation with millisecond precision. However, these studies require light delivery in complex patterns with cellular-scale resolution, while covering a large volume of tissue at depth in vivo. Here we describe a novel high-density silicon-based microscale light-emitting diode (μLED) array, consisting of up to ninety-six 25 μm-diameter μLEDs emitting at a wavelength of 450 nm with a peak irradiance of 400 mW/mm2. A width of 100 μm, tapering to a 1 μm point, and a 40 μm thickness help minimise tissue damage during insertion. Thermal properties permit a set of optogenetic operating regimes, with ~0.5 °C average temperature increase. We demonstrate depth-dependent activation of mouse neocortical neurons in vivo, offering an inexpensive novel tool for the precise manipulation of neural activity. PMID:27334849

  16. Designing a Scalable Fault Tolerance Model for High Performance Computational Chemistry: A Case Study with Coupled Cluster Perturbative Triples.

    PubMed

    van Dam, Hubertus J J; Vishnu, Abhinav; de Jong, Wibe A

    2011-01-11

    In the past couple of decades, the massive computational power provided by the most modern supercomputers has resulted in simulation of higher-order computational chemistry methods, previously considered intractable. As the system sizes continue to increase, the computational chemistry domain continues to escalate this trend using parallel computing with programming models such as Message Passing Interface (MPI) and Partitioned Global Address Space (PGAS) programming models such as Global Arrays. The ever increasing scale of these supercomputers comes at a cost of reduced Mean Time Between Failures (MTBF), currently on the order of days and projected to be on the order of hours for upcoming extreme scale systems. While traditional disk-based check pointing methods are ubiquitous for storing intermediate solutions, they suffer from high overhead of writing and recovering from checkpoints. In practice, checkpointing itself often brings the system down. Clearly, methods beyond checkpointing are imperative to handling the aggravating issue of reducing MTBF. In this paper, we address this challenge by designing and implementing an efficient fault tolerant version of the Coupled Cluster (CC) method with NWChem, using in-memory data redundancy. We present the challenges associated with our design, including an efficient data storage model, maintenance of at least one consistent data copy, and the recovery process. Our performance evaluation without faults shows that the current design exhibits a small overhead. In the presence of a simulated fault, the proposed design incurs negligible overhead in comparison to the state of the art implementation without faults. PMID:26606219

  17. Scalable integration of Li5FeO4 towards robust, high-performance lithium-ion hybrid capacitors.

    PubMed

    Park, Min-Sik; Lim, Young-Geun; Hwang, Soo Min; Kim, Jung Ho; Kim, Jeom-Soo; Dou, Shi Xue; Cho, Jaephil; Kim, Young-Jun

    2014-11-01

    Lithium-ion hybrid capacitors have attracted great interest due to their high specific energy relative to conventional electrical double-layer capacitors. Nevertheless, the safety issue still remains a drawback for lithium-ion capacitors in practical operational environments because of the use of metallic lithium. Herein, single-phase Li5FeO4 with an antifluorite structure that acts as an alternative lithium source (instead of metallic lithium) is employed and its potential use for lithium-ion capacitors is verified. Abundant Li(+) amounts can be extracted from Li5FeO4 incorporated in the positive electrode and efficiently doped into the negative electrode during the first electrochemical charging. After the first Li(+) extraction, Li(+) does not return to the Li5FeO4 host structure and is steadily involved in the electrochemical reactions of the negative electrode during subsequent cycling. Various electrochemical and structural analyses support its superior characteristics for use as a promising lithium source. This versatile approach can yield a sufficient Li(+)-doping efficiency of >90% and improved safety as a result of the removal of metallic lithium from the cell. PMID:25208971

  18. Depth-specific optogenetic control in vivo with a scalable, high-density μLED neural probe

    NASA Astrophysics Data System (ADS)

    Scharf, Robert; Tsunematsu, Tomomi; McAlinden, Niall; Dawson, Martin D.; Sakata, Shuzo; Mathieson, Keith

    2016-06-01

    Controlling neural circuits is a powerful approach to uncover a causal link between neural activity and behaviour. Optogenetics has been widely adopted by the neuroscience community as it offers cell-type-specific perturbation with millisecond precision. However, these studies require light delivery in complex patterns with cellular-scale resolution, while covering a large volume of tissue at depth in vivo. Here we describe a novel high-density silicon-based microscale light-emitting diode (μLED) array, consisting of up to ninety-six 25 μm-diameter μLEDs emitting at a wavelength of 450 nm with a peak irradiance of 400 mW/mm2. A width of 100 μm, tapering to a 1 μm point, and a 40 μm thickness help minimise tissue damage during insertion. Thermal properties permit a set of optogenetic operating regimes, with ~0.5 °C average temperature increase. We demonstrate depth-dependent activation of mouse neocortical neurons in vivo, offering an inexpensive novel tool for the precise manipulation of neural activity.

  19. Scalable optical quantum computer

    SciTech Connect

    Manykin, E A; Mel'nichenko, E V

    2014-12-31

    A way of designing a scalable optical quantum computer based on the photon echo effect is proposed. Individual rare earth ions Pr{sup 3+}, regularly located in the lattice of the orthosilicate (Y{sub 2}SiO{sub 5}) crystal, are suggested to be used as optical qubits. Operations with qubits are performed using coherent and incoherent laser pulses. The operation protocol includes both the method of measurement-based quantum computations and the technique of optical computations. Modern hybrid photon echo protocols, which provide a sufficient quantum efficiency when reading recorded states, are considered as most promising for quantum computations and communications. (quantum computer)

  20. High-performance flat data center network architecture based on scalable and flow-controlled optical switching system

    NASA Astrophysics Data System (ADS)

    Calabretta, Nicola; Miao, Wang; Dorren, Harm

    2016-03-01

    Traffic in data centers networks (DCNs) is steadily growing to support various applications and virtualization technologies. Multi-tenancy enabling efficient resource utilization is considered as a key requirement for the next generation DCs resulting from the growing demands for services and applications. Virtualization mechanisms and technologies can leverage statistical multiplexing and fast switch reconfiguration to further extend the DC efficiency and agility. We present a novel high performance flat DCN employing bufferless and distributed fast (sub-microsecond) optical switches with wavelength, space, and time switching operation. The fast optical switches can enhance the performance of the DCNs by providing large-capacity switching capability and efficiently sharing the data plane resources by exploiting statistical multiplexing. Benefiting from the Software-Defined Networking (SDN) control of the optical switches, virtual DCNs can be flexibly created and reconfigured by the DCN provider. Numerical and experimental investigations of the DCN based on the fast optical switches show the successful setup of virtual network slices for intra-data center interconnections. Experimental results to assess the DCN performance in terms of latency and packet loss show less than 10^-5 packet loss and 640ns end-to-end latency with 0.4 load and 16- packet size buffer. Numerical investigation on the performance of the systems when the port number of the optical switch is scaled to 32x32 system indicate that more than 1000 ToRs each with Terabit/s interface can be interconnected providing a Petabit/s capacity. The roadmap to photonic integration of large port optical switches will be also presented.

  1. Crickets Are Not a Free Lunch: Protein Capture from Scalable Organic Side-Streams via High-Density Populations of Acheta domesticus

    PubMed Central

    Lundy, Mark E.; Parrella, Michael P.

    2015-01-01

    It has been suggested that the ecological impact of crickets as a source of dietary protein is less than conventional forms of livestock due to their comparatively efficient feed conversion and ability to consume organic side-streams. This study measured the biomass output and feed conversion ratios of house crickets (Acheta domesticus) reared on diets that varied in quality, ranging from grain-based to highly cellulosic diets. The measurements were made at a much greater population scale and density than any previously reported in the scientific literature. The biomass accumulation was strongly influenced by the quality of the diet (p<0.001), with the nitrogen (N) content, the ratio of N to acid detergent fiber (ADF) content, and the crude fat (CF) content (y=N/ADF+CF) explaining most of the variability between feed treatments (p = 0.02; R2 = 0.96). In addition, for populations of crickets that were able to survive to a harvestable size, the feed conversion ratios measured were higher (less efficient) than those reported from studies conducted at smaller scales and lower population densities. Compared to the industrial-scale production of chickens, crickets fed a poultry feed diet showed little improvement in protein conversion efficiency, a key metric in determining the ecological footprint of grain-based livestock protein. Crickets fed the solid filtrate from food waste processed at an industrial scale via enzymatic digestion were able to reach a harvestable size and achieve feed and protein efficiencies similar to that of chickens. However, crickets fed minimally-processed, municipal-scale food waste and diets composed largely of straw experienced >99% mortality without reaching a harvestable size. Therefore, the potential for A. domesticus to sustainably supplement the global protein supply, beyond what is currently produced via grain-fed chickens, will depend on capturing regionally scalable organic side-streams of relatively high-quality that are not

  2. Crickets are not a free lunch: protein capture from scalable organic side-streams via high-density populations of Acheta domesticus.

    PubMed

    Lundy, Mark E; Parrella, Michael P

    2015-01-01

    It has been suggested that the ecological impact of crickets as a source of dietary protein is less than conventional forms of livestock due to their comparatively efficient feed conversion and ability to consume organic side-streams. This study measured the biomass output and feed conversion ratios of house crickets (Acheta domesticus) reared on diets that varied in quality, ranging from grain-based to highly cellulosic diets. The measurements were made at a much greater population scale and density than any previously reported in the scientific literature. The biomass accumulation was strongly influenced by the quality of the diet (p<0.001), with the nitrogen (N) content, the ratio of N to acid detergent fiber (ADF) content, and the crude fat (CF) content (y=N/ADF+CF) explaining most of the variability between feed treatments (p = 0.02; R2 = 0.96). In addition, for populations of crickets that were able to survive to a harvestable size, the feed conversion ratios measured were higher (less efficient) than those reported from studies conducted at smaller scales and lower population densities. Compared to the industrial-scale production of chickens, crickets fed a poultry feed diet showed little improvement in protein conversion efficiency, a key metric in determining the ecological footprint of grain-based livestock protein. Crickets fed the solid filtrate from food waste processed at an industrial scale via enzymatic digestion were able to reach a harvestable size and achieve feed and protein efficiencies similar to that of chickens. However, crickets fed minimally-processed, municipal-scale food waste and diets composed largely of straw experienced >99% mortality without reaching a harvestable size. Therefore, the potential for A. domesticus to sustainably supplement the global protein supply, beyond what is currently produced via grain-fed chickens, will depend on capturing regionally scalable organic side-streams of relatively high-quality that are not

  3. Synthesis of Pt-Ni Octahedra in Continuous-Flow Droplet Reactors for the Scalable Production of Highly Active Catalysts toward Oxygen Reduction.

    PubMed

    Niu, Guangda; Zhou, Ming; Yang, Xuan; Park, Jinho; Lu, Ning; Wang, Jinguo; Kim, Moon J; Wang, Liduo; Xia, Younan

    2016-06-01

    A number of groups have reported the syntheses of nanosized Pt-Ni octahedra with remarkable activities toward the oxygen reduction reaction (ORR), a process key to the operation of proton-exchange membrane fuel cells. However, the throughputs of those batch-based syntheses are typically limited to a scale of 5-25 mg Pt per batch, which is far below the amount needed for commercial evaluation. Here we report the use of droplet reactors for the continuous and scalable production of Pt-Ni octahedra with high activities toward ORR. In a typical synthesis, Pt(acac)2, Ni(acac)2, and W(CO)6 were dissolved in a mixture of oleylamine, oleic acid, and benzyl ether, and then pumped into a polytetrafluoroethylene tube. When the solution entered the reaction zone at a temperature held in the range of 170-230 °C, W(CO)6 quickly decomposed to generate CO gas, naturally separating the reaction solution into discrete, uniform droplets. Each droplet then served as a reactor for the nucleation and growth of Pt-Ni octahedra whose size and composition could be controlled by changing the composition of the solvent and/or adjusting the amount of Ni(acac)2 added into the reaction solution. For a catalyst based on Pt2.4Ni octahedra of 9 nm in edge length, it showed an ORR mass activity of 2.67 A mgPt(-1) at 0.9 V, representing an 11-fold improvement over a state-of-the-art commercial Pt/C catalyst (0.24 A mgPt(-1)). PMID:27135156

  4. Medusa: a scalable MR console using USB.

    PubMed

    Stang, Pascal P; Conolly, Steven M; Santos, Juan M; Pauly, John M; Scott, Greig C

    2012-02-01

    Magnetic resonance imaging (MRI) pulse sequence consoles typically employ closed proprietary hardware, software, and interfaces, making difficult any adaptation for innovative experimental technology. Yet MRI systems research is trending to higher channel count receivers, transmitters, gradient/shims, and unique interfaces for interventional applications. Customized console designs are now feasible for researchers with modern electronic components, but high data rates, synchronization, scalability, and cost present important challenges. Implementing large multichannel MR systems with efficiency and flexibility requires a scalable modular architecture. With Medusa, we propose an open system architecture using the universal serial bus (USB) for scalability, combined with distributed processing and buffering to address the high data rates and strict synchronization required by multichannel MRI. Medusa uses a modular design concept based on digital synthesizer, receiver, and gradient blocks, in conjunction with fast programmable logic for sampling and synchronization. Medusa is a form of synthetic instrument, being reconfigurable for a variety of medical/scientific instrumentation needs. The Medusa distributed architecture, scalability, and data bandwidth limits are presented, and its flexibility is demonstrated in a variety of novel MRI applications. PMID:21954200

  5. Medusa: A Scalable MR Console Using USB

    PubMed Central

    Stang, Pascal P.; Conolly, Steven M.; Santos, Juan M.; Pauly, John M.; Scott, Greig C.

    2012-01-01

    MRI pulse sequence consoles typically employ closed proprietary hardware, software, and interfaces, making difficult any adaptation for innovative experimental technology. Yet MRI systems research is trending to higher channel count receivers, transmitters, gradient/shims, and unique interfaces for interventional applications. Customized console designs are now feasible for researchers with modern electronic components, but high data rates, synchronization, scalability, and cost present important challenges. Implementing large multi-channel MR systems with efficiency and flexibility requires a scalable modular architecture. With Medusa, we propose an open system architecture using the Universal Serial Bus (USB) for scalability, combined with distributed processing and buffering to address the high data rates and strict synchronization required by multi-channel MRI. Medusa uses a modular design concept based on digital synthesizer, receiver, and gradient blocks, in conjunction with fast programmable logic for sampling and synchronization. Medusa is a form of synthetic instrument, being reconfigurable for a variety of medical/scientific instrumentation needs. The Medusa distributed architecture, scalability, and data bandwidth limits are presented, and its flexibility is demonstrated in a variety of novel MRI applications. PMID:21954200

  6. Optimized scalable network switch

    DOEpatents

    Blumrich, Matthias A.; Chen, Dong; Coteus, Paul W.

    2010-02-23

    In a massively parallel computing system having a plurality of nodes configured in m multi-dimensions, each node including a computing device, a method for routing packets towards their destination nodes is provided which includes generating at least one of a 2m plurality of compact bit vectors containing information derived from downstream nodes. A multilevel arbitration process in which downstream information stored in the compact vectors, such as link status information and fullness of downstream buffers, is used to determine a preferred direction and virtual channel for packet transmission. Preferred direction ranges are encoded and virtual channels are selected by examining the plurality of compact bit vectors. This dynamic routing method eliminates the necessity of routing tables, thus enhancing scalability of the switch.

  7. Engineering scalable biological systems

    PubMed Central

    2010-01-01

    Synthetic biology is focused on engineering biological organisms to study natural systems and to provide new solutions for pressing medical, industrial and environmental problems. At the core of engineered organisms are synthetic biological circuits that execute the tasks of sensing inputs, processing logic and performing output functions. In the last decade, significant progress has been made in developing basic designs for a wide range of biological circuits in bacteria, yeast and mammalian systems. However, significant challenges in the construction, probing, modulation and debugging of synthetic biological systems must be addressed in order to achieve scalable higher-complexity biological circuits. Furthermore, concomitant efforts to evaluate the safety and biocontainment of engineered organisms and address public and regulatory concerns will be necessary to ensure that technological advances are translated into real-world solutions. PMID:21468204

  8. Optimized scalable network switch

    DOEpatents

    Blumrich, Matthias A.; Chen, Dong; Coteus, Paul W.; Gara, Alan G.; Giampapa, Mark E.; Heidelberger, Philip; Steinmacher-Burow, Burkhard D.; Takken, Todd E.; Vranas, Pavlos M.

    2007-12-04

    In a massively parallel computing system having a plurality of nodes configured in m multi-dimensions, each node including a computing device, a method for routing packets towards their destination nodes is provided which includes generating at least one of a 2m plurality of compact bit vectors containing information derived from downstream nodes. A multilevel arbitration process in which downstream information stored in the compact vectors, such as link status information and fullness of downstream buffers, is used to determine a preferred direction and virtual channel for packet transmission. Preferred direction ranges are encoded and virtual channels are selected by examining the plurality of compact bit vectors. This dynamic routing method eliminates the necessity of routing tables, thus enhancing scalability of the switch.

  9. Scalable and Cost-Effective Synthesis of Highly Efficient Fe2N-Based Oxygen Reduction Catalyst Derived from Seaweed Biomass.

    PubMed

    Liu, Long; Yang, Xianfeng; Ma, Na; Liu, Haitao; Xia, Yanzhi; Chen, Chengmeng; Yang, Dongjiang; Yao, Xiangdong

    2016-03-01

    A simple and scalable synthesis of a 3D Fe2N-based nanoaerogel is reported with superior oxygen reduction reaction activity from waste seaweed biomass, addressed the growing energy scarcity. The merits are due to the synergistic effect of the 3D porous hybrid aerogel support with excellent electrical conductivity, convenient mass transport and O2 adsorption, and core/shell structured Fe2N/N-doped amorphous carbon nanoparticles. PMID:26753802

  10. Scalable Petascale Storage for HEP using Lustre

    NASA Astrophysics Data System (ADS)

    Walker, C. J.; Traynor and, D. P.; Martin, A. J.

    2012-12-01

    We have deployed a 1 PB clustered filesystem for High Energy Physics. The use of commodity storage arrays and bonded ethernet interconnects makes the array cost effective, whilst providing high bandwidth to the storage. The filesystem is a POSIX filesytem, presented to the Grid using the StoRM Storage Resource Manager (SRM). We describe an upgrade to 10 Gbit/s networking and we present benchmarks demonstrating the performance and scalability of the filesystem.

  11. Scalable SCPPM Decoder

    NASA Technical Reports Server (NTRS)

    Quir, Kevin J.; Gin, Jonathan W.; Nguyen, Danh H.; Nguyen, Huy; Nakashima, Michael A.; Moision, Bruce E.

    2012-01-01

    A decoder was developed that decodes a serial concatenated pulse position modulation (SCPPM) encoded information sequence. The decoder takes as input a sequence of four bit log-likelihood ratios (LLR) for each PPM slot in a codeword via a XAUI 10-Gb/s quad optical fiber interface. If the decoder is unavailable, it passes the LLRs on to the next decoder via a XAUI 10-Gb/s quad optical fiber interface. Otherwise, it decodes the sequence and outputs information bits through a 1-GB/s Ethernet UDP/IP (User Datagram Protocol/Internet Protocol) interface. The throughput for a single decoder unit is 150-Mb/s at an average of four decoding iterations; by connecting a number of decoder units in series, a decoding rate equal to that of the aggregate rate is achieved. The unit is controlled through a 1-GB/s Ethernet UDP/IP interface. This ground station decoder was developed to demonstrate a deep space optical communication link capability, and is unique in the scalable design to achieve real-time SCPP decoding at the aggregate data rate.

  12. Scalable coherent interface: Links to the future

    SciTech Connect

    Gustavson, D.B.; Kristiansen, E.

    1991-11-01

    Now that the Scalable Coherent Interface (SCI) has solved the bandwidth problem, what can we use it for? SCI was developed to support closely coupled multiprocessors and their caches in a distributed shared-memory environment, but its scalability and the efficient generality of its architecture make it work very well over a wide range of applications. It can replace a local area network for connecting workstations on a campus. It can be powerful I/O channel for a supercomputer. It can be the processor-cache-memory-I/O connection in a highly parallel computer. It can gather data from enormous particle detectors and distribute it among thousands of processors. It can connect a desktop microprocessor to memory chips a few millimeters away, disk drivers a few meters away, and servers a few kilometers away.

  13. Scalable coherent interface: Links to the future

    SciTech Connect

    Gustavson, D.B. ); Kristiansen, E. )

    1991-11-01

    Now that the Scalable Coherent Interface (SCI) has solved the bandwidth problem, what can we use it for SCI was developed to support closely coupled multiprocessors and their caches in a distributed shared-memory environment, but its scalability and the efficient generality of its architecture make it work very well over a wide range of applications. It can replace a local area network for connecting workstations on a campus. It can be powerful I/O channel for a supercomputer. It can be the processor-cache-memory-I/O connection in a highly parallel computer. It can gather data from enormous particle detectors and distribute it among thousands of processors. It can connect a desktop microprocessor to memory chips a few millimeters away, disk drivers a few meters away, and servers a few kilometers away.

  14. A Scalability Model for ECS's Data Server

    NASA Technical Reports Server (NTRS)

    Menasce, Daniel A.; Singhal, Mukesh

    1998-01-01

    This report presents in four chapters a model for the scalability analysis of the Data Server subsystem of the Earth Observing System Data and Information System (EOSDIS) Core System (ECS). The model analyzes if the planned architecture of the Data Server will support an increase in the workload with the possible upgrade and/or addition of processors, storage subsystems, and networks. The approaches in the report include a summary of the architecture of ECS's Data server as well as a high level description of the Ingest and Retrieval operations as they relate to ECS's Data Server. This description forms the basis for the development of the scalability model of the data server and the methodology used to solve it.

  15. Design and implementation of scalable tape archiver

    NASA Technical Reports Server (NTRS)

    Nemoto, Toshihiro; Kitsuregawa, Masaru; Takagi, Mikio

    1996-01-01

    In order to reduce costs, computer manufacturers try to use commodity parts as much as possible. Mainframes using proprietary processors are being replaced by high performance RISC microprocessor-based workstations, which are further being replaced by the commodity microprocessor used in personal computers. Highly reliable disks for mainframes are also being replaced by disk arrays, which are complexes of disk drives. In this paper we try to clarify the feasibility of a large scale tertiary storage system composed of 8-mm tape archivers utilizing robotics. In the near future, the 8-mm tape archiver will be widely used and become a commodity part, since recent rapid growth of multimedia applications requires much larger storage than disk drives can provide. We designed a scalable tape archiver which connects as many 8-mm tape archivers (element archivers) as possible. In the scalable archiver, robotics can exchange a cassette tape between two adjacent element archivers mechanically. Thus, we can build a large scalable archiver inexpensively. In addition, a sophisticated migration mechanism distributes frequently accessed tapes (hot tapes) evenly among all of the element archivers, which improves the throughput considerably. Even with the failures of some tape drives, the system dynamically redistributes hot tapes to the other element archivers which have live tape drives. Several kinds of specially tailored huge archivers are on the market, however, the 8-mm tape scalable archiver could replace them. To maintain high performance in spite of high access locality when a large number of archivers are attached to the scalable archiver, it is necessary to scatter frequently accessed cassettes among the element archivers and to use the tape drives efficiently. For this purpose, we introduce two cassette migration algorithms, foreground migration and background migration. Background migration transfers cassettes between element archivers to redistribute frequently accessed

  16. Memory Scalability and Efficiency Analysis of Parallel Codes

    SciTech Connect

    Janjusic, Tommy; Kartsaklis, Christos

    2015-01-01

    Memory scalability is an enduring problem and bottleneck that plagues many parallel codes. Parallel codes designed for High Performance Systems are typically designed over the span of several, and in some instances 10+, years. As a result, optimization practices which were appropriate for earlier systems may no longer be valid and thus require careful optimization consideration. Specifically, parallel codes whose memory footprint is a function of their scalability must be carefully considered for future exa-scale systems. In this paper we present a methodology and tool to study the memory scalability of parallel codes. Using our methodology we evaluate an application s memory footprint as a function of scalability, which we coined memory efficiency, and describe our results. In particular, using our in-house tools we can pinpoint the specific application components which contribute to the application s overall memory foot-print (application data- structures, libraries, etc.).

  17. Pursuing Scalability for hypre's Conceptual Interfaces

    SciTech Connect

    Falgout, R D; Jones, J E; Yang, U M

    2004-07-21

    The software library hypre provides high performance preconditioners and solvers for the solution of large, sparse linear systems on massively parallel computers as well as conceptual interfaces that allow users to access the library in the way they naturally think about their problems. These interfaces include a stencil-based structured interface (Struct); a semi-structured interface (semiStruct), which is appropriate for applications that are mostly structured, e.g. block structured grids, composite grids in structured adaptive mesh refinement applications, and overset grids; a finite element interface (FEI) for unstructured problems, as well as a conventional linear-algebraic interface (IJ). It is extremely important to provide an efficient, scalable implementation of these interfaces in order to support the scalable solvers of the library, especially when using tens of thousands of processors. This paper describes the data structures, parallel implementation and resulting performance of the IJ, Struct and semiStruct interfaces. It investigates their scalability, presents successes as well as pitfalls of some of the approaches and suggests ways of dealing with them.

  18. A scalable method for the production of high-titer and high-quality adeno-associated type 9 vectors using the HSV platform

    PubMed Central

    Adamson-Small, Laura; Potter, Mark; Falk, Darin J; Cleaver, Brian; Byrne, Barry J; Clément, Nathalie

    2016-01-01

    Recombinant adeno-associated vectors based on serotype 9 (rAAV9) have demonstrated highly effective gene transfer in multiple animal models of muscular dystrophies and other neurological indications. Current limitations in vector production and purification have hampered widespread implementation of clinical candidate vectors, particularly when systemic administration is considered. In this study, we describe a complete herpes simplex virus (HSV)-based production and purification process capable of generating greater than 1 × 1014 rAAV9 vector genomes per 10-layer CellSTACK of HEK 293 producer cells, or greater than 1 × 105 vector genome per cell, in a final, fully purified product. This represents a 5- to 10-fold increase over transfection-based methods. In addition, rAAV vectors produced by this method demonstrated improved biological characteristics when compared to transfection-based production, including increased infectivity as shown by higher transducing unit-to-vector genome ratios and decreased total capsid protein amounts, shown by lower empty-to-full ratios. Together, this data establishes a significant improvement in both rAAV9 yields and vector quality. Further, the method can be readily adapted to large-scale good laboratory practice (GLP) and good manufacturing practice (GMP) production of rAAV9 vectors to enable preclinical and clinical studies and provide a platform to build on toward late-phases and commercial production. PMID:27222839

  19. Libra: Scalable Load Balance Analysis

    SciTech Connect

    2009-09-16

    Libra is a tool for scalable analysis of load balance data from all processes in a parallel application. Libra contains an instrumentation module that collects model data from parallel applications and a parallel compression mechanism that uses distributed wavelet transforms to gather load balance model data in a scalable fashion. Data is output to files, and these files can be viewed in a GUI tool by Libra users. The GUI tool associates particular load balance data with regions for code, emabling users to view the load balance properties of distributed "slices" of their application code.

  20. Libra: Scalable Load Balance Analysis

    Energy Science and Technology Software Center (ESTSC)

    2009-09-16

    Libra is a tool for scalable analysis of load balance data from all processes in a parallel application. Libra contains an instrumentation module that collects model data from parallel applications and a parallel compression mechanism that uses distributed wavelet transforms to gather load balance model data in a scalable fashion. Data is output to files, and these files can be viewed in a GUI tool by Libra users. The GUI tool associates particular load balancemore » data with regions for code, emabling users to view the load balance properties of distributed "slices" of their application code.« less

  1. An HEVC extension for spatial and quality scalable video coding

    NASA Astrophysics Data System (ADS)

    Hinz, Tobias; Helle, Philipp; Lakshman, Haricharan; Siekmann, Mischa; Stegemann, Jan; Schwarz, Heiko; Marpe, Detlev; Wiegand, Thomas

    2013-02-01

    This paper describes an extension of the upcoming High Efficiency Video Coding (HEVC) standard for supporting spatial and quality scalable video coding. Besides scalable coding tools known from scalable profiles of prior video coding standards such as H.262/MPEG-2 Video and H.264/MPEG-4 AVC, the proposed scalable HEVC extension includes new coding tools that further improve the coding efficiency of the enhancement layer. In particular, new coding modes by which base and enhancement layer signals are combined for forming an improved enhancement layer prediction signal have been added. All scalable coding tools have been integrated in a way that the low-level syntax and decoding process of HEVC remain unchanged to a large extent. Simulation results for typical application scenarios demonstrate the effectiveness of the proposed design. For spatial and quality scalable coding with two layers, bit-rate savings of about 20-30% have been measured relative to simulcasting the layers, which corresponds to a bit-rate overhead of about 5-15% relative to single-layer coding of the enhancement layer.

  2. Scalability study of solid xenon

    SciTech Connect

    Yoo, J.; Cease, H.; Jaskierny, W. F.; Markley, D.; Pahlka, R. B.; Balakishiyeva, D.; Saab, T.; Filipenko, M.

    2015-04-01

    We report a demonstration of the scalability of optically transparent xenon in the solid phase for use as a particle detector above a kilogram scale. We employed a cryostat cooled by liquid nitrogen combined with a xenon purification and chiller system. A modified {\\it Bridgeman's technique} reproduces a large scale optically transparent solid xenon.

  3. The relation of scalability and execution time

    NASA Technical Reports Server (NTRS)

    Sun, Xian-He

    1995-01-01

    Scalability has been used extensively as a de facto performance criterion for evaluating parallel algorithms and architectures. However, for many, scalability has theoretical interests only since it does not reveal execution time. In this paper, the relation between scalability and execution time is carefully studied. Results show that the isospeed scalability well characterizes the variation of execution time: smaller scalability leads to larger execution time, the same scalability leads to the same execution time, etc. Three algorithms from scientific computing are implemented on an Intel Paragon and an IBM SP2 parallel computer. Experimental and theoretical results show that scalability is an important, distinct metric for parallel and distributed systems, and may be as important as execution time in a scalable parallel and distributed environment.

  4. Benchmarking and parallel scalability of MANCINTAP, a Parallel High-Performance Tool For Neutron Activation Analysis in Complex 4D Scenarios

    NASA Astrophysics Data System (ADS)

    Firpo, G.; Frambati, S.; Frignani, M.; Gerra, G.

    2014-06-01

    MANCINTAP is a parallel computational tool developed by Ansaldo Nucleare to perform 4D neutron transport, activation and time-resolved dose-rate calculations in very complex geometries for CPU-intensive fission and fusion applications. MANCINTAP creates an automated link between the 3D radiation transport code MCNP5—which is used to evaluate both the neutron fluxes for activation calculations and the resulting secondary gamma dose rates—and the zero-dimensional activation code Anita2000 by handling crucial processes such as data exchange, determination of material mixtures and generation of cumulative probability distributions. A brief description of the computational tool is given here, with particular emphasis on the key technical choices underlying the project. Benchmarking of MANCINTAP has been performed in three steps: (i) against a very simplified model, where an analytical solution is available for comparison; (ii) against the well-established deterministic transport and activation code ATTILA and (iii) against experimental data obtained at the Frascati Neutron Generator (FNG) facility. An analysis of MANCINTAP scalability performances is proposed to demonstrate the robustness of its parallel structure, tailored for HPC applications, which makes it—to the best of our knowledge—a novel tool.

  5. Scalable hardbody and plume optical signatures

    NASA Astrophysics Data System (ADS)

    Crow, Dennis R.; Hawes, Fred; Braunstein, Matthew; Coker, Charles F.; Smith, Thomas, Jr.

    2004-08-01

    The Fast Line-of-sight Imagery for Target and Exhaust Signatures (FLITES) is a High Performance Computing (HPC-CHSSI) and Missile Defense Agency (MDA) funded effort that provides a scalable program to compute highly resolved temporal, spatial, and spectral hardbody and plume optical signatures. Distributed processing capabilities are included to allow complex, high fidelity, solutions to be generated quickly generated. The distributed processing logic includes automated load balancing algorithms to facilitate scalability using large numbers of processors. To enhance exhaust plume optical signature capabilities, FLITES employs two different radiance transport algorithms. The first algorithm is the traditional Curtis-Godson bandmodel approach and is provided to support comparisons to historical results and high-frame rate production requirements. The second algorithm is the Quasi Bandmodel Line-by-line (QBL) approach, which uses randomly placed "cloned" spectral lines to yield highly resolved radiation spectra for increased accuracy while maintaining tractable runtimes. This capability will provide a significant advancement over the traditional SPURC/SIRRM radiance transport methodology.

  6. Scalable ATM encryption

    SciTech Connect

    1995-04-01

    In order to provide needed security assurances for traffic carried in Asynchronous Transfer Mode (ATM) networks, methods of protecting the integrity and privacy of traffic must be employed. Cryptographic methods can be used to assure authenticity and privacy, but are hard to scale and the incorporation of these methods into computer networks can severely impact functionality, reliability, and performance. To study these trade-offs, a research prototype encryptor/decryptor is under development. This prototype is to demonstrate the viability of implementing certain encryption techniques in high speed networks by processing Asynchronous Transfer Mode (ATM) cells in a SONET OC-3 payload. This paper describes the objectives and design trade-offs intended to be investigated with the prototype. User requirements for high performance computing and communication have driven Sandia to do work in the areas of functionality, reliability, security, and performance of high speed communication networks. Adherence to standards (including emerging standards) achieves greater functionality of high speed computer networks by providing wide interoperability of applications, network hardware, and network software.

  7. A scalable and operationally simple radical trifluoromethylation

    PubMed Central

    Beatty, Joel W.; Douglas, James J.; Cole, Kevin P.; Stephenson, Corey R. J.

    2015-01-01

    The large number of reagents that have been developed for the synthesis of trifluoromethylated compounds is a testament to the importance of the CF3 group as well as the associated synthetic challenge. Current state-of-the-art reagents for appending the CF3 functionality directly are highly effective; however, their use on preparative scale has minimal precedent because they require multistep synthesis for their preparation, and/or are prohibitively expensive for large-scale application. For a scalable trifluoromethylation methodology, trifluoroacetic acid and its anhydride represent an attractive solution in terms of cost and availability; however, because of the exceedingly high oxidation potential of trifluoroacetate, previous endeavours to use this material as a CF3 source have required the use of highly forcing conditions. Here we report a strategy for the use of trifluoroacetic anhydride for a scalable and operationally simple trifluoromethylation reaction using pyridine N-oxide and photoredox catalysis to affect a facile decarboxylation to the CF3 radical. PMID:26258541

  8. A scalable and operationally simple radical trifluoromethylation.

    PubMed

    Beatty, Joel W; Douglas, James J; Cole, Kevin P; Stephenson, Corey R J

    2015-01-01

    The large number of reagents that have been developed for the synthesis of trifluoromethylated compounds is a testament to the importance of the CF3 group as well as the associated synthetic challenge. Current state-of-the-art reagents for appending the CF3 functionality directly are highly effective; however, their use on preparative scale has minimal precedent because they require multistep synthesis for their preparation, and/or are prohibitively expensive for large-scale application. For a scalable trifluoromethylation methodology, trifluoroacetic acid and its anhydride represent an attractive solution in terms of cost and availability; however, because of the exceedingly high oxidation potential of trifluoroacetate, previous endeavours to use this material as a CF3 source have required the use of highly forcing conditions. Here we report a strategy for the use of trifluoroacetic anhydride for a scalable and operationally simple trifluoromethylation reaction using pyridine N-oxide and photoredox catalysis to affect a facile decarboxylation to the CF3 radical. PMID:26258541

  9. Scalable parallel communications

    NASA Technical Reports Server (NTRS)

    Maly, K.; Khanna, S.; Overstreet, C. M.; Mukkamala, R.; Zubair, M.; Sekhar, Y. S.; Foudriat, E. C.

    1992-01-01

    Coarse-grain parallelism in networking (that is, the use of multiple protocol processors running replicated software sending over several physical channels) can be used to provide gigabit communications for a single application. Since parallel network performance is highly dependent on real issues such as hardware properties (e.g., memory speeds and cache hit rates), operating system overhead (e.g., interrupt handling), and protocol performance (e.g., effect of timeouts), we have performed detailed simulations studies of both a bus-based multiprocessor workstation node (based on the Sun Galaxy MP multiprocessor) and a distributed-memory parallel computer node (based on the Touchstone DELTA) to evaluate the behavior of coarse-grain parallelism. Our results indicate: (1) coarse-grain parallelism can deliver multiple 100 Mbps with currently available hardware platforms and existing networking protocols (such as Transmission Control Protocol/Internet Protocol (TCP/IP) and parallel Fiber Distributed Data Interface (FDDI) rings); (2) scale-up is near linear in n, the number of protocol processors, and channels (for small n and up to a few hundred Mbps); and (3) since these results are based on existing hardware without specialized devices (except perhaps for some simple modifications of the FDDI boards), this is a low cost solution to providing multiple 100 Mbps on current machines. In addition, from both the performance analysis and the properties of these architectures, we conclude: (1) multiple processors providing identical services and the use of space division multiplexing for the physical channels can provide better reliability than monolithic approaches (it also provides graceful degradation and low-cost load balancing); (2) coarse-grain parallelism supports running several transport protocols in parallel to provide different types of service (for example, one TCP handles small messages for many users, other TCP's running in parallel provide high bandwidth

  10. A Scalable Database Infrastructure

    NASA Astrophysics Data System (ADS)

    Arko, R. A.; Chayes, D. N.

    2001-12-01

    The rapidly increasing volume and complexity of MG&G data, and the growing demand from funding agencies and the user community that it be easily accessible, demand that we improve our approach to data management in order to reach a broader user-base and operate more efficient and effectively. We have chosen an approach based on industry-standard relational database management systems (RDBMS) that use community-wide data specifications, where there is a clear and well-documented external interface that allows use of general purpose as well as customized clients. Rapid prototypes assembled with this approach show significant advantages over the traditional, custom-built data management systems that often use "in-house" legacy file formats, data specifications, and access tools. We have developed an effective database prototype based a public domain RDBMS (PostgreSQL) and metadata standard (FGDC), and used it as a template for several ongoing MG&G database management projects - including ADGRAV (Antarctic Digital Gravity Synthesis), MARGINS, the Community Review system of the Digital Library for Earth Science Education, multibeam swath bathymetry metadata, and the R/V Maurice Ewing onboard acquisition system. By using standard formats and specifications, and working from a common prototype, we are able to reuse code and deploy rapidly. Rather than spend time on low-level details such as storage and indexing (which are built into the RDBMS), we can focus on high-level details such as documentation and quality control. In addition, because many commercial off-the-shelf (COTS) and public domain data browsers and visualization tools have built-in RDBMS support, we can focus on backend development and leave the choice of a frontend client(s) up to the end user. While our prototype is running under an open source RDBMS on a single processor host, the choice of standard components allows this implementation to scale to commercial RDBMS products and multiprocessor servers as

  11. Computational scalability of large size image dissemination

    NASA Astrophysics Data System (ADS)

    Kooper, Rob; Bajcsy, Peter

    2011-01-01

    We have investigated the computational scalability of image pyramid building needed for dissemination of very large image data. The sources of large images include high resolution microscopes and telescopes, remote sensing and airborne imaging, and high resolution scanners. The term 'large' is understood from a user perspective which means either larger than a display size or larger than a memory/disk to hold the image data. The application drivers for our work are digitization projects such as the Lincoln Papers project (each image scan is about 100-150MB or about 5000x8000 pixels with the total number to be around 200,000) and the UIUC library scanning project for historical maps from 17th and 18th century (smaller number but larger images). The goal of our work is understand computational scalability of the web-based dissemination using image pyramids for these large image scans, as well as the preservation aspects of the data. We report our computational benchmarks for (a) building image pyramids to be disseminated using the Microsoft Seadragon library, (b) a computation execution approach using hyper-threading to generate image pyramids and to utilize the underlying hardware, and (c) an image pyramid preservation approach using various hard drive configurations of Redundant Array of Independent Disks (RAID) drives for input/output operations. The benchmarks are obtained with a map (334.61 MB, JPEG format, 17591x15014 pixels). The discussion combines the speed and preservation objectives.

  12. Unequal erasure protection technique for scalable multistreams.

    PubMed

    Dumitrescu, Sorina; Rivers, Geoffrey; Shirani, Shahram

    2010-02-01

    This paper presents a novel unequal erasure protection (UEP) strategy for the transmission of scalable data, formed by interleaving independently decodable and scalable streams, over packet erasure networks. The technique, termed multistream UEP (M-UEP), differs from the traditional UEP strategy by: 1) placing separate streams in separate packets to establish independence and 2) using permuted systematic Reed-Solomon codes to enhance the distribution of message symbols amongst the packets. M-UEP improves upon UEP by ensuring that all received source symbols are decoded. The R-D optimal redundancy allocation problem for M-UEP is formulated and its globally optimal solution is shown to have a time complexity of O(2(N)N(L+1)(N+1)) , where N is the number of packets and L is the packet length. To address the high complexity of the globally optimal solution, an efficient suboptimal algorithm is proposed which runs in O(N(2)L(2)) time. The proposed M-UEP algorithm is applied on SPIHT coded images in conjunction with an appropriate grouping of wavelet coefficients into streams. The experimental results reveal that M-UEP consistently outperforms the traditional UEP reaching peak improvements of 0.6 dB. Moreover, our tests show that M-UEP is more robust than UEP in adverse channel conditions. PMID:19783503

  13. Scalable broadband OPCPA in Lithium Niobate with signal angular dispersion

    NASA Astrophysics Data System (ADS)

    Tóth, György; Pálfalvi, László; Tokodi, Levente; Hebling, János; Fülöp, József András

    2016-07-01

    Angular dispersion of the signal beam is proposed for efficient, scalable high-power few-cycle pulse generation in LiNbO3 by optical parametric chirped-pulse amplification (OPCPA) in the 1.4 to 2.1 μm wavelength range. An optimized double-grating setup can provide the required angular dispersion. Calculations predict 16.8 fs (3 cycles) pulses with 13 TW peak power. Further scalability of the scheme towards the 100-TW power level is feasible by using efficient, cost-effective, compact diode-pumped solid-state lasers for pumping directly at 1 μm, without second-harmonic generation.

  14. Scalable chemical oxygen - iodine laser

    SciTech Connect

    Adamenkov, A A; Bakshin, V V; Vyskubenko, B A; Efremov, V I; Il'in, S P; Ilyushin, Yurii N; Kolobyanin, Yu V; Kudryashov, E A; Troshkin, M V

    2011-12-31

    The problem of scaling chemical oxygen - iodine lasers (COILs) is discussed. The results of experimental study of a twisted-aerosol singlet oxygen generator meeting the COIL scalability requirements are presented. The energy characteristics of a supersonic COIL with singlet oxygen and iodine mixing in parallel flows are also experimentally studied. The output power of {approx}7.5 kW, corresponding to a specific power of 230 W cm{sup -2}, is achieved. The maximum chemical efficiency of the COIL is {approx}30%.

  15. Scalable, enantioselective taxane total synthesis

    PubMed Central

    Mendoza, Abraham; Ishihara, Yoshihiro; Baran, Phil S.

    2011-01-01

    Taxanes are a large family of terpenes comprising over 350 members, the most famous of which is Taxol (paclitaxel) — a billion-dollar anticancer drug. Here, we describe the first practical and scalable synthetic entry to these natural products via a concise preparation of (+)-taxa-4(5),11(12)-dien-2-one, which possesses a suitable functional handle to access more oxidised members of its family. This route enabled a gram-scale preparation of the ”parent” taxane, taxadiene, representing the largest quantity of this naturally occurring terpene ever isolated or prepared in pure form. The taxane family’s characteristic 6-8-6 tricyclic system containing a bridgehead alkene is forged via a vicinal difunctionalisation/Diels–Alder strategy. Asymmetry is introduced by means of an enantioselective conjugate addition that forms an all-carbon quaternary centre, from which all other stereocentres are fixed via substrate control. This study lays a critical foundation for a planned access to minimally oxidised taxane analogs and a scalable laboratory preparation of Taxol itself. PMID:22169867

  16. Network selection, Information filtering and Scalable computation

    NASA Astrophysics Data System (ADS)

    Ye, Changqing

    -complete factorizations, possibly with a high percentage of missing values. This promotes additional sparsity beyond rank reduction. Computationally, we design methods based on a ``decomposition and combination'' strategy, to break large-scale optimization into many small subproblems to solve in a recursive and parallel manner. On this basis, we implement the proposed methods through multi-platform shared-memory parallel programming, and through Mahout, a library for scalable machine learning and data mining, for mapReduce computation. For example, our methods are scalable to a dataset consisting of three billions of observations on a single machine with sufficient memory, having good timings. Both theoretical and numerical investigations show that the proposed methods exhibit significant improvement in accuracy over state-of-the-art scalable methods.

  17. Practical, Scalable, High-Throughput Approaches to η3-Pyranyl and η3-Pyridinyl Organometallic Enantiomeric Scaffolds Using the Achmatowicz Reaction

    PubMed Central

    Coombs, Thomas C.; Lee, Maurice D.; Wong, Heilam; Armstrong, Matthew; Cheng, Bo; Chen, Wenyong; Moretto, Alessandro F.; Liebeskind, Lanny S.

    2009-01-01

    A unified strategy for the high throughput synthesis of multigram quantities of the η3-oxopyranyl- and η3-oxopyridinylmolybdenum complexes TpMo(CO)2(η3-oxopyranyl) and TpMo(CO)2(η3-oxopyridinyl) is described (Tp = hydridotrispyrazolylborato). The strategy uses the oxa- and aza-Achmatowicz reaction for the preparation of these organometallic enantiomeric scaffolds, in both racemic and high enantiopurity versions. PMID:18171075

  18. Scalable computer architecture for digital vascular systems

    NASA Astrophysics Data System (ADS)

    Goddard, Iain; Chao, Hui; Skalabrin, Mark

    1998-06-01

    Digital vascular computer systems are used for radiology and fluoroscopy (R/F), angiography, and cardiac applications. In the United States alone, about 26 million procedures of these types are performed annually: about 81% R/F, 11% cardiac, and 8% angiography. Digital vascular systems have a very wide range of performance requirements, especially in terms of data rates. In addition, new features are added over time as they are shown to be clinically efficacious. Application-specific processing modes such as roadmapping, peak opacification, and bolus chasing are particular to some vascular systems. New algorithms continue to be developed and proven, such as Cox and deJager's precise registration methods for masks and live images in digital subtraction angiography. A computer architecture must have high scalability and reconfigurability to meet the needs of this modality. Ideally, the architecture could also serve as the basis for a nonvascular R/F system.

  19. Efficient scalable solid-state neutron detector

    SciTech Connect

    Moses, Daniel

    2015-06-15

    We report on scalable solid-state neutron detector system that is specifically designed to yield high thermal neutron detection sensitivity. The basic detector unit in this system is made of a {sup 6}Li foil coupled to two crystalline silicon diodes. The theoretical intrinsic efficiency of a detector-unit is 23.8% and that of detector element comprising a stack of five detector-units is 60%. Based on the measured performance of this detector-unit, the performance of a detector system comprising a planar array of detector elements, scaled to encompass effective area of 0.43 m{sup 2}, is estimated to yield the minimum absolute efficiency required of radiological portal monitors used in homeland security.

  20. Efficient scalable solid-state neutron detector

    NASA Astrophysics Data System (ADS)

    Moses, Daniel

    2015-06-01

    We report on scalable solid-state neutron detector system that is specifically designed to yield high thermal neutron detection sensitivity. The basic detector unit in this system is made of a 6Li foil coupled to two crystalline silicon diodes. The theoretical intrinsic efficiency of a detector-unit is 23.8% and that of detector element comprising a stack of five detector-units is 60%. Based on the measured performance of this detector-unit, the performance of a detector system comprising a planar array of detector elements, scaled to encompass effective area of 0.43 m2, is estimated to yield the minimum absolute efficiency required of radiological portal monitors used in homeland security.

  1. Efficient scalable solid-state neutron detector.

    PubMed

    Moses, Daniel

    2015-06-01

    We report on scalable solid-state neutron detector system that is specifically designed to yield high thermal neutron detection sensitivity. The basic detector unit in this system is made of a (6)Li foil coupled to two crystalline silicon diodes. The theoretical intrinsic efficiency of a detector-unit is 23.8% and that of detector element comprising a stack of five detector-units is 60%. Based on the measured performance of this detector-unit, the performance of a detector system comprising a planar array of detector elements, scaled to encompass effective area of 0.43 m(2), is estimated to yield the minimum absolute efficiency required of radiological portal monitors used in homeland security. PMID:26133869

  2. Scalable Performance Measurement and Analysis

    SciTech Connect

    Gamblin, Todd

    2009-01-01

    Concurrency levels in large-scale, distributed-memory supercomputers are rising exponentially. Modern machines may contain 100,000 or more microprocessor cores, and the largest of these, IBM's Blue Gene/L, contains over 200,000 cores. Future systems are expected to support millions of concurrent tasks. In this dissertation, we focus on efficient techniques for measuring and analyzing the performance of applications running on very large parallel machines. Tuning the performance of large-scale applications can be a subtle and time-consuming task because application developers must measure and interpret data from many independent processes. While the volume of the raw data scales linearly with the number of tasks in the running system, the number of tasks is growing exponentially, and data for even small systems quickly becomes unmanageable. Transporting performance data from so many processes over a network can perturb application performance and make measurements inaccurate, and storing such data would require a prohibitive amount of space. Moreover, even if it were stored, analyzing the data would be extremely time-consuming. In this dissertation, we present novel methods for reducing performance data volume. The first draws on multi-scale wavelet techniques from signal processing to compress systemwide, time-varying load-balance data. The second uses statistical sampling to select a small subset of running processes to generate low-volume traces. A third approach combines sampling and wavelet compression to stratify performance data adaptively at run-time and to reduce further the cost of sampled tracing. We have integrated these approaches into Libra, a toolset for scalable load-balance analysis. We present Libra and show how it can be used to analyze data from large scientific applications scalably.

  3. Highly efficient blue organic light emitting device using indium-free transparent anode Ga:ZnO with scalability for large area coating

    SciTech Connect

    Wang Liang; Matson, Dean W.; Polikarpov, Evgueni; Swensen, James S.; Bonham, Charles C.; Cosimbescu, Lelia; Gaspar, Daniel J.; Padmaperuma, Asanga B.; Berry, Joseph J.; Ginley, David S.

    2010-02-15

    Organic light emitting devices have been achieved with an indium-free transparent anode, Ga doped ZnO (GZO). A large area coating technique was used (RF magnetron sputtering) to deposit the GZO films onto glass. The respective organic light emitting devices exhibited an operational voltage of 3.7 V, an external quantum efficiency of 17%, and a power efficiency of 39 lm/W at a current density of 1 mA/cm{sup 2}. These parameters are well within acceptable standards for blue OLEDs to generate a white light with high enough brightness for general lighting applications. It is expected that high-efficiency, long-lifetime, large area, and cost-effective white OLEDs can be made with these indium-free anode materials.

  4. A scalable parallel open architecture data acquisition system for low to high rate experiments, test beams and all SSC (Superconducting Super Collider) detectors

    SciTech Connect

    Barsotti, E.; Booth, A.; Bowden, M.; Swoboda, C. ); Lockyer, N.; VanBerg, R. )

    1989-12-01

    A new era of high-energy physics research is beginning requiring accelerators with much higher luminosities and interaction rates in order to discover new elementary particles. As a consequences, both orders of magnitude higher data rates from the detector and online processing power, well beyond the capabilities of current high energy physics data acquisition systems, are required. This paper describes a new data acquisition system architecture which draws heavily from the communications industry, is totally parallel (i.e., without any bottlenecks), is capable of data rates of hundreds of GigaBytes per second from the detector and into an array of online processors (i.e., processor farm), and uses an open systems architecture to guarantee compatibility with future commercially available online processor farms. The main features of the system architecture are standard interface ICs to detector subsystems wherever possible, fiber optic digital data transmission from the near-detector electronics, a self-routing parallel event builder, and the use of industry-supported and high-level language programmable processors in the proposed BCD system for both triggers and online filters. A brief status report of an ongoing project at Fermilab to build the self-routing parallel event builder will also be given in the paper. 3 figs., 1 tab.

  5. Stereoscopic video compression using temporal scalability

    NASA Astrophysics Data System (ADS)

    Puri, Atul; Kollarits, Richard V.; Haskell, Barry G.

    1995-04-01

    Despite the fact that human ability to perceive a high degree of realism is directly related to our ability to perceive depth accurately in a scene, most of the commonly used imaging and display technologies are able to provide only a 2D rendering of the 3D real world. Many current as well as emerging applications in areas of entertainment, remote operations, industrial and medicine can benefit from the depth perception offered by stereoscopic video systems which employ two views of a scene imaged under the constraints imposed by human visual system. Among the many challenges to be overcome for practical realization and widespread use of 3D/stereoscopic systems are efficient techniques for digital compression of enormous amounts of data while maintaining compatibility with normal video decoding and display systems. After a brief discussion on the relationship of digital stereoscopic 3DTV with digital TV and HDTV, we present an overview of tools in the MPEG-2 video standard that are relevant to our discussion on compression of stereoscopic video, which is the main topic of this paper. Next, we determine ways in which temporal scalability concepts can be applied to exploit redundancies inherent between the two views of a scene comprising stereoscopic video. Due consideration is given to masking properties of stereoscopic vision to determine bandwidth partitioning between the two views to realize an efficient coding scheme while providing sufficient quality. Simulations are performed on stereoscopic video of normal TV resolution to compare the performance of the two temporal scalability configurations with each other and with the simulcast solution. Preliminary results are quite promising and indicate that the configuration that exploits motion and disparity compensation significantly outperforms the one that exploits disparity compensation alone. Compression of both views of stereo video of normal TV resolution appears feasible in a total of 8 or 9 Mbit/s. Finally

  6. SWIFT—Scalable Clustering for Automated Identification of Rare Cell Populations in Large, High-Dimensional Flow Cytometry Datasets, Part 2: Biological Evaluation

    PubMed Central

    Mosmann, Tim R; Naim, Iftekhar; Rebhahn, Jonathan; Datta, Suprakash; Cavenaugh, James S; Weaver, Jason M; Sharma, Gaurav

    2014-01-01

    A multistage clustering and data processing method, SWIFT (detailed in a companion manuscript), has been developed to detect rare subpopulations in large, high-dimensional flow cytometry datasets. An iterative sampling procedure initially fits the data to multidimensional Gaussian distributions, then splitting and merging stages use a criterion of unimodality to optimize the detection of rare subpopulations, to converge on a consistent cluster number, and to describe non-Gaussian distributions. Probabilistic assignment of cells to clusters, visualization, and manipulation of clusters by their cluster medians, facilitate application of expert knowledge using standard flow cytometry programs. The dual problems of rigorously comparing similar complex samples, and enumerating absent or very rare cell subpopulations in negative controls, were solved by assigning cells in multiple samples to a cluster template derived from a single or combined sample. Comparison of antigen-stimulated and control human peripheral blood cell samples demonstrated that SWIFT could identify biologically significant subpopulations, such as rare cytokine-producing influenza-specific T cells. A sensitivity of better than one part per million was attained in very large samples. Results were highly consistent on biological replicates, yet the analysis was sensitive enough to show that multiple samples from the same subject were more similar than samples from different subjects. A companion manuscript (Part 1) details the algorithmic development of SWIFT. © 2014 The Authors. Published by Wiley Periodicals Inc. PMID:24532172

  7. Highly efficient blue organic light emitting device using indium-free transparent anode Ga:ZnO with scalability for large area coating

    SciTech Connect

    Wang, Liang; Matson, Dean W.; Polikarpov, Evgueni; Swensen, James S.; Bonham, Charles C.; Cosimbescu, Lelia; Berry, J. J.; Ginley, D. S.; Gaspar, Daniel J.; Padmaperuma, Asanga B.

    2010-02-15

    The availability of economically-produced and environmentally-stable transparent conductive oxide (TCO) coatings is critical for the development of a variety of electronic devices requiring transparent electrodes. Such devices include liquid crystal display pixels and organic light emitting diodes (OLEDs),[1, 2] solar cell applications,[3, 4] and electrically heated windows.[5, 6] The materials fulfilling these requirements are usually wide band gap inorganic transparent conductive oxides (TCOs). Tin-doped indium oxide, or ITO, has traditionally been used for electronic TCO applications because of its low resistivity, high work function and transparency. Due to the increasing cost and limited supply of indium and its tendency to migrate in to the device, there has been increasing research interest to substitute ITO with an indium-free material. A number of alternative metal oxides and doped oxides have been evaluated as TCO materials with varying degrees of success.[7, 8] Among these alternatives to ITO, gallium-doped zinc oxide (GZO) [2, 9] and aluminium-doped zinc oxide (AZO) [10, 11] have drawn particular attention. These materials have been demonstrated to have resistivities and transparencies approaching those of the best ITO, low toxicity, and much lower materials cost. Although AZO is attractive as a TCO electrode material, GZO features a greater resistance to oxidation as a result of gallium’s greater electronegativity compared to Submitted to 2 aluminum.[12, 13

  8. Scalable synthesis of core-shell structured SiOx/nitrogen-doped carbon composite as a high-performance anode material for lithium-ion batteries

    NASA Astrophysics Data System (ADS)

    Shi, Lu; Wang, Weikun; Wang, Anbang; Yuan, Keguo; Jin, Zhaoqing; Yang, Yusheng

    2016-06-01

    In this work, a novel core-shell structured SiOx/nitrogen-doped carbon composite has been prepared by simply dispersing the SiOx particles, which are synthesized by a thermal evaporation method from an equimolar mixture of Si and SiO2, into the dopamine solution, followed by a carbonization process. The SiOx core is well covered by the conformal and homogeneous nitrogen-doped carbon layer from the pyrolysis of polydopamine. By contrast with the bare SiOx, the electrochemical performance of the as-prepared core-shell structured SiOx/nitrogen-doped carbon composite has been improved significantly. It delivers a reversible capacity of 1514 mA h g-1 after 100 cycles at a current density of 100 mA g-1 and 933 mA h g-1 at 2 A g-1, much higher than those of commercial graphite anodes. The nitrogen-doped carbon layer ensures the excellent electrochemical performance of the SiOx/C composite. In addition, since dopamine can self-polymerize and coat virtually any surface, this versatile, facile and highly efficient coating process may be widely applicable to obtain various composites with uniform nitrogen-doped carbon coating layer.

  9. Final Project Report: DOE Award FG02-04ER25606 Overlay Transit Networking for Scalable, High Performance Data Communication across Heterogeneous Infrastructure

    SciTech Connect

    Beck, Micah; Moore, Terry

    2007-08-31

    As the flood of data associated with leading edge computational science continues to escalate, the challenge of supporting the distributed collaborations that are now characteristic of it becomes increasingly daunting. The chief obstacles to progress on this front lie less in the synchronous elements of collaboration, which have been reasonably well addressed by new global high performance networks, than in the asynchronous elements, where appropriate shared storage infrastructure seems to be lacking. The recent report from the Department of Energy on the emerging 'data management challenge' captures the multidimensional nature of this problem succinctly: Data inevitably needs to be buffered, for periods ranging from seconds to weeks, in order to be controlled as it moves through the distributed and collaborative research process. To meet the diverse and changing set of application needs that different research communities have, large amounts of non-archival storage are required for transitory buffering, and it needs to be widely dispersed, easily available, and configured to maximize flexibility of use. In today's grid fabric, however, massive storage is mostly concentrated in data centers, available only to those with user accounts and membership in the appropriate virtual organizations, allocated as if its usage were non-transitory, and encapsulated behind legacy interfaces that inhibit the flexibility of use and scheduling. This situation severely restricts the ability of application communities to access and schedule usable storage where and when they need to in order to make their workflow more productive. (p.69f) One possible strategy to deal with this problem lies in creating a storage infrastructure that can be universally shared because it provides only the most generic of asynchronous services. Different user communities then define higher level services as necessary to meet their needs. One model of such a service is a Storage Network, analogous to

  10. A scalable 2-D parallel sparse solver

    SciTech Connect

    Kothari, S.C.; Mitra, S.

    1995-12-01

    Scalability beyond a small number of processors, typically 32 or less, is known to be a problem for existing parallel general sparse (PGS) direct solvers. This paper presents a parallel general sparse PGS direct solver for general sparse linear systems on distributed memory machines. The algorithm is based on the well-known sequential sparse algorithm Y12M. To achieve efficient parallelization, a 2-D scattered decomposition of the sparse matrix is used. The proposed algorithm is more scalable than existing parallel sparse direct solvers. Its scalability is evaluated on a 256 processor nCUBE2s machine using Boeing/Harwell benchmark matrices.

  11. Scalable encryption using alpha rooting

    NASA Astrophysics Data System (ADS)

    Wharton, Eric J.; Panetta, Karen A.; Agaian, Sos S.

    2008-04-01

    Full and partial encryption methods are important for subscription based content providers, such as internet and cable TV pay channels. Providers need to be able to protect their products while at the same time being able to provide demonstrations to attract new customers without giving away the full value of the content. If an algorithm were introduced which could provide any level of full or partial encryption in a fast and cost effective manner, the applications to real-time commercial implementation would be numerous. In this paper, we present a novel application of alpha rooting, using it to achieve fast and straightforward scalable encryption with a single algorithm. We further present use of the measure of enhancement, the Logarithmic AME, to select optimal parameters for the partial encryption. When parameters are selected using the measure, the output image achieves a balance between protecting the important data in the image while still containing a good overall representation of the image. We will show results for this encryption method on a number of images, using histograms to evaluate the effectiveness of the encryption.

  12. Scalable Equation of State Capability

    SciTech Connect

    Epperly, T W; Fritsch, F N; Norquist, P D; Sanford, L A

    2007-12-03

    The purpose of this techbase project was to investigate the use of parallel array data types to reduce the memory footprint of the Livermore Equation Of State (LEOS) library. Addressing the memory scalability of LEOS is necessary to run large scientific simulations on IBM BG/L and future architectures with low memory per processing core. We considered using normal MPI, one-sided MPI, and Global Arrays to manage the distributed array and ended up choosing Global Arrays because it was the only communication library that provided the level of asynchronous access required. To reduce the runtime overhead using a parallel array data structure, a least recently used (LRU) caching algorithm was used to provide a local cache of commonly used parts of the parallel array. The approach was initially implemented in a isolated copy of LEOS and was later integrated into the main trunk of the LEOS Subversion repository. The approach was tested using a simple test. Testing indicated that the approach was feasible, and the simple LRU caching had a 86% hit rate.

  13. Generic algorithms for high performance scalable geocomputing

    NASA Astrophysics Data System (ADS)

    de Jong, Kor; Schmitz, Oliver; Karssenberg, Derek

    2016-04-01

    During the last decade, the characteristics of computing hardware have changed a lot. For example, instead of a single general purpose CPU core, personal computers nowadays contain multiple cores per CPU and often general purpose accelerators, like GPUs. Additionally, compute nodes are often grouped together to form clusters or a supercomputer, providing enormous amounts of compute power. For existing earth simulation models to be able to use modern hardware platforms, their compute intensive parts must be rewritten. This can be a major undertaking and may involve many technical challenges. Compute tasks must be distributed over CPU cores, offloaded to hardware accelerators, or distributed to different compute nodes. And ideally, all of this should be done in such a way that the compute task scales well with the hardware resources. This presents two challenges: 1) how to make good use of all the compute resources and 2) how to make these compute resources available for developers of simulation models, who may not (want to) have the required technical background for distributing compute tasks. The first challenge requires the use of specialized technology (e.g.: threads, OpenMP, MPI, OpenCL, CUDA). The second challenge requires the abstraction of the logic handling the distribution of compute tasks from the model-specific logic, hiding the technical details from the model developer. To assist the model developer, we are developing a C++ software library (called Fern) containing algorithms that can use all CPU cores available in a single compute node (distributing tasks over multiple compute nodes will be done at a later stage). The algorithms are grid-based (finite difference) and include local and spatial operations such as convolution filters. The algorithms handle distribution of the compute tasks to CPU cores internally. In the resulting model the low-level details of how this is done is separated from the model-specific logic representing the modeled system. This contrasts with practices in which code for distributing of compute tasks is mixed with model-specific code, and results in a better maintainable model. For flexibility and efficiency, the algorithms are configurable at compile-time with the respect to the following aspects: data type, value type, no-data handling, input value domain handling, and output value range handling. This makes the algorithms usable in very different contexts, without the need for making intrusive changes to existing models when using them. Applications that benefit from using the Fern library include the construction of forward simulation models in (global) hydrology (e.g. PCR-GLOBWB (Van Beek et al. 2011)), ecology, geomorphology, or land use change (e.g. PLUC (Verstegen et al. 2014)) and manipulation of hyper-resolution land surface data such as digital elevation models and remote sensing data. Using the Fern library, we have also created an add-on to the PCRaster Python Framework (Karssenberg et al. 2010) allowing its users to speed up their spatio-temporal models, sometimes by changing just a single line of Python code in their model. In our presentation we will give an overview of the design of the algorithms, providing examples of different contexts where they can be used to replace existing sequential algorithms, including the PCRaster environmental modeling software (www.pcraster.eu). We will show how the algorithms can be configured to behave differently when necessary. References Karssenberg, D., Schmitz, O., Salamon, P., De Jong, K. and Bierkens, M.F.P., 2010, A software framework for construction of process-based stochastic spatio-temporal models and data assimilation. Environmental Modelling & Software, 25, pp. 489-502, Link. Best Paper Award 2010: Software and Decision Support. Van Beek, L. P. H., Y. Wada, and M. F. P. Bierkens. 2011. Global monthly water stress: 1. Water balance and water availability. Water Resources Research. 47. Verstegen, J. A., D. Karssenberg, F. van der Hilst, and A. P. C. Faaij. 2014. Identifying a land use change cellular automaton by Bayesian data assimilation. Environmental Modelling & Software 53:121-136.

  14. Generic algorithms for high performance scalable geocomputing

    NASA Astrophysics Data System (ADS)

    de Jong, Kor; Schmitz, Oliver; Karssenberg, Derek

    2016-04-01

    During the last decade, the characteristics of computing hardware have changed a lot. For example, instead of a single general purpose CPU core, personal computers nowadays contain multiple cores per CPU and often general purpose accelerators, like GPUs. Additionally, compute nodes are often grouped together to form clusters or a supercomputer, providing enormous amounts of compute power. For existing earth simulation models to be able to use modern hardware platforms, their compute intensive parts must be rewritten. This can be a major undertaking and may involve many technical challenges. Compute tasks must be distributed over CPU cores, offloaded to hardware accelerators, or distributed to different compute nodes. And ideally, all of this should be done in such a way that the compute task scales well with the hardware resources. This presents two challenges: 1) how to make good use of all the compute resources and 2) how to make these compute resources available for developers of simulation models, who may not (want to) have the required technical background for distributing compute tasks. The first challenge requires the use of specialized technology (e.g.: threads, OpenMP, MPI, OpenCL, CUDA). The second challenge requires the abstraction of the logic handling the distribution of compute tasks from the model-specific logic, hiding the technical details from the model developer. To assist the model developer, we are developing a C++ software library (called Fern) containing algorithms that can use all CPU cores available in a single compute node (distributing tasks over multiple compute nodes will be done at a later stage). The algorithms are grid-based (finite difference) and include local and spatial operations such as convolution filters. The algorithms handle distribution of the compute tasks to CPU cores internally. In the resulting model the low-level details of how this is done is separated from the model-specific logic representing the modeled system. This contrasts with practices in which code for distributing of compute tasks is mixed with model-specific code, and results in a better maintainable model. For flexibility and efficiency, the algorithms are configurable at compile-time with the respect to the following aspects: data type, value type, no-data handling, input value domain handling, and output value range handling. This makes the algorithms usable in very different contexts, without the need for making intrusive changes to existing models when using them. Applications that benefit from using the Fern library include the construction of forward simulation models in (global) hydrology (e.g. PCR-GLOBWB (Van Beek et al. 2011)), ecology, geomorphology, or land use change (e.g. PLUC (Verstegen et al. 2014)) and manipulation of hyper-resolution land surface data such as digital elevation models and remote sensing data. Using the Fern library, we have also created an add-on to the PCRaster Python Framework (Karssenberg et al. 2010) allowing its users to speed up their spatio-temporal models, sometimes by changing just a single line of Python code in their model. In our presentation we will give an overview of the design of the algorithms, providing examples of different contexts where they can be used to replace existing sequential algorithms, including the PCRaster environmental modeling software (www.pcraster.eu). We will show how the algorithms can be configured to behave differently when necessary. References Karssenberg, D., Schmitz, O., Salamon, P., De Jong, K. and Bierkens, M.F.P., 2010, A software framework for construction of process-based stochastic spatio-temporal models and data assimilation. Environmental Modelling & Software, 25, pp. 489-502, Link. Best Paper Award 2010: Software and Decision Support. Van Beek, L. P. H., Y. Wada, and M. F. P. Bierkens. 2011. Global monthly water stress: 1. Water balance and water availability. Water Resources Research. 47. Verstegen, J. A., D. Karssenberg, F. van der Hilst, and A. P. C. Faaij. 2014. Identifying a land use change cellular automaton by Baye

  15. Dynamically scalable dual-core pipelined processor

    NASA Astrophysics Data System (ADS)

    Kumar, Nishant; Aggrawal, Ekta; Rajawat, Arvind

    2015-10-01

    This article proposes design and architecture of a dynamically scalable dual-core pipelined processor. Methodology of the design is the core fusion of two processors where two independent cores can dynamically morph into a larger processing unit, or they can be used as distinct processing elements to achieve high sequential performance and high parallel performance. Processor provides two execution modes. Mode1 is multiprogramming mode for execution of streams of instruction of lower data width, i.e., each core can perform 16-bit operations individually. Performance is improved in this mode due to the parallel execution of instructions in both the cores at the cost of area. In mode2, both the processing cores are coupled and behave like single, high data width processing unit, i.e., can perform 32-bit operation. Additional core-to-core communication is needed to realise this mode. The mode can switch dynamically; therefore, this processor can provide multifunction with single design. Design and verification of processor has been done successfully using Verilog on Xilinx 14.1 platform. The processor is verified in both simulation and synthesis with the help of test programs. This design aimed to be implemented on Xilinx Spartan 3E XC3S500E FPGA.

  16. Scalable Systems Software Enabling Technology Center

    SciTech Connect

    Michael T. Showerman

    2009-04-06

    NCSA’s role in the SCIDAC Scalable Systems Software (SSS) project was to develop interfaces and communication mechanisms for systems monitoring, and to implement a prototype demonstrating those standards. The Scalable Systems Monitoring component of the SSS suite was designed to provide a large volume of both static and dynamic systems data to the components within the SSS infrastructure as well as external data consumers.

  17. CMP: A memory-constrained scalability metric

    SciTech Connect

    Fienup, M.; Kothari, S.C.

    1995-12-01

    A scalability metric, called constant-memory-per-processor (CMP), is described for parallel architecture-algorithrn pairs. Its purpose is to predict the behavior of a specific algorithm on a distributed-memory machine as the number of processors is increased, but the memory per processor remains constant. While the CMP scalability metric predicts the asymptotic behavior, we show how to use it to predict expected performance on actual parallel machines, specifically the MasPar MP-I and MP-2.

  18. Scalable Total Synthesis of rac-Jungermannenones B and C.

    PubMed

    Liu, Weilong; Li, Houhua; Cai, Pei-Jun; Wang, Zhen; Yu, Zhi-Xiang; Lei, Xiaoguang

    2016-02-24

    Reported is the first scalable synthesis of rac-jungermannenones B and C starting from the commercially available and inexpensive geraniol in 10 and 9 steps, respectively. The unique jungermannenone framework is rapidly assembled by an unprecedented regioselective 1,6-dienyne reductive cyclization reaction which proceeds through a vinyl radical cyclization/allylic radical isomerization mechanism. DFT calculations explain the high regioselectivity observed in the 1,6-dienyne reductive radical cyclization. PMID:26823176

  19. Scalable Silicon Nanostructuring for Thermoelectric Applications

    NASA Astrophysics Data System (ADS)

    Koukharenko, E.; Boden, S. A.; Platzek, D.; Bagnall, D. M.; White, N. M.

    2013-07-01

    The current limitations of commercially available thermoelectric (TE) generators include their incompatibility with human body applications due to the toxicity of commonly used alloys and possible future shortage of raw materials (Bi-Sb-Te and Se). In this respect, exploiting silicon as an environmentally friendly candidate for thermoelectric applications is a promising alternative since it is an abundant, ecofriendly semiconductor for which there already exists an infrastructure for low-cost and high-yield processing. Contrary to the existing approaches, where n/ p-legs were either heavily doped to an optimal carrier concentration of 1019 cm-3 or morphologically modified by increasing their roughness, in this work improved thermoelectric performance was achieved in smooth silicon nanostructures with low doping concentration (1.5 × 1015 cm-3). Scalable, highly reproducible e-beam lithographies, which are compatible with nanoimprint and followed by deep reactive-ion etching (DRIE), were employed to produce arrays of regularly spaced nanopillars of 400 nm height with diameters varying from 140 nm to 300 nm. A potential Seebeck microprobe (PSM) was used to measure the Seebeck coefficients of such nanostructures. This resulted in values ranging from -75 μV/K to -120 μV/K for n-type and 100 μV/K to 140 μV/K for p-type, which are significant improvements over previously reported data.

  20. Scalable image coding for interactive image communication over networks

    NASA Astrophysics Data System (ADS)

    Yoon, Sung H.; Lee, Ji H.; Alexander, Winser E.

    2000-12-01

    This paper presents a new, scalable coding technique that can be used in interactive image/video communications over the Internet. The proposed technique generates a fully embedded bit stream that provides scalability with high quality for the whole image and it can be used to implement region based coding as well. The embedded bit stream is comprised of a basic layer and many enhancement layers. The enhancement layers add refinement to the quality of the image that has been reconstructed using the basic layer. The proposed coding technique uses multiple quantizers with thresholds (QT) for layering and it creates a bit plane for each layer. The bit plane is then partitioned into sets of small areas to be coded independently. Run length and entropy coding are applied to each of the sets to provide scalability for the entire image resulting in high picture quality in the user-specific area of interest (ROI). We tested this technique by applying it to various test images and the results consistently show high level of performance.

  1. Scalable, full-colour and controllable chromotropic plasmonic printing.

    PubMed

    Xue, Jiancai; Zhou, Zhang-Kai; Wei, Zhiqiang; Su, Rongbin; Lai, Juan; Li, Juntao; Li, Chao; Zhang, Tengwei; Wang, Xue-Hua

    2015-01-01

    Plasmonic colour printing has drawn wide attention as a promising candidate for the next-generation colour-printing technology. However, an efficient approach to realize full colour and scalable fabrication is still lacking, which prevents plasmonic colour printing from practical applications. Here we present a scalable and full-colour plasmonic printing approach by combining conjugate twin-phase modulation with a plasmonic broadband absorber. More importantly, our approach also demonstrates controllable chromotropic capability, that is, the ability of reversible colour transformations. This chromotropic capability affords enormous potentials in building functionalized prints for anticounterfeiting, special label, and high-density data encryption storage. With such excellent performances in functional colour applications, this colour-printing approach could pave the way for plasmonic colour printing in real-world commercial utilization. PMID:26567803

  2. Scalable Parallel Distance Field Construction for Large-Scale Applications.

    PubMed

    Yu, Hongfeng; Xie, Jinrong; Ma, Kwan-Liu; Kolla, Hemanth; Chen, Jacqueline H

    2015-10-01

    Computing distance fields is fundamental to many scientific and engineering applications. Distance fields can be used to direct analysis and reduce data. In this paper, we present a highly scalable method for computing 3D distance fields on massively parallel distributed-memory machines. A new distributed spatial data structure, named parallel distance tree, is introduced to manage the level sets of data and facilitate surface tracking over time, resulting in significantly reduced computation and communication costs for calculating the distance to the surface of interest from any spatial locations. Our method supports several data types and distance metrics from real-world applications. We demonstrate its efficiency and scalability on state-of-the-art supercomputers using both large-scale volume datasets and surface models. We also demonstrate in-situ distance field computation on dynamic turbulent flame surfaces for a petascale combustion simulation. Our work greatly extends the usability of distance fields for demanding applications. PMID:26357251

  3. Scalable parallel distance field construction for large-scale applications

    SciTech Connect

    Yu, Hongfeng; Xie, Jinrong; Ma, Kwan -Liu; Kolla, Hemanth; Chen, Jacqueline H.

    2015-10-01

    Computing distance fields is fundamental to many scientific and engineering applications. Distance fields can be used to direct analysis and reduce data. In this paper, we present a highly scalable method for computing 3D distance fields on massively parallel distributed-memory machines. Anew distributed spatial data structure, named parallel distance tree, is introduced to manage the level sets of data and facilitate surface tracking overtime, resulting in significantly reduced computation and communication costs for calculating the distance to the surface of interest from any spatial locations. Our method supports several data types and distance metrics from real-world applications. We demonstrate its efficiency and scalability on state-of-the-art supercomputers using both large-scale volume datasets and surface models. We also demonstrate in-situ distance field computation on dynamic turbulent flame surfaces for a petascale combustion simulation. In conclusion, our work greatly extends the usability of distance fields for demanding applications.

  4. Scalable, full-colour and controllable chromotropic plasmonic printing

    PubMed Central

    Xue, Jiancai; Zhou, Zhang-Kai; Wei, Zhiqiang; Su, Rongbin; Lai, Juan; Li, Juntao; Li, Chao; Zhang, Tengwei; Wang, Xue-Hua

    2015-01-01

    Plasmonic colour printing has drawn wide attention as a promising candidate for the next-generation colour-printing technology. However, an efficient approach to realize full colour and scalable fabrication is still lacking, which prevents plasmonic colour printing from practical applications. Here we present a scalable and full-colour plasmonic printing approach by combining conjugate twin-phase modulation with a plasmonic broadband absorber. More importantly, our approach also demonstrates controllable chromotropic capability, that is, the ability of reversible colour transformations. This chromotropic capability affords enormous potentials in building functionalized prints for anticounterfeiting, special label, and high-density data encryption storage. With such excellent performances in functional colour applications, this colour-printing approach could pave the way for plasmonic colour printing in real-world commercial utilization. PMID:26567803

  5. Scalable office-based health care

    PubMed Central

    Koepp, Gabriel A.; Manohar, Chinmay U.; McCrady-Spitzer, Shelly K.; Levine, James A.

    2014-01-01

    The goal of healthcare is to provide high quality care at an affordable cost for its patients. However, the population it serves has changed dramatically since the popularization of hospital-based healthcare. With available new technology, alternative healthcare delivery methods can be designed and tested. This study examines Scalable Office Based Healthcare for Small Business, where healthcare is delivered to the office floor. This delivery was tested in 18 individuals at a small business in Minneapolis, Minnesota. The goal was to deliver modular healthcare and mitigate conditions such as diabetes, hyperlipidemia, obesity, sedentariness, and metabolic disease. The modular healthcare system was welcomed by employees – 70% of those eligible enrolled. The findings showed that the modular healthcare deliverable was feasible and effective. The data demonstrated significant improvements in weight loss, fat loss, and blood variables for at risk participants. This study leaves room for improvement and further innovation. Expansion to include offerings such as physicals, diabetes management, smoking cessation, and pre-natal treatment would improve its utility. Future studies could include testing the adaptability of delivery method, as it should adapt to reach rural and underserved populations. PMID:21471576

  6. Scalable Production of Molybdenum Disulfide Based Biosensors.

    PubMed

    Naylor, Carl H; Kybert, Nicholas J; Schneier, Camilla; Xi, Jin; Romero, Gabriela; Saven, Jeffery G; Liu, Renyu; Johnson, A T Charlie

    2016-06-28

    We demonstrate arrays of opioid biosensors based on chemical vapor deposition grown molybdenum disulfide (MoS2) field effect transistors (FETs) coupled to a computationally redesigned, water-soluble variant of the μ-opioid receptor (MOR). By transferring dense films of monolayer MoS2 crystals onto prefabricated electrode arrays, we obtain high-quality FETs with clean surfaces that allow for reproducible protein attachment. The fabrication yield of MoS2 FETs and biosensors exceeds 95%, with an average mobility of 2.0 cm(2) V(-1) s(-1) (36 cm(2) V(-1) s(-1)) at room temperature under ambient (in vacuo). An atomic length nickel-mediated linker chemistry enables target binding events that occur very close to the MoS2 surface to maximize sensitivity. The biosensor response calibration curve for a synthetic opioid peptide known to bind to the wild-type MOR indicates binding affinity that matches values determined using traditional techniques and a limit of detection ∼3 nM (1.5 ng/mL). The combination of scalable array fabrication and rapid, precise binding readout enabled by the MoS2 transistor offers the prospect of a solid-state drug testing platform for rapid readout of the interactions between novel drugs and their intended protein targets. PMID:27227361

  7. Lightweight and scalable secure communication in VANET

    NASA Astrophysics Data System (ADS)

    Zhu, Xiaoling; Lu, Yang; Zhu, Xiaojuan; Qiu, Shuwei

    2015-05-01

    To avoid a message to be tempered and forged in vehicular ad hoc network (VANET), the digital signature method is adopted by IEEE1609.2. However, the costs of the method are excessively high for large-scale networks. The paper efficiently copes with the issue with a secure communication framework by introducing some lightweight cryptography primitives. In our framework, point-to-point and broadcast communications for vehicle-to-infrastructure (V2I) and vehicle-to-vehicle (V2V) are studied, mainly based on symmetric cryptography. A new issue incurred is symmetric key management. Thus, we develop key distribution and agreement protocols for two-party key and group key under different environments, whether a road side unit (RSU) is deployed or not. The analysis shows that our protocols provide confidentiality, authentication, perfect forward secrecy, forward secrecy and backward secrecy. The proposed group key agreement protocol especially solves the key leak problem caused by members joining or leaving in existing key agreement protocols. Due to aggregated signature and substitution of XOR for point addition, the average computation and communication costs do not significantly increase with the increase in the number of vehicles; hence, our framework provides good scalability.

  8. Towards Scalable Optimal Sequence Homology Detection

    SciTech Connect

    Daily, Jeffrey A.; Krishnamoorthy, Sriram; Kalyanaraman, Anantharaman

    2012-12-26

    Abstract—The field of bioinformatics and computational biol- ogy is experiencing a data revolution — experimental techniques to procure data have increased in throughput, improved in accuracy and reduced in costs. This has spurred an array of high profile sequencing and data generation projects. While the data repositories represent untapped reservoirs of rich information critical for scientific breakthroughs, the analytical software tools that are needed to analyze large volumes of such sequence data have significantly lagged behind in their capacity to scale. In this paper, we address homology detection, which is a funda- mental problem in large-scale sequence analysis with numerous applications. We present a scalable framework to conduct large- scale optimal homology detection on massively parallel super- computing platforms. Our approach employs distributed memory work stealing to effectively parallelize optimal pairwise alignment computation tasks. Results on 120,000 cores of the Hopper Cray XE6 supercomputer demonstrate strong scaling and up to 2.42 × 107 optimal pairwise sequence alignments computed per second (PSAPS), the highest reported in the literature.

  9. Scalable Combinatorial Tools for Health Disparities Research

    PubMed Central

    Langston, Michael A.; Levine, Robert S.; Kilbourne, Barbara J.; Rogers, Gary L.; Kershenbaum, Anne D.; Baktash, Suzanne H.; Coughlin, Steven S.; Saxton, Arnold M.; Agboto, Vincent K.; Hood, Darryl B.; Litchveld, Maureen Y.; Oyana, Tonny J.; Matthews-Juarez, Patricia; Juarez, Paul D.

    2014-01-01

    Despite staggering investments made in unraveling the human genome, current estimates suggest that as much as 90% of the variance in cancer and chronic diseases can be attributed to factors outside an individual’s genetic endowment, particularly to environmental exposures experienced across his or her life course. New analytical approaches are clearly required as investigators turn to complicated systems theory and ecological, place-based and life-history perspectives in order to understand more clearly the relationships between social determinants, environmental exposures and health disparities. While traditional data analysis techniques remain foundational to health disparities research, they are easily overwhelmed by the ever-increasing size and heterogeneity of available data needed to illuminate latent gene x environment interactions. This has prompted the adaptation and application of scalable combinatorial methods, many from genome science research, to the study of population health. Most of these powerful tools are algorithmically sophisticated, highly automated and mathematically abstract. Their utility motivates the main theme of this paper, which is to describe real applications of innovative transdisciplinary models and analyses in an effort to help move the research community closer toward identifying the causal mechanisms and associated environmental contexts underlying health disparities. The public health exposome is used as a contemporary focus for addressing the complex nature of this subject. PMID:25310540

  10. Scalable multichannel MRI data acquisition system.

    PubMed

    Bodurka, Jerzy; Ledden, Patrick J; van Gelderen, Peter; Chu, Renxin; de Zwart, Jacco A; Morris, Doug; Duyn, Jeff H

    2004-01-01

    A scalable multichannel digital MRI receiver system was designed to achieve high bandwidth echo-planar imaging (EPI) acquisitions for applications such as BOLD-fMRI. The modular system design allows for easy extension to an arbitrary number of channels. A 16-channel receiver was developed and integrated with a General Electric (GE) Signa 3T VH/3 clinical scanner. Receiver performance was evaluated on phantoms and human volunteers using a custom-built 16-element receive-only brain surface coil array. At an output bandwidth of 1 MHz, a 100% acquisition duty cycle was achieved. Overall system noise figure and dynamic range were better than 0.85 dB and 84 dB, respectively. During repetitive EPI scanning on phantoms, the relative temporal standard deviation of the image intensity time-course was below 0.2%. As compared to the product birdcage head coil, 16-channel reception with the custom array yielded a nearly 6-fold SNR gain in the cerebral cortex and a 1.8-fold SNR gain in the center of the brain. The excellent system stability combined with the increased sensitivity and SENSE capabilities of 16-channel coils are expected to significantly benefit and enhance fMRI applications. PMID:14705057

  11. A scalable neuristor built with Mott memristors

    NASA Astrophysics Data System (ADS)

    Pickett, Matthew D.; Medeiros-Ribeiro, Gilberto; Williams, R. Stanley

    2013-02-01

    The Hodgkin-Huxley model for action potential generation in biological axons is central for understanding the computational capability of the nervous system and emulating its functionality. Owing to the historical success of silicon complementary metal-oxide-semiconductors, spike-based computing is primarily confined to software simulations and specialized analogue metal-oxide-semiconductor field-effect transistor circuits. However, there is interest in constructing physical systems that emulate biological functionality more directly, with the goal of improving efficiency and scale. The neuristor was proposed as an electronic device with properties similar to the Hodgkin-Huxley axon, but previous implementations were not scalable. Here we demonstrate a neuristor built using two nanoscale Mott memristors, dynamical devices that exhibit transient memory and negative differential resistance arising from an insulating-to-conducting phase transition driven by Joule heating. This neuristor exhibits the important neural functions of all-or-nothing spiking with signal gain and diverse periodic spiking, using materials and structures that are amenable to extremely high-density integration with or without silicon transistors.

  12. A scalable neuristor built with Mott memristors.

    PubMed

    Pickett, Matthew D; Medeiros-Ribeiro, Gilberto; Williams, R Stanley

    2013-02-01

    The Hodgkin-Huxley model for action potential generation in biological axons is central for understanding the computational capability of the nervous system and emulating its functionality. Owing to the historical success of silicon complementary metal-oxide-semiconductors, spike-based computing is primarily confined to software simulations and specialized analogue metal-oxide-semiconductor field-effect transistor circuits. However, there is interest in constructing physical systems that emulate biological functionality more directly, with the goal of improving efficiency and scale. The neuristor was proposed as an electronic device with properties similar to the Hodgkin-Huxley axon, but previous implementations were not scalable. Here we demonstrate a neuristor built using two nanoscale Mott memristors, dynamical devices that exhibit transient memory and negative differential resistance arising from an insulating-to-conducting phase transition driven by Joule heating. This neuristor exhibits the important neural functions of all-or-nothing spiking with signal gain and diverse periodic spiking, using materials and structures that are amenable to extremely high-density integration with or without silicon transistors. PMID:23241533

  13. Towards scalable electronic structure calculations for alloys

    SciTech Connect

    Stocks, G.M.; Nicholson, D.M.C.; Wang, Y.; Shelton, W.A.; Szotek, Z.; Temmermann, W.M.

    1994-06-01

    A new approach to calculating the properties of large systems within the local density approximation (LDA) that offers the promise of scalability on massively parallel supercomputers is outlined. The electronic structure problem is formulated in real space using multiple scattering theory. The standard LDA algorithm is divided into two parts. Firstly, finding the self-consistent field (SCF) electron density, Secondly, calculating the energy corresponding to the SCF density. We show, at least for metals and alloys, that the former problem is easily solved using real space methods. For the second we take advantage of the variational properties of a generalized Harris-Foulkes free energy functional, a new conduction band Fermi function, and a fictitious finite electron temperature that again allow us to use real-space methods. Using a compute-node {R_arrow} atom equivalence the new method is naturally highly parallel and leads to O(N) scaling where N is the number of atoms making up the system. We show scaling data gathered on the Intel XP/S 35 Paragon for systems up to 512-atoms/simulation cell. To demonstrate that we can achieve metallurgical-precision, we apply the new method to the calculation the energies of disordered CuO{sub 0.5}Zn{sub 0.5} alloys using a large random sample.

  14. Wanted: Scalable Tracers for Diffusion Measurements

    PubMed Central

    2015-01-01

    Scalable tracers are potentially a useful tool to examine diffusion mechanisms and to predict diffusion coefficients, particularly for hindered diffusion in complex, heterogeneous, or crowded systems. Scalable tracers are defined as a series of tracers varying in size but with the same shape, structure, surface chemistry, deformability, and diffusion mechanism. Both chemical homology and constant dynamics are required. In particular, branching must not vary with size, and there must be no transition between ordinary diffusion and reptation. Measurements using scalable tracers yield the mean diffusion coefficient as a function of size alone; measurements using nonscalable tracers yield the variation due to differences in the other properties. Candidate scalable tracers are discussed for two-dimensional (2D) diffusion in membranes and three-dimensional diffusion in aqueous solutions. Correlations to predict the mean diffusion coefficient of globular biomolecules from molecular mass are reviewed briefly. Specific suggestions for the 3D case include the use of synthetic dendrimers or random hyperbranched polymers instead of dextran and the use of core–shell quantum dots. Another useful tool would be a series of scalable tracers varying in deformability alone, prepared by varying the density of crosslinking in a polymer to make say “reinforced Ficoll” or “reinforced hyperbranched polyglycerol.” PMID:25319586

  15. SuperLU{_}DIST: A scalable distributed-memory sparse direct solver for unsymmetric linear systems

    SciTech Connect

    Li, Xiaoye S.; Demmel, James W.

    2002-03-27

    In this paper, we present the main algorithmic features in the software package SuperLU{_}DIST, a distributed-memory sparse direct solver for large sets of linear equations. We give in detail our parallelization strategies, with focus on scalability issues, and demonstrate the parallel performance and scalability on current machines. The solver is based on sparse Gaussian elimination, with an innovative static pivoting strategy proposed earlier by the authors. The main advantage of static pivoting over classical partial pivoting is that it permits a priori determination of data structures and communication pattern for sparse Gaussian elimination, which makes it more scalable on distributed memory machines. Based on this a priori knowledge, we designed highly parallel and scalable algorithms for both LU decomposition and triangular solve and we show that they are suitable for large-scale distributed memory machines.

  16. Lilith: A software framework for the rapid development of scalable tools for distributed computing

    SciTech Connect

    Gentile, A.C.; Evensky, D.A.; Armstrong, R.C.

    1998-03-01

    Lilith is a general purpose framework, written in Java, that provides a highly scalable distribution of user code across a heterogeneous computing platform. By creation of suitable user code, the Lilith framework can be used for tool development. The scalable performance provided by Lilith is crucial to the development of effective tools for large distributed systems. Furthermore, since Lilith handles the details of code distribution and communication, the user code need focus primarily on the tool functionality, thus, greatly decreasing the time required for tool development. In this paper, the authors concentrate on the use of the Lilith framework to develop scalable tools. The authors review the functionality of Lilith and introduce a typical tool capitalizing on the features of the framework. They present new Objects directly involved with tool creation. They explain details of development and illustrate with an example. They present timing results demonstrating scalability.

  17. A Robust Scalable Transportation System Concept

    NASA Technical Reports Server (NTRS)

    Hahn, Andrew; DeLaurentis, Daniel

    2006-01-01

    This report documents the 2005 Revolutionary System Concept for Aeronautics (RSCA) study entitled "A Robust, Scalable Transportation System Concept". The objective of the study was to generate, at a high-level of abstraction, characteristics of a new concept for the National Airspace System, or the new NAS, under which transportation goals such as increased throughput, delay reduction, and improved robustness could be realized. Since such an objective can be overwhelmingly complex if pursued at the lowest levels of detail, instead a System-of-Systems (SoS) approach was adopted to model alternative air transportation architectures at a high level. The SoS approach allows the consideration of not only the technical aspects of the NAS", but also incorporates policy, socio-economic, and alternative transportation system considerations into one architecture. While the representations of the individual systems are basic, the higher level approach allows for ways to optimize the SoS at the network level, determining the best topology (i.e. configuration of nodes and links). The final product (concept) is a set of rules of behavior and network structure that not only satisfies national transportation goals, but represents the high impact rules that accomplish those goals by getting the agents to "do the right thing" naturally. The novel combination of Agent Based Modeling and Network Theory provides the core analysis methodology in the System-of-Systems approach. Our method of approach is non-deterministic which means, fundamentally, it asks and answers different questions than deterministic models. The nondeterministic method is necessary primarily due to our marriage of human systems with technological ones in a partially unknown set of future worlds. Our goal is to understand and simulate how the SoS, human and technological components combined, evolve.

  18. Space Situational Awareness Data Processing Scalability Utilizing Google Cloud Services

    NASA Astrophysics Data System (ADS)

    Greenly, D.; Duncan, M.; Wysack, J.; Flores, F.

    Space Situational Awareness (SSA) is a fundamental and critical component of current space operations. The term SSA encompasses the awareness, understanding and predictability of all objects in space. As the population of orbital space objects and debris increases, the number of collision avoidance maneuvers grows and prompts the need for accurate and timely process measures. The SSA mission continually evolves to near real-time assessment and analysis demanding the need for higher processing capabilities. By conventional methods, meeting these demands requires the integration of new hardware to keep pace with the growing complexity of maneuver planning algorithms. SpaceNav has implemented a highly scalable architecture that will track satellites and debris by utilizing powerful virtual machines on the Google Cloud Platform. SpaceNav algorithms for processing CDMs outpace conventional means. A robust processing environment for tracking data, collision avoidance maneuvers and various other aspects of SSA can be created and deleted on demand. Migrating SpaceNav tools and algorithms into the Google Cloud Platform will be discussed and the trials and tribulations involved. Information will be shared on how and why certain cloud products were used as well as integration techniques that were implemented. Key items to be presented are: 1.Scientific algorithms and SpaceNav tools integrated into a scalable architecture a) Maneuver Planning b) Parallel Processing c) Monte Carlo Simulations d) Optimization Algorithms e) SW Application Development/Integration into the Google Cloud Platform 2. Compute Engine Processing a) Application Engine Automated Processing b) Performance testing and Performance Scalability c) Cloud MySQL databases and Database Scalability d) Cloud Data Storage e) Redundancy and Availability

  19. Scalable fault tolerant image communication and storage grid

    NASA Astrophysics Data System (ADS)

    Slik, David; Seiler, Oliver; Altman, Tym; Montour, Mike; Kermani, Mohammad; Proseilo, Walter; Terry, David; Kawahara, Midori; Leckie, Chris; Muir, Dale

    2003-05-01

    Increasing production and use of digital medical imagery are driving new approaches to information storage and management. Traditional, centralized approaches to image communication, storage and archiving are becoming increasingly expensive to scale and operate with high levels of reliability. Multi-site, geographically-distributed deployments connected by limited-bandwidth networks present further scalability, reliability, and availability challenges. A grid storage architecture built from a distributed network of low cost, off-the-shelf servers (nodes) provides scalable data and metadata storage, processing, and communication without single points of failure. Imaging studies are stored, replicated, cached, managed, and retrieved based on defined rules, and nodes within the grid can acquire studies and respond to queries. Grid nodes transparently load-balance queries, storage/retrieval requests, and replicate data for automated backup and disaster recovery. This approach reduces latency, increases availability, provides near-linear scalability and allows the creation of a geographically distributed medical imaging network infrastructure. This paper presents some key concepts in grid storage and discusses the results of a clinical deployment of a multi-site storage grid for cancer care in the province of British Columbia.

  20. The intergroup protocols: Scalable group communication for the internet

    SciTech Connect

    Berket, K.

    2000-11-01

    Reliable group ordered delivery of multicast messages in a distributed system is a useful service that simplifies the programming of distributed applications. Such a service helps to maintain the consistency of replicated information and to coordinate the activities of the various processes. With the increasing popularity of the Internet, there is an increasing interest in scaling the protocols that provide this service to the environment of the Internet. The InterGroup protocol suite, described in this dissertation, provides such a service, and is intended for the environment of the Internet with scalability to large numbers of nodes and high latency links. The InterGroup protocols approach the scalability problem from various directions. They redefine the meaning of group membership, allow voluntary membership changes, add a receiver-oriented selection of delivery guarantees that permits heterogeneity of the receiver set, and provide a scalable reliability service. The InterGroup system comprises several components, executing at various sites within the system. Each component provides part of the services necessary to implement a group communication system for the wide-area. The components can be categorized as: (1) control hierarchy, (2) reliable multicast, (3) message distribution and delivery, and (4) process group membership. We have implemented a prototype of the InterGroup protocols in Java, and have tested the system performance in both local-area and wide-area networks.

  1. Superlinearly scalable noise robustness of redundant coupled dynamical systems

    NASA Astrophysics Data System (ADS)

    Kohar, Vivek; Kia, Behnam; Lindner, John F.; Ditto, William L.

    2016-03-01

    We illustrate through theory and numerical simulations that redundant coupled dynamical systems can be extremely robust against local noise in comparison to uncoupled dynamical systems evolving in the same noisy environment. Previous studies have shown that the noise robustness of redundant coupled dynamical systems is linearly scalable and deviations due to noise can be minimized by increasing the number of coupled units. Here, we demonstrate that the noise robustness can actually be scaled superlinearly if some conditions are met and very high noise robustness can be realized with very few coupled units. We discuss these conditions and show that this superlinear scalability depends on the nonlinearity of the individual dynamical units. The phenomenon is demonstrated in discrete as well as continuous dynamical systems. This superlinear scalability not only provides us an opportunity to exploit the nonlinearity of physical systems without being bogged down by noise but may also help us in understanding the functional role of coupled redundancy found in many biological systems. Moreover, engineers can exploit superlinear noise suppression by starting a coupled system near (not necessarily at) the appropriate initial condition.

  2. An overview on scalable encryption for wireless multimedia access

    NASA Astrophysics Data System (ADS)

    Yu, Hong Heather

    2003-08-01

    Wireless environments present many challenges for secure multimedia access, especial streaming media. The availability of varying network bandwidths and diverse receiver device processing powers and storage spaces demand scalable and flexible approaches that are capable of adapting to changing network conditions as well as device capabilities. To meet these requirements, scalable and fine granularity scalable (FGS) compression algorithms were proposed and widely adopted to provide scalable access of multimedia with interoperability between different services and flexible support to receivers with different device capabilities. Encryption is one of the most important security tools to protect content from unauthorized use. If a medium data stream is encrypted using non-scalable cryptography algorithms, decryption at arbitrary bit rate to provide scalable services can hardly be accomplished. If a medium compressed using scalable coding needs to be protected and non-scalable cryptography algorithms are used, the advantages of scalable coding may be lost. Therefore scalable encryption techniques are needed to provide scalability or to preserve the FGS adaptation capability (if the media stream is FGS coded) and enable intermediate processing of encrypted data without unnecessary decryption. In this paper, we will give an overview of scalable encryption schemes and present a fine grained scalable encryption algorithm. One desirable feature is its simplicity and flexibility in supporting scalable multimedia communication and multimedia content access control in wireless environments.

  3. An Efficient, Scalable Content-Based Messaging System

    SciTech Connect

    Gorton, Ian; Almquist, Justin P.; Cramer, Nick O.; Haack, Jereme N.; Hoza, Mark

    2003-09-16

    Large-scale information processing environments must rapidly search through massive streams of raw data to locate useful information. These data streams contain textual and numeric data items, and may be highly structured or mostly freeform text. This project aims to create a high performance and scalable engine for locating relevant content in data streams. Based on the J2EE Java Messaging Service (JMS), the content-based messaging (CBM) engine provides highly efficient message formatting and filtering. This paper describes the design of the CBM engine, and presents empirical results that compare the performance with a standard JMS to demonstrate the performance improvements that are achieved.

  4. Area scalable optically induced photorefractive photonic microstructures

    NASA Astrophysics Data System (ADS)

    Jin, Wentao; Xue, Yan Ling; Jiang, Dongdong

    2016-07-01

    A convenient approach to fabricate area scalable two-dimensional photonic microstructures was experimentally demonstrated by multi-face optical wedges. The approach is quite compact and stable without complex optical alignment equipment. Large-area square lattice microstructures are optically induced inside an iron-doped lithium niobate photorefractive crystal. The induced large-area microstructures are analyzed and verified by plane wave guiding, Brillouin-zone spectroscopy, angle-dependent transmission spectrum, and lateral Bragg reflection patterns. The method can be easily extended to generate other more complex area scalable photonic microstructures, such as quasicrystal lattices, by designing the multi-face optical wedge appropriately. The induced area scalable photonic microstructures can be fixed or erased even re-recorded in the photorefractive crystal, which suggests potential applications in micro-nano photonic devices.

  5. Scalable k-means statistics with Titan.

    SciTech Connect

    Thompson, David C.; Bennett, Janine C.; Pebay, Philippe Pierre

    2009-11-01

    This report summarizes existing statistical engines in VTK/Titan and presents both the serial and parallel k-means statistics engines. It is a sequel to [PT08], [BPRT09], and [PT09] which studied the parallel descriptive, correlative, multi-correlative, principal component analysis, and contingency engines. The ease of use of the new parallel k-means engine is illustrated by the means of C++ code snippets and algorithm verification is provided. This report justifies the design of the statistics engines with parallel scalability in mind, and provides scalability and speed-up analysis results for the k-means engine.

  6. Validation of a Scalable Solar Sailcraft

    NASA Technical Reports Server (NTRS)

    Murphy, D. M.

    2006-01-01

    The NASA In-Space Propulsion (ISP) program sponsored intensive solar sail technology and systems design, development, and hardware demonstration activities over the past 3 years. Efforts to validate a scalable solar sail system by functional demonstration in relevant environments, together with test-analysis correlation activities on a scalable solar sail system have recently been successfully completed. A review of the program, with descriptions of the design, results of testing, and analytical model validations of component and assembly functional, strength, stiffness, shape, and dynamic behavior are discussed. The scaled performance of the validated system is projected to demonstrate the applicability to flight demonstration and important NASA road-map missions.

  7. Parallel Heuristics for Scalable Community Detection

    SciTech Connect

    Lu, Howard; Kalyanaraman, Anantharaman; Halappanavar, Mahantesh; Choudhury, Sutanay

    2014-05-17

    Community detection has become a fundamental operation in numerous graph-theoretic applications. It is used to reveal natural divisions that exist within real world networks without imposing prior size or cardinality constraints on the set of communities. Despite its potential for application, there is only limited support for community detection on large-scale parallel computers, largely owing to the irregular and inherently sequential nature of the underlying heuristics. In this paper, we present parallelization heuristics for fast community detection using the Louvain method as the serial template. The Louvain method is an iterative heuristic for modularity optimization. Originally developed by Blondel et al. in 2008, the method has become increasingly popular owing to its ability to detect high modularity community partitions in a fast and memory-efficient manner. However, the method is also inherently sequential, thereby limiting its scalability to problems that can be solved on desktops. Here, we observe certain key properties of this method that present challenges for its parallelization, and consequently propose multiple heuristics that are designed to break the sequential barrier. Our heuristics are agnostic to the underlying parallel architecture. For evaluation purposes, we implemented our heuristics on shared memory (OpenMP) and distributed memory (MapReduce-MPI) machines, and tested them over real world graphs derived from multiple application domains (internet, biological, natural language processing). Experimental results demonstrate the ability of our heuristics to converge to high modularity solutions comparable to those output by the serial algorithm in nearly the same number of iterations, while also drastically reducing time to solution.

  8. Scalability of Localized Arc Filament Plasma Actuators

    NASA Technical Reports Server (NTRS)

    Brown, Clifford A.

    2008-01-01

    Temporal flow control of a jet has been widely studied in the past to enhance jet mixing or reduce jet noise. Most of this research, however, has been done using small diameter low Reynolds number jets that often have little resemblance to the much larger jets common in real world applications because the flow actuators available lacked either the power or bandwidth to sufficiently impact these larger higher energy jets. The Localized Arc Filament Plasma Actuators (LAFPA), developed at the Ohio State University (OSU), have demonstrated the ability to impact a small high speed jet in experiments conducted at OSU and the power to perturb a larger high Reynolds number jet in experiments conducted at the NASA Glenn Research Center. However, the response measured in the large-scale experiments was significantly reduced for the same number of actuators compared to the jet response found in the small-scale experiments. A computational study has been initiated to simulate the LAFPA system with additional actuators on a large-scale jet to determine the number of actuators required to achieve the same desired response for a given jet diameter. Central to this computational study is a model for the LAFPA that both accurately represents the physics of the actuator and can be implemented into a computational fluid dynamics solver. One possible model, based on pressure waves created by the rapid localized heating that occurs at the actuator, is investigated using simplified axisymmetric simulations. The results of these simulations will be used to determine the validity of the model before more realistic and time consuming three-dimensional simulations are conducted to ultimately determine the scalability of the LAFPA system.

  9. Scalable microreactors and methods for using same

    DOEpatents

    Lawal, Adeniyi; Qian, Dongying

    2010-03-02

    The present invention provides a scalable microreactor comprising a multilayered reaction block having alternating reaction plates and heat exchanger plates that have a plurality of microchannels; a multilaminated reactor input manifold, a collecting reactor output manifold, a heat exchange input manifold and a heat exchange output manifold. The present invention also provides methods of using the microreactor for multiphase chemical reactions.

  10. Scalable Domain Decomposed Monte Carlo Particle Transport

    SciTech Connect

    O'Brien, Matthew Joseph

    2013-12-05

    In this dissertation, we present the parallel algorithms necessary to run domain decomposed Monte Carlo particle transport on large numbers of processors (millions of processors). Previous algorithms were not scalable, and the parallel overhead became more computationally costly than the numerical simulation.

  11. Physical principles for scalable neural recording.

    PubMed

    Marblestone, Adam H; Zamft, Bradley M; Maguire, Yael G; Shapiro, Mikhail G; Cybulski, Thaddeus R; Glaser, Joshua I; Amodei, Dario; Stranges, P Benjamin; Kalhor, Reza; Dalrymple, David A; Seo, Dongjin; Alon, Elad; Maharbiz, Michel M; Carmena, Jose M; Rabaey, Jan M; Boyden, Edward S; Church, George M; Kording, Konrad P

    2013-01-01

    Simultaneously measuring the activities of all neurons in a mammalian brain at millisecond resolution is a challenge beyond the limits of existing techniques in neuroscience. Entirely new approaches may be required, motivating an analysis of the fundamental physical constraints on the problem. We outline the physical principles governing brain activity mapping using optical, electrical, magnetic resonance, and molecular modalities of neural recording. Focusing on the mouse brain, we analyze the scalability of each method, concentrating on the limitations imposed by spatiotemporal resolution, energy dissipation, and volume displacement. Based on this analysis, all existing approaches require orders of magnitude improvement in key parameters. Electrical recording is limited by the low multiplexing capacity of electrodes and their lack of intrinsic spatial resolution, optical methods are constrained by the scattering of visible light in brain tissue, magnetic resonance is hindered by the diffusion and relaxation timescales of water protons, and the implementation of molecular recording is complicated by the stochastic kinetics of enzymes. Understanding the physical limits of brain activity mapping may provide insight into opportunities for novel solutions. For example, unconventional methods for delivering electrodes may enable unprecedented numbers of recording sites, embedded optical devices could allow optical detectors to be placed within a few scattering lengths of the measured neurons, and new classes of molecularly engineered sensors might obviate cumbersome hardware architectures. We also study the physics of powering and communicating with microscale devices embedded in brain tissue and find that, while radio-frequency electromagnetic data transmission suffers from a severe power-bandwidth tradeoff, communication via infrared light or ultrasound may allow high data rates due to the possibility of spatial multiplexing. The use of embedded local recording and

  12. Scalable Machine Learning for Massive Astronomical Datasets

    NASA Astrophysics Data System (ADS)

    Ball, Nicholas M.; Gray, A.

    2014-04-01

    We present the ability to perform data mining and machine learning operations on a catalog of half a billion astronomical objects. This is the result of the combination of robust, highly accurate machine learning algorithms with linear scalability that renders the applications of these algorithms to massive astronomical data tractable. We demonstrate the core algorithms kernel density estimation, K-means clustering, linear regression, nearest neighbors, random forest and gradient-boosted decision tree, singular value decomposition, support vector machine, and two-point correlation function. Each of these is relevant for astronomical applications such as finding novel astrophysical objects, characterizing artifacts in data, object classification (including for rare objects), object distances, finding the important features describing objects, density estimation of distributions, probabilistic quantities, and exploring the unknown structure of new data. The software, Skytree Server, runs on any UNIX-based machine, a virtual machine, or cloud-based and distributed systems including Hadoop. We have integrated it on the cloud computing system of the Canadian Astronomical Data Centre, the Canadian Advanced Network for Astronomical Research (CANFAR), creating the world's first cloud computing data mining system for astronomy. We demonstrate results showing the scaling of each of our major algorithms on large astronomical datasets, including the full 470,992,970 objects of the 2 Micron All-Sky Survey (2MASS) Point Source Catalog. We demonstrate the ability to find outliers in the full 2MASS dataset utilizing multiple methods, e.g., nearest neighbors. This is likely of particular interest to the radio astronomy community given, for example, that survey projects contain groups dedicated to this topic. 2MASS is used as a proof-of-concept dataset due to its convenience and availability. These results are of interest to any astronomical project with large and/or complex

  13. Physical principles for scalable neural recording

    PubMed Central

    Zamft, Bradley M.; Maguire, Yael G.; Shapiro, Mikhail G.; Cybulski, Thaddeus R.; Glaser, Joshua I.; Amodei, Dario; Stranges, P. Benjamin; Kalhor, Reza; Dalrymple, David A.; Seo, Dongjin; Alon, Elad; Maharbiz, Michel M.; Carmena, Jose M.; Rabaey, Jan M.; Boyden, Edward S.; Church, George M.; Kording, Konrad P.

    2013-01-01

    Simultaneously measuring the activities of all neurons in a mammalian brain at millisecond resolution is a challenge beyond the limits of existing techniques in neuroscience. Entirely new approaches may be required, motivating an analysis of the fundamental physical constraints on the problem. We outline the physical principles governing brain activity mapping using optical, electrical, magnetic resonance, and molecular modalities of neural recording. Focusing on the mouse brain, we analyze the scalability of each method, concentrating on the limitations imposed by spatiotemporal resolution, energy dissipation, and volume displacement. Based on this analysis, all existing approaches require orders of magnitude improvement in key parameters. Electrical recording is limited by the low multiplexing capacity of electrodes and their lack of intrinsic spatial resolution, optical methods are constrained by the scattering of visible light in brain tissue, magnetic resonance is hindered by the diffusion and relaxation timescales of water protons, and the implementation of molecular recording is complicated by the stochastic kinetics of enzymes. Understanding the physical limits of brain activity mapping may provide insight into opportunities for novel solutions. For example, unconventional methods for delivering electrodes may enable unprecedented numbers of recording sites, embedded optical devices could allow optical detectors to be placed within a few scattering lengths of the measured neurons, and new classes of molecularly engineered sensors might obviate cumbersome hardware architectures. We also study the physics of powering and communicating with microscale devices embedded in brain tissue and find that, while radio-frequency electromagnetic data transmission suffers from a severe power–bandwidth tradeoff, communication via infrared light or ultrasound may allow high data rates due to the possibility of spatial multiplexing. The use of embedded local recording and

  14. Scalable Machine Learning for Massive Astronomical Datasets

    NASA Astrophysics Data System (ADS)

    Ball, Nicholas M.; Astronomy Data Centre, Canadian

    2014-01-01

    We present the ability to perform data mining and machine learning operations on a catalog of half a billion astronomical objects. This is the result of the combination of robust, highly accurate machine learning algorithms with linear scalability that renders the applications of these algorithms to massive astronomical data tractable. We demonstrate the core algorithms kernel density estimation, K-means clustering, linear regression, nearest neighbors, random forest and gradient-boosted decision tree, singular value decomposition, support vector machine, and two-point correlation function. Each of these is relevant for astronomical applications such as finding novel astrophysical objects, characterizing artifacts in data, object classification (including for rare objects), object distances, finding the important features describing objects, density estimation of distributions, probabilistic quantities, and exploring the unknown structure of new data. The software, Skytree Server, runs on any UNIX-based machine, a virtual machine, or cloud-based and distributed systems including Hadoop. We have integrated it on the cloud computing system of the Canadian Astronomical Data Centre, the Canadian Advanced Network for Astronomical Research (CANFAR), creating the world's first cloud computing data mining system for astronomy. We demonstrate results showing the scaling of each of our major algorithms on large astronomical datasets, including the full 470,992,970 objects of the 2 Micron All-Sky Survey (2MASS) Point Source Catalog. We demonstrate the ability to find outliers in the full 2MASS dataset utilizing multiple methods, e.g., nearest neighbors, and the local outlier factor. 2MASS is used as a proof-of-concept dataset due to its convenience and availability. These results are of interest to any astronomical project with large and/or complex datasets that wishes to extract the full scientific value from its data.

  15. Responsive, Flexible and Scalable Broader Impacts (Invited)

    NASA Astrophysics Data System (ADS)

    Decharon, A.; Companion, C.; Steinman, M.

    2010-12-01

    investment of time. Initiated in summer 2010, the webinars are interactive and highly flexible: people can participate from their homes anywhere and can interact according to their comfort levels (i.e., submitting questions in “chat boxes” rather than orally). Expansion - To expand scientists’ research beyond educators attending a workshop or webinar, COSEE-OS uses a blog as an additional mode of communication. Topically focused by concept maps, blogs serve as a forum for scalable content. The varied types of formatting allow scientists to create long-lived resources that remain attributed to them while supporting sustained educator engagement. Blogs are another point of contact and allow educators further asynchronous access to scientists. Based on COSEE-OS evaluations, interacting on a blog was found to be educators’ preferred method of following up with scientists. Sustained engagement of scientists or educators requires a specific return on investment. Workshops and web tools can be used together to maximize scientist impact with a relatively small investment of time. As one educator stated, “It really helps my students’ interest when we discuss concepts and I tell them my knowledge comes directly from a scientist!” [A. deCharon et al. (2009), Online tools help get scientists and educators on the same page, Eos Transactions, American Geophysical Union, 90(34), 289-290.

  16. Simplex-stochastic collocation method with improved scalability

    NASA Astrophysics Data System (ADS)

    Edeling, W. N.; Dwight, R. P.; Cinnella, P.

    2016-04-01

    The Simplex-Stochastic Collocation (SSC) method is a robust tool used to propagate uncertain input distributions through a computer code. However, it becomes prohibitively expensive for problems with dimensions higher than 5. The main purpose of this paper is to identify bottlenecks, and to improve upon this bad scalability. In order to do so, we propose an alternative interpolation stencil technique based upon the Set-Covering problem, and we integrate the SSC method in the High-Dimensional Model-Reduction framework. In addition, we address the issue of ill-conditioned sample matrices, and we present an analytical map to facilitate uniformly-distributed simplex sampling.

  17. 2-D scalable optical controlled phased-array antenna system

    NASA Astrophysics Data System (ADS)

    Chen, Maggie Yihong; Howley, Brie; Wang, Xiaolong; Basile, Panoutsopoulos; Chen, Ray T.

    2006-02-01

    A novel optoelectronically-controlled wideband 2-D phased-array antenna system is demonstrated. The inclusion of WDM devices makes a highly scalable system structure. Only (M+N) delay lines are required to control a M×N array. The optical true-time delay lines are combination of polymer waveguides and optical switches, using a single polymeric platform and are monolithically integrated on a single substrate. The 16 time delays generated by the device are measured to range from 0 to 175 ps in 11.6 ps. Far-field patterns at different steering angles in X-band are measured.

  18. pcircle - A Suite of Scalable Parallel File System Tools

    SciTech Connect

    WANG, FEIYI

    2015-10-01

    Most of the software related to file system are written for conventional local file system, they are serialized and can't take advantage of the benefit of a large scale parallel file system. "pcircle" software builds on top of ubiquitous MPI in cluster computing environment and "work-stealing" pattern to provide a scalable, high-performance suite of file system tools. In particular - it implemented parallel data copy and parallel data checksumming, with advanced features such as async progress report, checkpoint and restart, as well as integrity checking.

  19. pcircle - A Suite of Scalable Parallel File System Tools

    Energy Science and Technology Software Center (ESTSC)

    2015-10-01

    Most of the software related to file system are written for conventional local file system, they are serialized and can't take advantage of the benefit of a large scale parallel file system. "pcircle" software builds on top of ubiquitous MPI in cluster computing environment and "work-stealing" pattern to provide a scalable, high-performance suite of file system tools. In particular - it implemented parallel data copy and parallel data checksumming, with advanced features such as asyncmore » progress report, checkpoint and restart, as well as integrity checking.« less

  20. Extended spatial scalability for non dyadic video formats: from SDTV to HDTV

    NASA Astrophysics Data System (ADS)

    Marquant, Gwena"lle; Francois, Edouard; Burdin, Nicolas; Lopez, Patrick; Viéron, Jérôme

    2005-07-01

    High Definition Television is currently a hot topic that generates research and commercial interests in video industry. This paper relates to video coding devices and methods for easy and fast conversion of video formats for digital TV. The proposed algorithm provides a generic framework that allows any video format conversions in a scalable way. We focused particularly on solutions to extract Standard Definition video signals from High Definition ones. Practically, we address spatial scalability solutions between video formats that are not linked up by dyadic decomposition and/or istropic decomposition. The challenge is to generate a single scalable bitstream in a compliant way that permits any spatial resolution decodings while saving bandwidth compared to classical simulcast approaches. Practically, our scalable coder is based on a multi-layer approach and on frame scales and borders introduction. Moreover, to encode/decode the macroblocks of the high resolution pictures, usual video coding approaches exploit the inheritance knowing the macroblocks of the decoded low resolution pictures. Approaches for inter-layer prediction of motion data have already been proposed for dyadic decompositions but do not work in case of non dyadic inter-layer resolutions. Consequently we propose a different managing of the inter layer prediction. All these techniques have been successfully implemented in the MPEG-SVC reference software. Obtained results achieve good coding efficiency and are comparable to the simulcast state of the art while providing features induced by scalability.

  1. Scalable extensions of HEVC for next generation services

    NASA Astrophysics Data System (ADS)

    Misra, Kiran; Segall, Andrew; Zhao, Jie; Kim, Seung-Hwan

    2013-02-01

    The high efficiency video coding (HEVC) standard being developed by ITU-T VCEG and ISO/IEC MPEG achieves a compression goal of reducing the bitrate by half for the same visual quality when compared with earlier video compression standards such as H.264/AVC. It achieves this goal with the use of several new tools such as quad-tree based partitioning of data, larger block sizes, improved intra prediction, the use of sophisticated prediction of motion information, inclusion of an in-loop sample adaptive offset process etc. This paper describes an approach where the HEVC framework is extended to achieve spatial scalability using a multi-loop approach. The enhancement layer inter-predictive coding efficiency is improved by including within the decoded picture buffer multiple up-sampled versions of the decoded base layer picture. This approach has the advantage of achieving significant coding gains with a simple extension of the base layer tools such as inter-prediction, motion information signaling etc. Coding efficiency of the enhancement layer is further improved using adaptive loop filter and internal bit-depth increment. The performance of the proposed scalable video coding approach is compared to simulcast transmission of video data using high efficiency model version 6.1 (HM-6.1). The bitrate savings are measured using Bjontegaard Delta (BD) rate for a spatial scalability factor of 2 and 1.5 respectively when compared with simulcast anchors. It is observed that the proposed approach provides an average luma BD rate gains of 33.7% and 50.5% respectively.

  2. Experimental verification of highly scalable OXC that consists of subsystem-modular express-switch part and multicast-switch-based add/drop part enabling total throughput of 314 Tbps.

    PubMed

    Takashina, Shoichi; Ishida, Hiroto; Niwa, Masaki; Mori, Yojiro; Hasegawa, Hiroshi; Sato, Ken-Ichi; Watanabe, Toshio

    2015-06-01

    We propose a cost-effective and scalable OXC/ROADM that consists of a subsystem-modular express switch part and a transponder-bank-based add/drop part. The effectiveness of the proposed architecture is verified via a hardware scale evaluation, network performance simulations, and transmission experiments. The architecture enables large throughput and offers significant hardware-scale reductions with marginal fiber-utilization penalty against the conventional architectures. A part of the OXC/ROADM designed to accommodate 35x35 express fiber ports and 2,800 transponders for add/drop is constructed. Its net throughput reaches 314 Tbps using 80 channels of 120-Gbps signal (30-Gbaud dual-polarization quadrature phase-shift-keying signals with 7% overhead are assumed). PMID:26072838

  3. Scalable Molecular Dynamics with NAMD

    PubMed Central

    Phillips, James C.; Braun, Rosemary; Wang, Wei; Gumbart, James; Tajkhorshid, Emad; Villa, Elizabeth; Chipot, Christophe; Skeel, Robert D.; Kalé, Laxmikant; Schulten, Klaus

    2008-01-01

    NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems. NAMD scales to hundreds of processors on high-end parallel platforms, as well as tens of processors on low-cost commodity clusters, and also runs on individual desktop and laptop computers. NAMD works with AMBER and CHARMM potential functions, parameters, and file formats. This paper, directed to novices as well as experts, first introduces concepts and methods used in the NAMD program, describing the classical molecular dynamics force field, equations of motion, and integration methods along with the efficient electrostatics evaluation algorithms employed and temperature and pressure controls used. Features for steering the simulation across barriers and for calculating both alchemical and conformational free energy differences are presented. The motivations for and a roadmap to the internal design of NAMD, implemented in C++ and based on Charm++ parallel objects, are outlined. The factors affecting the serial and parallel performance of a simulation are discussed. Next, typical NAMD use is illustrated with representative applications to a small, a medium, and a large biomolecular system, highlighting particular features of NAMD, e.g., the Tcl scripting language. Finally, the paper provides a list of the key features of NAMD and discusses the benefits of combining NAMD with the molecular graphics/sequence analysis software VMD and the grid computing/collaboratory software BioCoRE. NAMD is distributed free of charge with source code at www.ks.uiuc.edu. PMID:16222654

  4. Scalable video coding in frequency domain

    NASA Astrophysics Data System (ADS)

    Civanlar, Mehmet R.; Puri, Atul

    1992-11-01

    Scalable video coding is important in a number of applications where video needs to be decoded and displayed at a variety of resolution scales. It is more efficient than simulcasting, in which all desired resolution scales are coded totally independent of one another within the constraint of a fixed available bandwidth. In this paper, we focus on scalability using the frequency domain approach. We employ the framework proposed for the ongoing second phase of Motion Picture Experts Group (MPEG-2) standard to study the performance of one such scheme and investigate improvements aimed at increasing its efficiency. Practical issues related to multiplexing of encoded data of various resolution scales to facilitate decoding are considered. Simulations are performed to investigate the potential of a chosen frequency domain scheme. Various prospects and limitations are also discussed.

  5. Scalable descriptive and correlative statistics with Titan.

    SciTech Connect

    Thompson, David C.; Pebay, Philippe Pierre

    2008-12-01

    This report summarizes the existing statistical engines in VTK/Titan and presents the parallel versions thereof which have already been implemented. The ease of use of these parallel engines is illustrated by the means of C++ code snippets. Furthermore, this report justifies the design of these engines with parallel scalability in mind; then, this theoretical property is verified with test runs that demonstrate optimal parallel speed-up with up to 200 processors.

  6. Scalable Computer Performance and Analysis (Hierarchical INTegration)

    Energy Science and Technology Software Center (ESTSC)

    1999-09-02

    HINT is a program to measure a wide variety of scalable computer systems. It is capable of demonstrating the benefits of using more memory or processing power, and of improving communications within the system. HINT can be used for measurement of an existing system, while the associated program ANALYTIC HINT can be used to explain the measurements or as a design tool for proposed systems.

  7. Scalable Domain Decomposed Monte Carlo Particle Transport

    NASA Astrophysics Data System (ADS)

    O'Brien, Matthew Joseph

    In this dissertation, we present the parallel algorithms necessary to run domain decomposed Monte Carlo particle transport on large numbers of processors (millions of processors). Previous algorithms were not scalable, and the parallel overhead became more computationally costly than the numerical simulation. The main algorithms we consider are: • Domain decomposition of constructive solid geometry: enables extremely large calculations in which the background geometry is too large to fit in the memory of a single computational node. • Load Balancing: keeps the workload per processor as even as possible so the calculation runs efficiently. • Global Particle Find: if particles are on the wrong processor, globally resolve their locations to the correct processor based on particle coordinate and background domain. • Visualizing constructive solid geometry, sourcing particles, deciding that particle streaming communication is completed and spatial redecomposition. These algorithms are some of the most important parallel algorithms required for domain decomposed Monte Carlo particle transport. We demonstrate that our previous algorithms were not scalable, prove that our new algorithms are scalable, and run some of the algorithms up to 2 million MPI processes on the Sequoia supercomputer.

  8. A scalable infrastructure for CMS data analysis based on OpenStack Cloud and Gluster file system

    NASA Astrophysics Data System (ADS)

    Toor, S.; Osmani, L.; Eerola, P.; Kraemer, O.; Lindén, T.; Tarkoma, S.; White, J.

    2014-06-01

    The challenge of providing a resilient and scalable computational and data management solution for massive scale research environments requires continuous exploration of new technologies and techniques. In this project the aim has been to design a scalable and resilient infrastructure for CERN HEP data analysis. The infrastructure is based on OpenStack components for structuring a private Cloud with the Gluster File System. We integrate the state-of-the-art Cloud technologies with the traditional Grid middleware infrastructure. Our test results show that the adopted approach provides a scalable and resilient solution for managing resources without compromising on performance and high availability.

  9. Scalable and sustainable electrochemical allylic C-H oxidation.

    PubMed

    Horn, Evan J; Rosen, Brandon R; Chen, Yong; Tang, Jiaze; Chen, Ke; Eastgate, Martin D; Baran, Phil S

    2016-05-01

    New methods and strategies for the direct functionalization of C-H bonds are beginning to reshape the field of retrosynthetic analysis, affecting the synthesis of natural products, medicines and materials. The oxidation of allylic systems has played a prominent role in this context as possibly the most widely applied C-H functionalization, owing to the utility of enones and allylic alcohols as versatile intermediates, and their prevalence in natural and unnatural materials. Allylic oxidations have featured in hundreds of syntheses, including some natural product syntheses regarded as "classics". Despite many attempts to improve the efficiency and practicality of this transformation, the majority of conditions still use highly toxic reagents (based around toxic elements such as chromium or selenium) or expensive catalysts (such as palladium or rhodium). These requirements are problematic in industrial settings; currently, no scalable and sustainable solution to allylic oxidation exists. This oxidation strategy is therefore rarely used for large-scale synthetic applications, limiting the adoption of this retrosynthetic strategy by industrial scientists. Here we describe an electrochemical C-H oxidation strategy that exhibits broad substrate scope, operational simplicity and high chemoselectivity. It uses inexpensive and readily available materials, and represents a scalable allylic C-H oxidation (demonstrated on 100 grams), enabling the adoption of this C-H oxidation strategy in large-scale industrial settings without substantial environmental impact. PMID:27096371

  10. Design of scalable optical interconnection network using wavelength division multiplexing

    NASA Astrophysics Data System (ADS)

    Jing, Wencai; Tian, Jindong; Zhou, Ge; Zhang, Yimo; Liu, Wei; Zhang, Xun

    2000-04-01

    This paper describes the two-layer scalable wavelength routing optical interconnection network being developed in Tianjin University. The top layer of the network is multi- wavelength bi-directional optical bus, which has high bandwidth and low latency. The optical bus is made up of passive components, no wavelength-tunable devices have been sued. As a result, the optical bus has low communication latency that is mainly decided by the optical fiber length. The sub-layer of the network is single-wavelength ring, which has low communication latency and high-scalability. In each ring, a wavelength routing node is used for data transmission between the ring and the optical bus. Each node computer is connected to the ring using an optical network interface card, which is based on peripheral component interconnect bus. The communication latency inside the ring is decreased using synchronous pipelining transmission technique. The scale of the ring is mainly limited by the efficient bandwidth required by each node computer. The number of rings is mainly decided by the optical power of the laser diodes and the sensitivity of the optical detectors. If Erbium doped fiber amplifier is used in the optical bus, the scale of the network can be further developed.

  11. Scalable and sustainable electrochemical allylic C–H oxidation

    NASA Astrophysics Data System (ADS)

    Horn, Evan J.; Rosen, Brandon R.; Chen, Yong; Tang, Jiaze; Chen, Ke; Eastgate, Martin D.; Baran, Phil S.

    2016-05-01

    New methods and strategies for the direct functionalization of C–H bonds are beginning to reshape the field of retrosynthetic analysis, affecting the synthesis of natural products, medicines and materials. The oxidation of allylic systems has played a prominent role in this context as possibly the most widely applied C–H functionalization, owing to the utility of enones and allylic alcohols as versatile intermediates, and their prevalence in natural and unnatural materials. Allylic oxidations have featured in hundreds of syntheses, including some natural product syntheses regarded as “classics”. Despite many attempts to improve the efficiency and practicality of this transformation, the majority of conditions still use highly toxic reagents (based around toxic elements such as chromium or selenium) or expensive catalysts (such as palladium or rhodium). These requirements are problematic in industrial settings; currently, no scalable and sustainable solution to allylic oxidation exists. This oxidation strategy is therefore rarely used for large-scale synthetic applications, limiting the adoption of this retrosynthetic strategy by industrial scientists. Here we describe an electrochemical C–H oxidation strategy that exhibits broad substrate scope, operational simplicity and high chemoselectivity. It uses inexpensive and readily available materials, and represents a scalable allylic C–H oxidation (demonstrated on 100 grams), enabling the adoption of this C–H oxidation strategy in large-scale industrial settings without substantial environmental impact.

  12. Scalable Track Initiation for Optical Space Surveillance

    NASA Astrophysics Data System (ADS)

    Schumacher, P.; Wilkins, M. P.

    2012-09-01

    The advent of high-sensitivity, high-capacity optical sensors for space surveillance presents us with interesting and challenging tracking problems. Accounting for the origin of every detection made by such systems is generally agreed to belong to the "most difficult" category of tracking problems. Especially in the early phases of the tracking scenario, when a catalog of targets is being compiled, or when many new objects appear in space because of on-orbit explosion or collision, one faces a combinatorially large number of orbit (data association) hypotheses to evaluate. The number of hypotheses is reduced to a more feasible number if observations close together in time can, with high confidence, be associated by the sensor into extended tracks on single objects. Most current space surveillance techniques are predicated on the sensor systems' ability to form such tracks reliably. However, the required operational tempo of space surveillance, the very large number of objects in Earth orbit and the difficulties of detecting dim, fast-moving targets at long ranges means that individual sensor track reports are often inadequate for computing initial orbit hypotheses. In fact, this situation can occur with optical sensors even when the probability of detection is high. For example, the arc of orbit that has been observed may be too short or may have been sampled too sparsely to allow well-conditioned, usable orbit estimates from single tracks. In that case, one has no choice but to solve a data association problem involving an unknown number of targets and many widely spaced observations of uncertain origin. In the present paper, we are motivated by this more difficult aspect of the satellite cataloging problem. However, the results of this analysis may find use in a variety of less stressing tracking applications. The computational complexity of track initiation using only angle measurements is polynomial in time. However, the polynomial degree can be high, always at

  13. Optical Properties of Scalable Nano-Mesh Films

    NASA Astrophysics Data System (ADS)

    Alvine, Kyle; Bernacki, Bruce; Bennett, Wendy; Schemer-Kohrn, Alan

    2015-03-01

    We describe here the optical properties of a scalable nano-mesh film both experimentally measured and calculated by FDTD numerical modeling. Typically, applications for optically responsive nano-plasmonic or photonic films are limited by virtue of tractable fabrication techniques to several hundred microns or a few millimeters in size. The films described here have been demonstrated over an extent of several inches and could be readily scaled to larger sizes. The films are comprised of a quasi-regular periodic array of nanoscale holes in a metallic film. The nanostructure is fabricated in a scalable fashion in a multi-step fashion via sputtering on a nanoscale template created by nanoparticle self-assembly. Both the numerical modeling and experimentally measured scattering demonstrate that these films are highly resonant with the resonance location in the visible or near infrared and set by the hole size and pattern geometry. Such films can also be readily be made on flexible substrates if desired. Potential applications include new proposed photonic thermal management coatings or plasmoelectric devices.

  14. Scalable tuning of building models to hourly data

    SciTech Connect

    Garrett, Aaron; New, Joshua Ryan

    2015-03-31

    Energy models of existing buildings are unreliable unless calibrated so they correlate well with actual energy usage. Manual tuning requires a skilled professional, is prohibitively expensive for small projects, imperfect, non-repeatable, non-transferable, and not scalable to the dozens of sensor channels that smart meters, smart appliances, and cheap/ubiquitous sensors are beginning to make available today. A scalable, automated methodology is needed to quickly and intelligently calibrate building energy models to all available data, increase the usefulness of those models, and facilitate speed-and-scale penetration of simulation-based capabilities into the marketplace for actualized energy savings. The "Autotune'' project is a novel, model-agnostic methodology which leverages supercomputing, large simulation ensembles, and big data mining with multiple machine learning algorithms to allow automatic calibration of simulations that match measured experimental data in a way that is deployable on commodity hardware. This paper shares several methodologies employed to reduce the combinatorial complexity to a computationally tractable search problem for hundreds of input parameters. Furthermore, accuracy metrics are provided which quantify model error to measured data for either monthly or hourly electrical usage from a highly-instrumented, emulated-occupancy research home.

  15. Scalable tuning of building models to hourly data

    DOE PAGESBeta

    Garrett, Aaron; New, Joshua Ryan

    2015-03-31

    Energy models of existing buildings are unreliable unless calibrated so they correlate well with actual energy usage. Manual tuning requires a skilled professional, is prohibitively expensive for small projects, imperfect, non-repeatable, non-transferable, and not scalable to the dozens of sensor channels that smart meters, smart appliances, and cheap/ubiquitous sensors are beginning to make available today. A scalable, automated methodology is needed to quickly and intelligently calibrate building energy models to all available data, increase the usefulness of those models, and facilitate speed-and-scale penetration of simulation-based capabilities into the marketplace for actualized energy savings. The "Autotune'' project is a novel, model-agnosticmore » methodology which leverages supercomputing, large simulation ensembles, and big data mining with multiple machine learning algorithms to allow automatic calibration of simulations that match measured experimental data in a way that is deployable on commodity hardware. This paper shares several methodologies employed to reduce the combinatorial complexity to a computationally tractable search problem for hundreds of input parameters. Furthermore, accuracy metrics are provided which quantify model error to measured data for either monthly or hourly electrical usage from a highly-instrumented, emulated-occupancy research home.« less

  16. Scalable parallel distance field construction for large-scale applications

    DOE PAGESBeta

    Yu, Hongfeng; Xie, Jinrong; Ma, Kwan -Liu; Kolla, Hemanth; Chen, Jacqueline H.

    2015-10-01

    Computing distance fields is fundamental to many scientific and engineering applications. Distance fields can be used to direct analysis and reduce data. In this paper, we present a highly scalable method for computing 3D distance fields on massively parallel distributed-memory machines. Anew distributed spatial data structure, named parallel distance tree, is introduced to manage the level sets of data and facilitate surface tracking overtime, resulting in significantly reduced computation and communication costs for calculating the distance to the surface of interest from any spatial locations. Our method supports several data types and distance metrics from real-world applications. We demonstrate itsmore » efficiency and scalability on state-of-the-art supercomputers using both large-scale volume datasets and surface models. We also demonstrate in-situ distance field computation on dynamic turbulent flame surfaces for a petascale combustion simulation. In conclusion, our work greatly extends the usability of distance fields for demanding applications.« less

  17. Performance-scalable volumetric data classification for online industrial inspection

    NASA Astrophysics Data System (ADS)

    Abraham, Aby J.; Sadki, Mustapha; Lea, R. M.

    2002-03-01

    Non-intrusive inspection and non-destructive testing of manufactured objects with complex internal structures typically requires the enhancement, analysis and visualization of high-resolution volumetric data. Given the increasing availability of fast 3D scanning technology (e.g. cone-beam CT), enabling on-line detection and accurate discrimination of components or sub-structures, the inherent complexity of classification algorithms inevitably leads to throughput bottlenecks. Indeed, whereas typical inspection throughput requirements range from 1 to 1000 volumes per hour, depending on density and resolution, current computational capability is one to two orders-of-magnitude less. Accordingly, speeding up classification algorithms requires both reduction of algorithm complexity and acceleration of computer performance. A shape-based classification algorithm, offering algorithm complexity reduction, by using ellipses as generic descriptors of solids-of-revolution, and supporting performance-scalability, by exploiting the inherent parallelism of volumetric data, is presented. A two-stage variant of the classical Hough transform is used for ellipse detection and correlation of the detected ellipses facilitates position-, scale- and orientation-invariant component classification. Performance-scalability is achieved cost-effectively by accelerating a PC host with one or more COTS (Commercial-Off-The-Shelf) PCI multiprocessor cards. Experimental results are reported to demonstrate the feasibility and cost-effectiveness of the data-parallel classification algorithm for on-line industrial inspection applications.

  18. The Node Monitoring Component of a Scalable Systems Software Environment

    SciTech Connect

    Samuel James Miller

    2006-08-09

    This research describes Fountain, a suite of programs used to monitor the resources of a cluster. A cluster is a collection of individual computers that are connected via a high speed communication network. They are traditionally used by users who desire more resources, such as processing power and memory, than any single computer can provide. A common drawback to effectively utilizing such a large-scale system is the management infrastructure, which often does not often scale well as the system grows. Large-scale parallel systems provide new research challenges in the area of systems software, the programs or tools that manage the system from boot-up to running a parallel job. The approach presented in this thesis utilizes a collection of separate components that communicate with each other to achieve a common goal. While systems software comprises a broad array of components, this thesis focuses on the design choices for a node monitoring component. We will describe Fountain, an implementation of the Scalable Systems Software (SSS) node monitor specification. It is targeted at aggregate node monitoring for clusters, focusing on both scalability and fault tolerance as its design goals. It leverages widely used technologies such as XML and HTTP to present an interface to other components in the SSS environment.

  19. Scalable mobile image retrieval by exploring contextual saliency.

    PubMed

    Yang, Xiyu; Qian, Xueming; Xue, Yao

    2015-06-01

    Nowadays, it is very convenient to capture photos by a smart phone. As using, the smart phone is a convenient way to share what users experienced anytime and anywhere through social networks, it is very possible that we capture multiple photos to make sure the content is well photographed. In this paper, an effective scalable mobile image retrieval approach is proposed by exploring contextual salient information for the input query image. Our goal is to explore the high-level semantic information of an image by finding the contextual saliency from multiple relevant photos rather than solely using the input image. Thus, the proposed mobile image retrieval approach first determines the relevant photos according to visual similarity, then mines salient features by exploring contextual saliency from multiple relevant images, and finally determines contributions of salient features for scalable retrieval. Compared with the existing mobile-based image retrieval approaches, our approach requires less bandwidth and has better retrieval performance. We can carry out retrieval with <200-B data, which is <5% of existing approaches. Most importantly, when the bandwidth is limited, we can rank the transmitted features according to their contributions to retrieval. Experimental results show the effectiveness of the proposed approach. PMID:25775488

  20. Scalable and Fault Tolerant Failure Detection and Consensus

    SciTech Connect

    Katti, Amogh; Di Fatta, Giuseppe; Naughton III, Thomas J; Engelmann, Christian

    2015-01-01

    Future extreme-scale high-performance computing systems will be required to work under frequent component failures. The MPI Forum's User Level Failure Mitigation proposal has introduced an operation, MPI_Comm_shrink, to synchronize the alive processes on the list of failed processes, so that applications can continue to execute even in the presence of failures by adopting algorithm-based fault tolerance techniques. This MPI_Comm_shrink operation requires a fault tolerant failure detection and consensus algorithm. This paper presents and compares two novel failure detection and consensus algorithms. The proposed algorithms are based on Gossip protocols and are inherently fault-tolerant and scalable. The proposed algorithms were implemented and tested using the Extreme-scale Simulator. The results show that in both algorithms the number of Gossip cycles to achieve global consensus scales logarithmically with system size. The second algorithm also shows better scalability in terms of memory and network bandwidth usage and a perfect synchronization in achieving global consensus.

  1. First experience with the scalable coherent interface

    SciTech Connect

    Mueller, H. . ECP Division); RD24 Collaboration

    1994-02-01

    The research project RD24 is studying applications of the Scalable Coherent Interface (IEEE-1596) standard for the large hadron collider (LHC). First SCI node chips from Dolphin were used to demonstrate the use and functioning of SCI's packet protocols and to measure data rates. The authors present results from a first, two-node SCI ringlet at CERN, based on a R3000 RISC processor node and DMA node on a MC68040 processor bus. A diagnostic link analyzer monitors the SCI packet protocols up to full link bandwidth. In its second phase, RD24 will build a first implementation of a multi-ringlet SCI data merger.

  2. Scalable analog wavefront sensor with subpixel resolution

    NASA Astrophysics Data System (ADS)

    Wilcox, Michael

    2006-06-01

    Standard Shack-Hartman wavefront sensors use a CCD element to sample position and distortion of a target or guide star. Digital sampling of the element and transfer to a memory space for subsequent computation adds significant temporal delay, thus, limiting the spatial frequency and scalability of the system as a wavefront sensor. A new approach to sampling uses information processing principles in an insect compound eye. Analog circuitry eliminates digital sampling and extends the useful range of the system to control a deformable mirror and make a faster, more capable wavefront sensor.

  3. Overcoming Scalability Challenges for Tool Daemon Launching

    SciTech Connect

    Ahn, D H; Arnold, D C; de Supinski, B R; Lee, G L; Miller, B P; Schulz, M

    2008-02-15

    Many tools that target parallel and distributed environments must co-locate a set of daemons with the distributed processes of the target application. However, efficient and portable deployment of these daemons on large scale systems is an unsolved problem. We overcome this gap with LaunchMON, a scalable, robust, portable, secure, and general purpose infrastructure for launching tool daemons. Its API allows tool builders to identify all processes of a target job, launch daemons on the relevant nodes and control daemon interaction. Our results show that Launch-MON scales to very large daemon counts and substantially enhances performance over existing ad hoc mechanisms.

  4. Scalable Unix tools on parallel processors

    SciTech Connect

    Gropp, W.; Lusk, E.

    1994-12-31

    The introduction of parallel processors that run a separate copy of Unix on each process has introduced new problems in managing the user`s environment. This paper discusses some generalizations of common Unix commands for managing files (e.g. 1s) and processes (e.g. ps) that are convenient and scalable. These basic tools, just like their Unix counterparts, are text-based. We also discuss a way to use these with a graphical user interface (GUI). Some notes on the implementation are provided. Prototypes of these commands are publicly available.

  5. SPRNG Scalable Parallel Random Number Generator LIbrary

    Energy Science and Technology Software Center (ESTSC)

    2010-03-16

    This revision corrects some errors in SPRNG 1. Users of newer SPRNG versions can obtain the corrected files and build their version with it. This version also improves the scalability of some of the application-based tests in the SPRNG test suite. It also includes an interface to a parallel Mersenne Twister, so that if users install the Mersenne Twister, then they can test this generator with the SPRNG test suite and also use some SPRNGmore » features with that generator.« less

  6. SWAP-Assembler: scalable and efficient genome assembly towards thousands of cores

    PubMed Central

    2014-01-01

    Background There is a widening gap between the throughput of massive parallel sequencing machines and the ability to analyze these sequencing data. Traditional assembly methods requiring long execution time and large amount of memory on a single workstation limit their use on these massive data. Results This paper presents a highly scalable assembler named as SWAP-Assembler for processing massive sequencing data using thousands of cores, where SWAP is an acronym for Small World Asynchronous Parallel model. In the paper, a mathematical description of multi-step bi-directed graph (MSG) is provided to resolve the computational interdependence on merging edges, and a highly scalable computational framework for SWAP is developed to automatically preform the parallel computation of all operations. Graph cleaning and contig extension are also included for generating contigs with high quality. Experimental results show that SWAP-Assembler scales up to 2048 cores on Yanhuang dataset using only 26 minutes, which is better than several other parallel assemblers, such as ABySS, Ray, and PASHA. Results also show that SWAP-Assembler can generate high quality contigs with good N50 size and low error rate, especially it generated the longest N50 contig sizes for Fish and Yanhuang datasets. Conclusions In this paper, we presented a highly scalable and efficient genome assembly software, SWAP-Assembler. Compared with several other assemblers, it showed very good performance in terms of scalability and contig quality. This software is available at: https://sourceforge.net/projects/swapassembler PMID:25253533

  7. Scalable low complexity image coder for remote volume visualization

    NASA Astrophysics Data System (ADS)

    Lalgudi, Hariharan G.; Marcellin, Michael W.; Bilgin, Ali; Nadar, Mariappan S.

    2008-08-01

    Remote visualization of volumetric data has gained importance over the past few years in order to realize the full potential of tele-radiology. Volume rendering is a computationally intensive process, often requiring hardware acceleration to achieve real time visualization. Hence a remote visualization model that is well-suited for high speed networks would be to transmit rendered images from the server (with dedicated hardware) based on view point requests from clients. In this regard, a compression scheme for the rendered images is vital for efficient utilization of the server-client bandwidth. Also, the complexity of the decompressor should be considered so that a low end client workstation can decode images at the desired frame rate. We present a scalable low complexity image coder that has good compression efficiency and high throughput.

  8. Lilith: A scalable secure tool for massively parallel distributed computing

    SciTech Connect

    Armstrong, R.C.; Camp, L.J.; Evensky, D.A.; Gentile, A.C.

    1997-06-01

    Changes in high performance computing have necessitated the ability to utilize and interrogate potentially many thousands of processors. The ASCI (Advanced Strategic Computing Initiative) program conducted by the United States Department of Energy, for example, envisions thousands of distinct operating systems connected by low-latency gigabit-per-second networks. In addition multiple systems of this kind will be linked via high-capacity networks with latencies as low as the speed of light will allow. Code which spans systems of this sort must be scalable; yet constructing such code whether for applications, debugging, or maintenance is an unsolved problem. Lilith is a research software platform that attempts to answer these questions with an end toward meeting these needs. Presently, Lilith exists as a test-bed, written in Java, for various spanning algorithms and security schemes. The test-bed software has, and enforces, hooks allowing implementation and testing of various security schemes.

  9. An Open Infrastructure for Scalable, Reconfigurable Analysis

    SciTech Connect

    de Supinski, B R; Fowler, R; Gamblin, T; Mueller, F; Ratn, P; Schulz, M

    2008-05-15

    Petascale systems will have hundreds of thousands of processor cores so their applications must be massively parallel. Effective use of petascale systems will require efficient interprocess communication through memory hierarchies and complex network topologies. Tools to collect and analyze detailed data about this communication would facilitate its optimization. However, several factors complicate tool design. First, large-scale runs on petascale systems will be a precious commodity, so scalable tools must have almost no overhead. Second, the volume of performance data from petascale runs could easily overwhelm hand analysis and, thus, tools must collect only data that is relevant to diagnosing performance problems. Analysis must be done in-situ, when available processing power is proportional to the data. We describe a tool framework that overcomes these complications. Our approach allows application developers to combine existing techniques for measurement, analysis, and data aggregation to develop application-specific tools quickly. Dynamic configuration enables application developers to select exactly the measurements needed and generic components support scalable aggregation and analysis of this data with little additional effort.

  10. Using the scalable nonlinear equations solvers package

    SciTech Connect

    Gropp, W.D.; McInnes, L.C.; Smith, B.F.

    1995-02-01

    SNES (Scalable Nonlinear Equations Solvers) is a software package for the numerical solution of large-scale systems of nonlinear equations on both uniprocessors and parallel architectures. SNES also contains a component for the solution of unconstrained minimization problems, called SUMS (Scalable Unconstrained Minimization Solvers). Newton-like methods, which are known for their efficiency and robustness, constitute the core of the package. As part of the multilevel PETSc library, SNES incorporates many features and options from other parts of PETSc. In keeping with the spirit of the PETSc library, the nonlinear solution routines are data-structure-neutral, making them flexible and easily extensible. This users guide contains a detailed description of uniprocessor usage of SNES, with some added comments regarding multiprocessor usage. At this time the parallel version is undergoing refinement and extension, as we work toward a common interface for the uniprocessor and parallel cases. Thus, forthcoming versions of the software will contain additional features, and changes to parallel interface may result at any time. The new parallel version will employ the MPI (Message Passing Interface) standard for interprocessor communication. Since most of these details will be hidden, users will need to perform only minimal message-passing programming.

  11. Scalable hybrid unstructured and structured grid raycasting.

    PubMed

    Muigg, Philipp; Hadwiger, Markus; Doleisch, Helmut; Hauser, Helwig

    2007-01-01

    This paper presents a scalable framework for real-time raycasting of large unstructured volumes that employs a hybrid bricking approach. It adaptively combines original unstructured bricks in important (focus) regions, with structured bricks that are resampled on demand in less important (context) regions. The basis of this focus+context approach is interactive specification of a scalar degree of interest (DOI) function. Thus, rendering always considers two volumes simultaneously: a scalar data volume, and the current DOI volume. The crucial problem of visibility sorting is solved by raycasting individual bricks and compositing in visibility order from front to back. In order to minimize visual errors at the grid boundary, it is always rendered accurately, even for resampled bricks. A variety of different rendering modes can be combined, including contour enhancement. A very important property of our approach is that it supports a variety of cell types natively, i.e., it is not constrained to tetrahedral grids, even when interpolation within cells is used. Moreover, our framework can handle multi-variate data, e.g., multiple scalar channels such as temperature or pressure, as well as time-dependent data. The combination of unstructured and structured bricks with different quality characteristics such as the type of interpolation or resampling resolution in conjunction with custom texture memory management yields a very scalable system. PMID:17968114

  12. Towards Scalable Graph Computation on Mobile Devices

    PubMed Central

    Chen, Yiqi; Lin, Zhiyuan; Pienta, Robert; Kahng, Minsuk; Chau, Duen Horng

    2015-01-01

    Mobile devices have become increasingly central to our everyday activities, due to their portability, multi-touch capabilities, and ever-improving computational power. Such attractive features have spurred research interest in leveraging mobile devices for computation. We explore a novel approach that aims to use a single mobile device to perform scalable graph computation on large graphs that do not fit in the device's limited main memory, opening up the possibility of performing on-device analysis of large datasets, without relying on the cloud. Based on the familiar memory mapping capability provided by today's mobile operating systems, our approach to scale up computation is powerful and intentionally kept simple to maximize its applicability across the iOS and Android platforms. Our experiments demonstrate that an iPad mini can perform fast computation on large real graphs with as many as 272 million edges (Google+ social graph), at a speed that is only a few times slower than a 13″ Macbook Pro. Through creating a real world iOS app with this technique, we demonstrate the strong potential application for scalable graph computation on a single mobile device using our approach. PMID:25859564

  13. Scalable enantioselective total synthesis of taxanes

    NASA Astrophysics Data System (ADS)

    Mendoza, Abraham; Ishihara, Yoshihiro; Baran, Phil S.

    2012-01-01

    Taxanes form a large family of terpenes comprising over 350 members, the most famous of which is Taxol (paclitaxel), a billion-dollar anticancer drug. Here, we describe the first practical and scalable synthetic entry to these natural products via a concise preparation of (+)-taxa-4(5),11(12)-dien-2-one, which has a suitable functional handle with which to access more oxidized members of its family. This route enables a gram-scale preparation of the ‘parent’ taxane—taxadiene—which is the largest quantity of this naturally occurring terpene ever isolated or prepared in pure form. The characteristic 6-8-6 tricyclic system of the taxane family, containing a bridgehead alkene, is forged via a vicinal difunctionalization/Diels-Alder strategy. Asymmetry is introduced by means of an enantioselective conjugate addition that forms an all-carbon quaternary centre, from which all other stereocentres are fixed through substrate control. This study lays a critical foundation for a planned access to minimally oxidized taxane analogues and a scalable laboratory preparation of Taxol itself.

  14. Scalable Quantum Computing Over the Rainbow

    NASA Astrophysics Data System (ADS)

    Pfister, Olivier; Menicucci, Nicolas C.; Flammia, Steven T.

    2011-03-01

    The physical implementation of nontrivial quantum computing is an experimental challenge due to decoherence and the need for scalability. Recently we proved a novel theoretical scheme for realizing a scalable quantum register of very large size, entangled in a cluster state, in the optical frequency comb (OFC) defined by the eigenmodes of a single optical parametric oscillator (OPO). The classical OFC is well known as implemented by the femtosecond, carrier-envelope-phase- and mode-locked lasers which have redefined frequency metrology in recent years. The quantum OFC is a set of harmonic oscillators, or Qmodes, whose amplitude and phase quadratures are continuous variables, the manipulation of which is a mature field for one or two Qmodes. We have shown that the nonlinear optical medium of a single OPO can be engineered, in a sophisticated but already demonstrated manner, so as to entangle in constant time the OPO's OFC into a finitely squeezed, Gaussian cluster state suitable for universal quantum computing over continuous variables. Here we summarize our theoretical result and survey the ongoing experimental efforts in this direction.

  15. Scalable Multi-Platform Distribution of Spatial 3d Contents

    NASA Astrophysics Data System (ADS)

    Klimke, J.; Hagedorn, B.; Döllner, J.

    2013-09-01

    Virtual 3D city models provide powerful user interfaces for communication of 2D and 3D geoinformation. Providing high quality visualization of massive 3D geoinformation in a scalable, fast, and cost efficient manner is still a challenging task. Especially for mobile and web-based system environments, software and hardware configurations of target systems differ significantly. This makes it hard to provide fast, visually appealing renderings of 3D data throughout a variety of platforms and devices. Current mobile or web-based solutions for 3D visualization usually require raw 3D scene data such as triangle meshes together with textures delivered from server to client, what makes them strongly limited in terms of size and complexity of the models they can handle. In this paper, we introduce a new approach for provisioning of massive, virtual 3D city models on different platforms namely web browsers, smartphones or tablets, by means of an interactive map assembled from artificial oblique image tiles. The key concept is to synthesize such images of a virtual 3D city model by a 3D rendering service in a preprocessing step. This service encapsulates model handling and 3D rendering techniques for high quality visualization of massive 3D models. By generating image tiles using this service, the 3D rendering process is shifted from the client side, which provides major advantages: (a) The complexity of the 3D city model data is decoupled from data transfer complexity (b) the implementation of client applications is simplified significantly as 3D rendering is encapsulated on server side (c) 3D city models can be easily deployed for and used by a large number of concurrent users, leading to a high degree of scalability of the overall approach. All core 3D rendering techniques are performed on a dedicated 3D rendering server, and thin-client applications can be compactly implemented for various devices and platforms.

  16. Scalable Track Initiation for Optical Space Surveillance

    NASA Astrophysics Data System (ADS)

    Schumacher, P.; Wilkins, M. P.

    2012-09-01

    The advent of high-sensitivity, high-capacity optical sensors for space surveillance presents us with interesting and challenging tracking problems. Accounting for the origin of every detection made by such systems is generally agreed to belong to the "most difficult" category of tracking problems. Especially in the early phases of the tracking scenario, when a catalog of targets is being compiled, or when many new objects appear in space because of on-orbit explosion or collision, one faces a combinatorially large number of orbit (data association) hypotheses to evaluate. The number of hypotheses is reduced to a more feasible number if observations close together in time can, with high confidence, be associated by the sensor into extended tracks on single objects. Most current space surveillance techniques are predicated on the sensor systems' ability to form such tracks reliably. However, the required operational tempo of space surveillance, the very large number of objects in Earth orbit and the difficulties of detecting dim, fast-moving targets at long ranges means that individual sensor track reports are often inadequate for computing initial orbit hypotheses. In fact, this situation can occur with optical sensors even when the probability of detection is high. For example, the arc of orbit that has been observed may be too short or may have been sampled too sparsely to allow well-conditioned, usable orbit estimates from single tracks. In that case, one has no choice but to solve a data association problem involving an unknown number of targets and many widely spaced observations of uncertain origin. In the present paper, we are motivated by this more difficult aspect of the satellite cataloging problem. However, the results of this analysis may find use in a variety of less stressing tracking applications. The computational complexity of track initiation using only angle measurements is polynomial in time. However, the polynomial degree can be high, always at

  17. Fundamental research on scalable DNA molecular computation

    NASA Astrophysics Data System (ADS)

    Wang, Sixue

    Beginning with the ground-breaking work on DNA computation by Adleman in 1994 [2], the idea of using DNA molecules to perform computations has been explored extensively. In this thesis, a computation based on a scalable DNA neural network was discussed and a neuron model was partially implemented using DNA molecules. In order to understand the behavior of short DNA strands in a polyacrylamide gel, we have measured the mobilities of various short single-stranded DNA (ssDNA) and double-stranded DNA (dsDNA) shorter than 100 bases. We found that sufficiently short lengths of ssDNA had a higher mobility than same lengths of dsDNA, with a crossover length Lx at which the mobilities are equal. The crossover length decreases approximately linearly with polyacrylamide gel acrylamide concentration. At the same time, the influence of DNA structure on its mobility was studied and the effect of single-stranded overhangs on dsDNA was discussed. The idea to make a scalable DNA neural network was discussed. To prepare our basis vector DNA oligomers, a 90 base DNA template with 50 base random strand in the middle and two 20 base primers on the ends was designed and purchased. By a series of dilutions, we obtained several aliquots, containing only 30 random sequence molecules each. These were amplified to roughly 5 pico mole quantities by 38 cycles of PCR with hot start DNA polymerase. We then used asymmetric PCR followed by polyacrylamide gel purification to get the necessary single-stranded basis vectors (ssDNA) and their complements. We tested the suitability of this scheme by adding two vectors formed from different linear of the basis vectors. The full scheme for DNA neural network computation was tested using two determinate ssDNA strands. We successfully transformed an input DNA oligomer to a different output oligomer using the polymerase reaction required by the proposed DNA neural network algorithm. Isothermal linear amplification was used to obtain a sufficient quantity of

  18. Overview of the Scalable Coherent Interface, IEEE STD 1596 (SCI)

    SciTech Connect

    Gustavson, D.B.; James, D.V.; Wiggers, H.A.

    1992-10-01

    The Scalable Coherent Interface standard defines a new generation of interconnection that spans the full range from supercomputer memory `bus` to campus-wide network. SCI provides bus-like services and a shared-memory software model while using an underlying, packet protocol on many independent communication links. Initially these links are 1 GByte/s (wires) and 1 GBit/s (fiber), but the protocol scales well to future faster or lower-cost technologies. The interconnect may use switches, meshes, and rings. The SCI distributed-shared-memory model is simple and versatile, enabling for the first time a smooth integration of highly parallel multiprocessors, workstations, personal computers, I/O, networking and data acquisition.

  19. Suspended in culture--human pluripotent cells for scalable technologies.

    PubMed

    O'Brien, Carmel; Laslett, Andrew L

    2012-09-01

    Human embryonic stem cells (hESCs) and human induced pluripotent stem cells (hiPSCs), collectively termed human pluripotent stem cells (hPSCs), are typically derived and maintained in adherent and semi-defined culture conditions. Recently a number of groups, including Chen et al., 2012, have demonstrated that hESCs can now be expanded efficiently and maintain pluripotency over long-term passaging as aggregates in a serum-free defined suspension culture system, permitting the preparation of scalable cGMP derived hPSC cultures for cell banking, high throughput research programs and clinical applications. In this short commentary we describe the utility and potential future uses of suspension culture systems for hPSCs. PMID:22771716

  20. Center for Programming Models for Scalable Parallel Computing

    SciTech Connect

    John Mellor-Crummey

    2008-02-29

    Rice University's achievements as part of the Center for Programming Models for Scalable Parallel Computing include: (1) design and implemention of cafc, the first multi-platform CAF compiler for distributed and shared-memory machines, (2) performance studies of the efficiency of programs written using the CAF and UPC programming models, (3) a novel technique to analyze explicitly-parallel SPMD programs that facilitates optimization, (4) design, implementation, and evaluation of new language features for CAF, including communication topologies, multi-version variables, and distributed multithreading to simplify development of high-performance codes in CAF, and (5) a synchronization strength reduction transformation for automatically replacing barrier-based synchronization with more efficient point-to-point synchronization. The prototype Co-array Fortran compiler cafc developed in this project is available as open source software from http://www.hipersoft.rice.edu/caf.

  1. Fermilab's multi-petabyte scalable mass storage system

    SciTech Connect

    Oleynik, Gene; Alcorn, Bonnie; Baisley, Wayne; Bakken, Jon; Berg, David; Berman, Eileen; Huang, Chih-Hao; Jones, Terry; Kennedy, Robert D.; Kulyavtsev, Alexander; Moibenko, Alexander; Perelmutov, Timur; Petravick, Don; Podstavkov, Vladimir; Szmuksta, George; Zalokar, Michael; /Fermilab

    2005-01-01

    Fermilab provides a multi-Petabyte scale mass storage system for High Energy Physics (HEP) Experiments and other scientific endeavors. We describe the scalability aspects of the hardware and software architecture that were designed into the Mass Storage System to permit us to scale to multiple petabytes of storage capacity, manage tens of terabytes per day in data transfers, support hundreds of users, and maintain data integrity. We discuss in detail how we scale the system over time to meet the ever-increasing needs of the scientific community, and relate our experiences with many of the technical and economic issues related to scaling the system. Since the 2003 MSST conference, the experiments at Fermilab have generated more than 1.9 PB of additional data. We present results on how this system has scaled and performed for the Fermilab CDF and D0 Run II experiments as well as other HEP experiments and scientific endeavors.

  2. Scalable antifouling reverse osmosis membranes utilizing perfluorophenyl azide photochemistry.

    PubMed

    McVerry, Brian T; Wong, Mavis C Y; Marsh, Kristofer L; Temple, James A T; Marambio-Jones, Catalina; Hoek, Eric M V; Kaner, Richard B

    2014-09-01

    We present a method to produce anti-fouling reverse osmosis (RO) membranes that maintains the process and scalability of current RO membrane manufacturing. Utilizing perfluorophenyl azide (PFPA) photochemistry, commercial reverse osmosis membranes were dipped into an aqueous solution containing PFPA-terminated poly(ethyleneglycol) species and then exposed to ultraviolet light under ambient conditions, a process that can easily be adapted to a roll-to-roll process. Successful covalent modification of commercial reverse osmosis membranes was confirmed with attenuated total reflectance infrared spectroscopy and contact angle measurements. By employing X-ray photoelectron spectroscopy, it was determined that PFPAs undergo UV-generated nitrene addition and bind to the membrane through an aziridine linkage. After modification with the PFPA-PEG derivatives, the reverse osmosis membranes exhibit high fouling-resistance. PMID:25042670

  3. Scalable syntheses of the BET bromodomain inhibitor JQ1

    PubMed Central

    Syeda, Shameem Sultana; Jakkaraj, Sudhakar; Georg, Gunda I.

    2015-01-01

    We have developed methods involving the use of alternate, safer reagents for the scalable syntheses of the potent BET bromodomain inhibitor JQ1. A one-pot three step method, involving the conversion of a benzodiazepine to a thioamde using Lawesson’s reagent, followed by amidrazone formation and installation of the triazole moiety furnished JQ1. This method provides good yields and a facile purification process. For the synthesis of enantiomerically enriched (+)-JQ1, the highly toxic reagent diethyl chlorophosphate, used in a previous synthesis, was replaced with the safer reagent diphenyl chlorophosphate in the three-step one-pot triazole formation without effecting yields and enantiomeric purity of (+)-JQ1. PMID:26034331

  4. A Practical and Scalable Tool to Find Overlaps between Sequences

    PubMed Central

    Haj Rachid, Maan

    2015-01-01

    The evolution of the next generation sequencing technology increases the demand for efficient solutions, in terms of space and time, for several bioinformatics problems. This paper presents a practical and easy-to-implement solution for one of these problems, namely, the all-pairs suffix-prefix problem, using a compact prefix tree. The paper demonstrates an efficient construction of this time-efficient and space-economical tree data structure. The paper presents techniques for parallel implementations of the proposed solution. Experimental evaluation indicates superior results in terms of space and time over existing solutions. Results also show that the proposed technique is highly scalable in a parallel execution environment. PMID:25961045

  5. A Practical and Scalable Tool to Find Overlaps between Sequences.

    PubMed

    Rachid, Maan Haj; Malluhi, Qutaibah

    2015-01-01

    The evolution of the next generation sequencing technology increases the demand for efficient solutions, in terms of space and time, for several bioinformatics problems. This paper presents a practical and easy-to-implement solution for one of these problems, namely, the all-pairs suffix-prefix problem, using a compact prefix tree. The paper demonstrates an efficient construction of this time-efficient and space-economical tree data structure. The paper presents techniques for parallel implementations of the proposed solution. Experimental evaluation indicates superior results in terms of space and time over existing solutions. Results also show that the proposed technique is highly scalable in a parallel execution environment. PMID:25961045

  6. Stability and scalability of piezoelectric flag

    NASA Astrophysics Data System (ADS)

    Wang, Xiaolin; Alben, Silas; Li, Chenyang; Young, Yin Lu

    2015-11-01

    Piezoelectric material (PZT) has drawn enormous attention in the past decades due to its ability to convert mechanical deformation energy into electrical potential energy, and vice versa, and has been applied to energy harvesting and vibration control. In this work, we consider the effect of PZT on the stability of a flexible flag using the inviscid vortex-sheet model. We find that the critical flutter speed is increased due to the extra damping effect of the PZT, and can also be altered by tuning the output inductance-resistance circuit. Optimal resistance and inductance are found to either maximize or minimize the flutter speed. The former application is useful for the vibration control while the latter is important for energy harvesting. We also discuss the scalability of above system to the actual application in air and water.

  7. Scalable problems and memory bounded speedup

    NASA Technical Reports Server (NTRS)

    Sun, Xian-He; Ni, Lionel M.

    1992-01-01

    In this paper three models of parallel speedup are studied. They are fixed-size speedup, fixed-time speedup and memory-bounded speedup. The latter two consider the relationship between speedup and problem scalability. Two sets of speedup formulations are derived for these three models. One set considers uneven workload allocation and communication overhead and gives more accurate estimation. Another set considers a simplified case and provides a clear picture on the impact of the sequential portion of an application on the possible performance gain from parallel processing. The simplified fixed-size speedup is Amdahl's law. The simplified fixed-time speedup is Gustafson's scaled speedup. The simplified memory-bounded speedup contains both Amdahl's law and Gustafson's scaled speedup as special cases. This study leads to a better understanding of parallel processing.

  8. A versatile scalable PET processing system

    SciTech Connect

    H. Dong, A. Weisenberger, J. McKisson, Xi Wenze, C. Cuevas, J. Wilson, L. Zukerman

    2011-06-01

    Positron Emission Tomography (PET) historically has major clinical and preclinical applications in cancerous oncology, neurology, and cardiovascular diseases. Recently, in a new direction, an application specific PET system is being developed at Thomas Jefferson National Accelerator Facility (Jefferson Lab) in collaboration with Duke University, University of Maryland at Baltimore (UMAB), and West Virginia University (WVU) targeted for plant eco-physiology research. The new plant imaging PET system is versatile and scalable such that it could adapt to several plant imaging needs - imaging many important plant organs including leaves, roots, and stems. The mechanical arrangement of the detectors is designed to accommodate the unpredictable and random distribution in space of the plant organs without requiring the plant be disturbed. Prototyping such a system requires a new data acquisition system (DAQ) and data processing system which are adaptable to the requirements of these unique and versatile detectors.

  9. Parallel scalability of Hartree–Fock calculations

    SciTech Connect

    Chow, Edmond Liu, Xing; Smelyanskiy, Mikhail; Hammond, Jeff R.

    2015-03-14

    Quantum chemistry is increasingly performed using large cluster computers consisting of multiple interconnected nodes. For a fixed molecular problem, the efficiency of a calculation usually decreases as more nodes are used, due to the cost of communication between the nodes. This paper empirically investigates the parallel scalability of Hartree–Fock calculations. The construction of the Fock matrix and the density matrix calculation are analyzed separately. For the former, we use a parallelization of Fock matrix construction based on a static partitioning of work followed by a work stealing phase. For the latter, we use density matrix purification from the linear scaling methods literature, but without using sparsity. When using large numbers of nodes for moderately sized problems, density matrix computations are network-bandwidth bound, making purification methods potentially faster than eigendecomposition methods.

  10. Scalable ranked retrieval using document images

    NASA Astrophysics Data System (ADS)

    Jain, Rajiv; Oard, Douglas W.; Doermann, David

    2013-12-01

    Despite the explosion of text on the Internet, hard copy documents that have been scanned as images still play a significant role for some tasks. The best method to perform ranked retrieval on a large corpus of document images, however, remains an open research question. The most common approach has been to perform text retrieval using terms generated by optical character recognition. This paper, by contrast, examines whether a scalable segmentation-free image retrieval algorithm, which matches sub-images containing text or graphical objects, can provide additional benefit in satisfying a user's information needs on a large, real world dataset. Results on 7 million scanned pages from the CDIP v1.0 test collection show that content based image retrieval finds a substantial number of documents that text retrieval misses, and that when used as a basis for relevance feedback can yield improvements in retrieval effectiveness.

  11. iSIGHT-FD scalability test report.

    SciTech Connect

    Clay, Robert L.; Shneider, Max S.

    2008-07-01

    The engineering analysis community at Sandia National Laboratories uses a number of internal and commercial software codes and tools, including mesh generators, preprocessors, mesh manipulators, simulation codes, post-processors, and visualization packages. We define an analysis workflow as the execution of an ordered, logical sequence of these tools. Various forms of analysis (and in particular, methodologies that use multiple function evaluations or samples) involve executing parameterized variations of these workflows. As part of the DART project, we are evaluating various commercial workflow management systems, including iSIGHT-FD from Engineous. This report documents the results of a scalability test that was driven by DAKOTA and conducted on a parallel computer (Thunderbird). The purpose of this experiment was to examine the suitability and performance of iSIGHT-FD for large-scale, parameterized analysis workflows. As the results indicate, we found iSIGHT-FD to be suitable for this type of application.

  12. Porphyrins as Catalysts in Scalable Organic Reactions.

    PubMed

    Barona-Castaño, Juan C; Carmona-Vargas, Christian C; Brocksom, Timothy J; de Oliveira, Kleber T

    2016-01-01

    Catalysis is a topic of continuous interest since it was discovered in chemistry centuries ago. Aiming at the advance of reactions for efficient processes, a number of approaches have been developed over the last 180 years, and more recently, porphyrins occupy an important role in this field. Porphyrins and metalloporphyrins are fascinating compounds which are involved in a number of synthetic transformations of great interest for industry and academy. The aim of this review is to cover the most recent progress in reactions catalysed by porphyrins in scalable procedures, thus presenting the state of the art in reactions of epoxidation, sulfoxidation, oxidation of alcohols to carbonyl compounds and C-H functionalization. In addition, the use of porphyrins as photocatalysts in continuous flow processes is covered. PMID:27005601

  13. BASSET: Scalable Gateway Finder in Large Graphs

    SciTech Connect

    Tong, H; Papadimitriou, S; Faloutsos, C; Yu, P S; Eliassi-Rad, T

    2010-11-03

    Given a social network, who is the best person to introduce you to, say, Chris Ferguson, the poker champion? Or, given a network of people and skills, who is the best person to help you learn about, say, wavelets? The goal is to find a small group of 'gateways': persons who are close enough to us, as well as close enough to the target (person, or skill) or, in other words, are crucial in connecting us to the target. The main contributions are the following: (a) we show how to formulate this problem precisely; (b) we show that it is sub-modular and thus it can be solved near-optimally; (c) we give fast, scalable algorithms to find such gateways. Experiments on real data sets validate the effectiveness and efficiency of the proposed methods, achieving up to 6,000,000x speedup.

  14. Toward Scalable Benchmarks for Mass Storage Systems

    NASA Technical Reports Server (NTRS)

    Miller, Ethan L.

    1996-01-01

    This paper presents guidelines for the design of a mass storage system benchmark suite, along with preliminary suggestions for programs to be included. The benchmarks will measure both peak and sustained performance of the system as well as predicting both short- and long-term behavior. These benchmarks should be both portable and scalable so they may be used on storage systems from tens of gigabytes to petabytes or more. By developing a standard set of benchmarks that reflect real user workload, we hope to encourage system designers and users to publish performance figures that can be compared with those of other systems. This will allow users to choose the system that best meets their needs and give designers a tool with which they can measure the performance effects of improvements to their systems.

  15. Scalable graphene production: perspectives and challenges of plasma applications

    NASA Astrophysics Data System (ADS)

    Levchenko, Igor; Ostrikov, Kostya (Ken); Zheng, Jie; Li, Xingguo; Keidar, Michael; B. K. Teo, Kenneth

    2016-05-01

    Graphene, a newly discovered and extensively investigated material, has many unique and extraordinary properties which promise major technological advances in fields ranging from electronics to mechanical engineering and food production. Unfortunately, complex techniques and high production costs hinder commonplace applications. Scaling of existing graphene production techniques to the industrial level without compromising its properties is a current challenge. This article focuses on the perspectives and challenges of scalability, equipment, and technological perspectives of the plasma-based techniques which offer many unique possibilities for the synthesis of graphene and graphene-containing products. The plasma-based processes are amenable for scaling and could also be useful to enhance the controllability of the conventional chemical vapour deposition method and some other techniques, and to ensure a good quality of the produced graphene. We examine the unique features of the plasma-enhanced graphene production approaches, including the techniques based on inductively-coupled and arc discharges, in the context of their potential scaling to mass production following the generic scaling approaches applicable to the existing processes and systems. This work analyses a large amount of the recent literature on graphene production by various techniques and summarizes the results in a tabular form to provide a simple and convenient comparison of several available techniques. Our analysis reveals a significant potential of scalability for plasma-based technologies, based on the scaling-related process characteristics. Among other processes, a greater yield of 1 g × h-1 m-2 was reached for the arc discharge technology, whereas the other plasma-based techniques show process yields comparable to the neutral-gas based methods. Selected plasma-based techniques show lower energy consumption than in thermal CVD processes, and the ability to produce graphene flakes of various

  16. Scalable graphene production: perspectives and challenges of plasma applications.

    PubMed

    Levchenko, Igor; Ostrikov, Kostya Ken; Zheng, Jie; Li, Xingguo; Keidar, Michael; B K Teo, Kenneth

    2016-05-19

    Graphene, a newly discovered and extensively investigated material, has many unique and extraordinary properties which promise major technological advances in fields ranging from electronics to mechanical engineering and food production. Unfortunately, complex techniques and high production costs hinder commonplace applications. Scaling of existing graphene production techniques to the industrial level without compromising its properties is a current challenge. This article focuses on the perspectives and challenges of scalability, equipment, and technological perspectives of the plasma-based techniques which offer many unique possibilities for the synthesis of graphene and graphene-containing products. The plasma-based processes are amenable for scaling and could also be useful to enhance the controllability of the conventional chemical vapour deposition method and some other techniques, and to ensure a good quality of the produced graphene. We examine the unique features of the plasma-enhanced graphene production approaches, including the techniques based on inductively-coupled and arc discharges, in the context of their potential scaling to mass production following the generic scaling approaches applicable to the existing processes and systems. This work analyses a large amount of the recent literature on graphene production by various techniques and summarizes the results in a tabular form to provide a simple and convenient comparison of several available techniques. Our analysis reveals a significant potential of scalability for plasma-based technologies, based on the scaling-related process characteristics. Among other processes, a greater yield of 1 g × h(-1) m(-2) was reached for the arc discharge technology, whereas the other plasma-based techniques show process yields comparable to the neutral-gas based methods. Selected plasma-based techniques show lower energy consumption than in thermal CVD processes, and the ability to produce graphene flakes of

  17. Fully scalable video transmission using the SSM adaptation framework

    NASA Astrophysics Data System (ADS)

    Mukherjee, Debargha; Chen, Peisong; Hsiang, Shih-Ta; Woods, John W.; Said, Amir

    2003-06-01

    Recently a methodology for representation and adaptation of arbitrary scalable bit-streams in a fully content non-specific manner has been proposed on the basis of a universal model for all scalable bit-streams called Scalable Structured Meta-formats (SSM). According to this model, elementary scalable bit-streams are naturally organized in a symmetric multi-dimensional logical structure. The model parameters for a specific bit-stream along with information guiding decision-making among possible adaptation choices are represented in a binary or XML descriptor to accompany the bit-stream flowing downstream. The capabilities and preferences of receiving terminals flow upstream and are also specified in binary or XML form to represent constraints that guide adaptation. By interpreting the descriptor and the constraint specifications, a universal adaptation engine sitting on a network node can adapt the content appropriately to suit the specified needs and preferences of recipients, without knowledge of the specifics of the content, its encoding and/or encryption. In this framework, different adaptation infrastructures are no longer needed for different types of scalable media. In this work, we show how this framework can be used to adapt fully scalable video bit-streams, specifically ones obtained by the fully scalable MC-EZBC video coding system. MC-EZBC uses a 3-D subband/wavelet transform that exploits correlation by filtering along motion trajectories, to obtain a 3-dimensional scalable bit-stream combining temporal, spatial and SNR scalability in a compact bit-stream. Several adaptation use cases are presented to demonstrate the flexibility and advantages of a fully scalable video bit-stream when used in conjunction with a network adaptation engine for transmission.

  18. An efficient scalable intra coding algorithm for spatial scalability in enhancement layer

    NASA Astrophysics Data System (ADS)

    Wang, Zhang; Lu, Lijun

    2011-05-01

    Scalable video coding (SVC) is attractive due to the capability of reconstructing lower resolution or lower quality signals from partial bit streams, which allows for simple solutions adaptted to network and terminal capabilities. This article addresses the spatial scalability of SVC and proposes an efficient H.264-based scalable intra coding algorithm. In comparison with precious single layer intra prediction (SLIP) method, the proposed algorithm aims to improve the intra coding performance of the enhancement layer by a new inter layer intra prediction (ILIP) method. The main idea of ILIP is that up-sampled and reconstructed pixels of the base layer are very useful to predict and encode those pixels of the enhancement layer, especially when those neighbouring pixels are not available. Experimental results show that the peak signal to noise ratio (PSNR) data of luminance component of encoded frames are improved, and both bit-rates and computation complexity are maintained very well. For sequence Football, the average increase of PSNR is up to 0.21 dB, while for Foreman and Bus, they are 0.14 dB and 0.17 dB, respectively.

  19. LVFS: A Scalable Petabye/Exabyte Data Storage System

    NASA Astrophysics Data System (ADS)

    Golpayegani, N.; Halem, M.; Masuoka, E. J.; Ye, G.; Devine, N. K.

    2013-12-01

    . The LVFS system replaces the NFS disk mounting approach of LAADS and utilizes the already existing highly optimized metadata database server, which is applicable to most scientific big data intensive compute systems. Thus, LVFS ties the existing storage system with the existing metadata infrastructure system which we believe leads to a scalable exabyte virtual file system. The uniqueness of the implemented design is not limited to LAADS but can be employed with most scientific data processing systems. By utilizing the Filesystem In Userspace (FUSE), a kernel module available in many operating systems, LVFS was able to replace the NFS system while staying POSIX compliant. As a result, the LVFS system becomes scalable to exabyte sizes owing to the use of highly scalable database servers optimized for metadata storage. The flexibility of the LVFS design allows it to organize data on the fly in different ways, such as by region, date, instrument or product without the need for duplication, symbolic links, or any other replication methods. We proposed here a strategic reference architecture that addresses the inefficiencies of scientific petabyte/exabyte file system access through the dynamic integration of the observing system's large metadata file.

  20. Scalable Photogrammetric Motion Capture System "mosca": Development and Application

    NASA Astrophysics Data System (ADS)

    Knyaz, V. A.

    2015-05-01

    Wide variety of applications (from industrial to entertainment) has a need for reliable and accurate 3D information about motion of an object and its parts. Very often the process of movement is rather fast as in cases of vehicle movement, sport biomechanics, animation of cartoon characters. Motion capture systems based on different physical principles are used for these purposes. The great potential for obtaining high accuracy and high degree of automation has vision-based system due to progress in image processing and analysis. Scalable inexpensive motion capture system is developed as a convenient and flexible tool for solving various tasks requiring 3D motion analysis. It is based on photogrammetric techniques of 3D measurements and provides high speed image acquisition, high accuracy of 3D measurements and highly automated processing of captured data. Depending on the application the system can be easily modified for different working areas from 100 mm to 10 m. The developed motion capture system uses from 2 to 4 technical vision cameras for video sequences of object motion acquisition. All cameras work in synchronization mode at frame rate up to 100 frames per second under the control of personal computer providing the possibility for accurate calculation of 3D coordinates of interest points. The system was used for a set of different applications fields and demonstrated high accuracy and high level of automation.

  1. Scalable desktop visualisation of very large radio astronomy data cubes

    NASA Astrophysics Data System (ADS)

    Perkins, Simon; Questiaux, Jacques; Finniss, Stephen; Tyler, Robin; Blyth, Sarah; Kuttel, Michelle M.

    2014-07-01

    Observation data from radio telescopes is typically stored in three (or higher) dimensional data cubes, the resolution, coverage and size of which continues to grow as ever larger radio telescopes come online. The Square Kilometre Array, tabled to be the largest radio telescope in the world, will generate multi-terabyte data cubes - several orders of magnitude larger than the current norm. Despite this imminent data deluge, scalable approaches to file access in Astronomical visualisation software are rare: most current software packages cannot read astronomical data cubes that do not fit into computer system memory, or else provide access only at a serious performance cost. In addition, there is little support for interactive exploration of 3D data. We describe a scalable, hierarchical approach to 3D visualisation of very large spectral data cubes to enable rapid visualisation of large data files on standard desktop hardware. Our hierarchical approach, embodied in the AstroVis prototype, aims to provide a means of viewing large datasets that do not fit into system memory. The focus is on rapid initial response: our system initially rapidly presents a reduced, coarse-grained 3D view of the data cube selected, which is gradually refined. The user may select sub-regions of the cube to be explored in more detail, or extracted for use in applications that do not support large files. We thus shift the focus from data analysis informed by narrow slices of detailed information, to analysis informed by overview information, with details on demand. Our hierarchical solution to the rendering of large data cubes reduces the overall time to complete file reading, provides user feedback during file processing and is memory efficient. This solution does not require high performance computing hardware and can be implemented on any platform supporting the OpenGL rendering library.

  2. Scalable Conjunction Processing using Spatiotemporally Indexed Ephemeris Data

    NASA Astrophysics Data System (ADS)

    Budianto-Ho, I.; Johnson, S.; Sivilli, R.; Alberty, C.; Scarberry, R.

    2014-09-01

    The collision warnings produced by the Joint Space Operations Center (JSpOC) are of critical importance in protecting U.S. and allied spacecraft against destructive collisions and protecting the lives of astronauts during space flight. As the Space Surveillance Network (SSN) improves its sensor capabilities for tracking small and dim space objects, the number of tracked objects increases from thousands to hundreds of thousands of objects, while the number of potential conjunctions increases with the square of the number of tracked objects. Classical filtering techniques such as apogee and perigee filters have proven insufficient. Novel and orders of magnitude faster conjunction analysis algorithms are required to find conjunctions in a timely manner. Stellar Science has developed innovative filtering techniques for satellite conjunction processing using spatiotemporally indexed ephemeris data that efficiently and accurately reduces the number of objects requiring high-fidelity and computationally-intensive conjunction analysis. Two such algorithms, one based on the k-d Tree pioneered in robotics applications and the other based on Spatial Hash Tables used in computer gaming and animation, use, at worst, an initial O(N log N) preprocessing pass (where N is the number of tracked objects) to build large O(N) spatial data structures that substantially reduce the required number of O(N^2) computations, substituting linear memory usage for quadratic processing time. The filters have been implemented as Open Services Gateway initiative (OSGi) plug-ins for the Continuous Anomalous Orbital Situation Discriminator (CAOS-D) conjunction analysis architecture. We have demonstrated the effectiveness, efficiency, and scalability of the techniques using a catalog of 100,000 objects, an analysis window of one day, on a 64-core computer with 1TB shared memory. Each algorithm can process the full catalog in 6 minutes or less, almost a twenty-fold performance improvement over the

  3. ParaText : scalable text modeling and analysis.

    SciTech Connect

    Dunlavy, Daniel M.; Stanton, Eric T.; Shead, Timothy M.

    2010-06-01

    Automated processing, modeling, and analysis of unstructured text (news documents, web content, journal articles, etc.) is a key task in many data analysis and decision making applications. As data sizes grow, scalability is essential for deep analysis. In many cases, documents are modeled as term or feature vectors and latent semantic analysis (LSA) is used to model latent, or hidden, relationships between documents and terms appearing in those documents. LSA supplies conceptual organization and analysis of document collections by modeling high-dimension feature vectors in many fewer dimensions. While past work on the scalability of LSA modeling has focused on the SVD, the goal of our work is to investigate the use of distributed memory architectures for the entire text analysis process, from data ingestion to semantic modeling and analysis. ParaText is a set of software components for distributed processing, modeling, and analysis of unstructured text. The ParaText source code is available under a BSD license, as an integral part of the Titan toolkit. ParaText components are chained-together into data-parallel pipelines that are replicated across processes on distributed-memory architectures. Individual components can be replaced or rewired to explore different computational strategies and implement new functionality. ParaText functionality can be embedded in applications on any platform using the native C++ API, Python, or Java. The ParaText MPI Process provides a 'generic' text analysis pipeline in a command-line executable that can be used for many serial and parallel analysis tasks. ParaText can also be deployed as a web service accessible via a RESTful (HTTP) API. In the web service configuration, any client can access the functionality provided by ParaText using commodity protocols ... from standard web browsers to custom clients written in any language.

  4. A Scalable Framework to Detect Personal Health Mentions on Twitter

    PubMed Central

    Fabbri, Daniel; Rosenbloom, S Trent

    2015-01-01

    Background Biomedical research has traditionally been conducted via surveys and the analysis of medical records. However, these resources are limited in their content, such that non-traditional domains (eg, online forums and social media) have an opportunity to supplement the view of an individual’s health. Objective The objective of this study was to develop a scalable framework to detect personal health status mentions on Twitter and assess the extent to which such information is disclosed. Methods We collected more than 250 million tweets via the Twitter streaming API over a 2-month period in 2014. The corpus was filtered down to approximately 250,000 tweets, stratified across 34 high-impact health issues, based on guidance from the Medical Expenditure Panel Survey. We created a labeled corpus of several thousand tweets via a survey, administered over Amazon Mechanical Turk, that documents when terms correspond to mentions of personal health issues or an alternative (eg, a metaphor). We engineered a scalable classifier for personal health mentions via feature selection and assessed its potential over the health issues. We further investigated the utility of the tweets by determining the extent to which Twitter users disclose personal health status. Results Our investigation yielded several notable findings. First, we find that tweets from a small subset of the health issues can train a scalable classifier to detect health mentions. Specifically, training on 2000 tweets from four health issues (cancer, depression, hypertension, and leukemia) yielded a classifier with precision of 0.77 on all 34 health issues. Second, Twitter users disclosed personal health status for all health issues. Notably, personal health status was disclosed over 50% of the time for 11 out of 34 (33%) investigated health issues. Third, the disclosure rate was dependent on the health issue in a statistically significant manner (P<.001). For instance, more than 80% of the tweets about

  5. Scalable Algorithms for Unsupervised Classification and Anomaly Detection in Large Geospatiotemporal Data Sets

    NASA Astrophysics Data System (ADS)

    Mills, R. T.; Hoffman, F. M.; Kumar, J.

    2015-12-01

    The increasing availability of high-resolution geospatiotemporal datasets from sources such as observatory networks, remote sensing platforms, and computational Earth system models has opened new possibilities for knowledge discovery and mining of ecological data sets fused from disparate sources. Traditional algorithms and computing platforms are impractical for the analysis and synthesis of data sets of this size; however, new algorithmic approaches that can effectively utilize the complex memory hierarchies and the extremely high levels of available parallelism in state-of-the-art high-performance computing platforms can enable such analysis. We describe some unsupervised knowledge discovery and anomaly detection approaches based on highly scalable parallel algorithms for k-means clustering and singular value decomposition, consider a few practical applications thereof to the analysis of climatic and remotely-sensed vegetation phenology data sets, and speculate on some of the new applications that such scalable analysis methods may enable.

  6. On the scalability of parallel genetic algorithms.

    PubMed

    Cantú-Paz, E; Goldberg, D E

    1999-01-01

    This paper examines the scalability of several types of parallel genetic algorithms (GAs). The objective is to determine the optimal number of processors that can be used by each type to minimize the execution time. The first part of the paper considers algorithms with a single population. The investigation focuses on an implementation where the population is distributed to several processors, but the results are applicable to more common master-slave implementations, where the population is entirely stored in a master processor and multiple slaves are used to evaluate the fitness. The second part of the paper deals with parallel GAs with multiple populations. It first considers a bounding case where the connectivity, the migration rate, and the frequency of migrations are set to their maximal values. Then, arbitrary regular topologies with lower migration rates are considered and the frequency of migrations is set to its lowest value. The investigationis mainly theoretical, but experimental evidence with an additively-decomposable function is included to illustrate the accuracy of the theory. In all cases, the calculations show that the optimal number of processors that minimizes the execution time is directly proportional to the square root of the population size and the fitness evaluation time. Since these two factors usually increase as the domain becomes more difficult, the results of the paper suggest that parallel GAs can integrate large numbers of processors and significantly reduce the execution time of many practical applications. PMID:10578030

  7. Quantum Information Processing using Scalable Techniques

    NASA Astrophysics Data System (ADS)

    Hanneke, D.; Bowler, R.; Jost, J. D.; Home, J. P.; Lin, Y.; Tan, T.-R.; Leibfried, D.; Wineland, D. J.

    2011-05-01

    We report progress towards improving our previous demonstrations that combined all the fundamental building blocks required for scalable quantum information processing using trapped atomic ions. Included elements are long-lived qubits; a laser-induced universal gate set; state initialization and readout; and information transport, including co-trapping a second ion species to reinitialize motion without qubit decoherence. Recent efforts have focused on reducing experimental overhead and increasing gate fidelity. Most of the experimental duty cycle was previously used for transport, separation, and recombination of ion chains as well as re-cooling of motional excitation. We have addressed these issues by developing and implementing an arbitrary waveform generator with an update rate far above the ions' motional frequencies. To reduce gate errors, we actively stabilize the position of several UV (313 nm) laser beams. We have also switched the two-qubit entangling gate to one that acts directly on 9Be+ hyperfine qubit states whose energy separation is magnetic-fluctuation insensitive. This work is supported by DARPA, NSA, ONR, IARPA, Sandia, and the NIST Quantum Information Program.

  8. SCTP as scalable video coding transport

    NASA Astrophysics Data System (ADS)

    Ortiz, Jordi; Graciá, Eduardo Martínez; Skarmeta, Antonio F.

    2013-12-01

    This study presents an evaluation of the Stream Transmission Control Protocol (SCTP) for the transport of the scalable video codec (SVC), proposed by MPEG as an extension to H.264/AVC. Both technologies fit together properly. On the one hand, SVC permits to split easily the bitstream into substreams carrying different video layers, each with different importance for the reconstruction of the complete video sequence at the receiver end. On the other hand, SCTP includes features, such as the multi-streaming and multi-homing capabilities, that permit to transport robustly and efficiently the SVC layers. Several transmission strategies supported on baseline SCTP and its concurrent multipath transfer (CMT) extension are compared with the classical solutions based on the Transmission Control Protocol (TCP) and the Realtime Transmission Protocol (RTP). Using ns-2 simulations, it is shown that CMT-SCTP outperforms TCP and RTP in error-prone networking environments. The comparison is established according to several performance measurements, including delay, throughput, packet loss, and peak signal-to-noise ratio of the received video.

  9. Scalable histopathological image analysis via active learning.

    PubMed

    Zhu, Yan; Zhang, Shaoting; Liu, Wei; Metaxas, Dimitris N

    2014-01-01

    Training an effective and scalable system for medical image analysis usually requires a large amount of labeled data, which incurs a tremendous annotation burden for pathologists. Recent progress in active learning can alleviate this issue, leading to a great reduction on the labeling cost without sacrificing the predicting accuracy too much. However, most existing active learning methods disregard the "structured information" that may exist in medical images (e.g., data from individual patients), and make a simplifying assumption that unlabeled data is independently and identically distributed. Both may not be suitable for real-world medical images. In this paper, we propose a novel batch-mode active learning method which explores and leverages such structured information in annotations of medical images to enforce diversity among the selected data, therefore maximizing the information gain. We formulate the active learning problem as an adaptive submodular function maximization problem subject to a partition matroid constraint, and further present an efficient greedy algorithm to achieve a good solution with a theoretically proven bound. We demonstrate the efficacy of our algorithm on thousands of histopathological images of breast microscopic tissues. PMID:25320821

  10. Scalability and interoperability within glideinWMS

    SciTech Connect

    Bradley, D.; Sfiligoi, I.; Padhi, S.; Frey, J.; Tannenbaum, T.; /Wisconsin U., Madison

    2010-01-01

    Physicists have access to thousands of CPUs in grid federations such as OSG and EGEE. With the start-up of the LHC, it is essential for individuals or groups of users to wrap together available resources from multiple sites across multiple grids under a higher user-controlled layer in order to provide a homogeneous pool of available resources. One such system is glideinWMS, which is based on the Condor batch system. A general discussion of glideinWMS can be found elsewhere. Here, we focus on recent advances in extending its reach: scalability and integration of heterogeneous compute elements. We demonstrate that the new developments exceed the design goal of over 10,000 simultaneous running jobs under a single Condor schedd, using strong security protocols across global networks, and sustaining a steady-state job completion rate of a few Hz. We also show interoperability across heterogeneous computing elements achieved using client-side methods. We discuss this technique and the challenges in direct access to NorduGrid and CREAM compute elements, in addition to Globus based systems.

  11. Improving the Performance Scalability of the Community Atmosphere Model

    SciTech Connect

    Mirin, Arthur; Worley, Patrick H

    2012-01-01

    The Community Atmosphere Model (CAM), which serves as the atmosphere component of the Community Climate System Model (CCSM), is the most computationally expensive CCSM component in typical configurations. On current and next-generation leadership class computing systems, the performance of CAM is tied to its parallel scalability. Improving performance scalability in CAM has been a challenge, due largely to algorithmic restrictions necessitated by the polar singularities in its latitude-longitude computational grid. Nevertheless, through a combination of exploiting additional parallelism, implementing improved communication protocols, and eliminating scalability bottlenecks, we have been able to more than double the maximum throughput rate of CAM on production platforms. We describe these improvements and present results on the Cray XT5 and IBM BG/P. The approaches taken are not specific to CAM and may inform similar scalability enhancement activities for other codes.

  12. Scalable Geiger/APD/PIN multi-channel sensing platform

    NASA Astrophysics Data System (ADS)

    Buckley, Steve; Bellis, Stephen; Wilcock, Reuben; Mathewson, Alan; Jackson, J. Carlton

    2006-02-01

    Previous generation low light detection platforms have been based on the photomultiplier tube (PMT) or the silicon single photon counting module (SPCM) from Perkin Elmer1. A new generation of silicon CMOS compatible photon counting sensors are being developed offering high quantum efficiency, low operating voltage, high levels of robustness and compatibility with CMOS processing for integration into large format imaging arrays. This latest generation yields a new detector for emerging applications which demand photon counting performance providing high performance and flexibility not possible to date. We describe a 4-channel photon detection platform, which allows the use of 4 separate photon counting detectors in either free space or fibre-coupled mode. The platform is scalable up to 16 channels with plug in modules allowing active quenching or Peltier cooling as required. A graphical user interface allows feedback and control of all device parameters. We show a novel ability to integrate separate detection modules to extend the dynamic range of the system. This allows a PIN or APD mode detector to be used alongside sensitive photon counting detectors. An advanced FPGA and microcontroller interface has been designed which allows simultaneous time binning of counting rates and readout of the analog signals when used with linear detectors. This new architecture will be discussed, presenting a full characterization of count rate, quantum efficiency, time binning and sensitivity across the broad spectrum of light flux applicable to PIN diodes, APDs and Geiger-mode photon counting sensors.

  13. TriG: Next Generation Scalable Spaceborne GNSS Receiver

    NASA Technical Reports Server (NTRS)

    Tien, Jeffrey Y.; Okihiro, Brian Bachman; Esterhuizen, Stephan X.; Franklin, Garth W.; Meehan, Thomas K.; Munson, Timothy N.; Robison, David E.; Turbiner, Dmitry; Young, Lawrence E.

    2012-01-01

    TriG is the next generation NASA scalable space GNSS Science Receiver. It will track all GNSS and additional signals (i.e. GPS, GLONASS, Galileo, Compass and Doris). Scalable 3U architecture and fully software and firmware recofigurable, enabling optimization to meet specific mission requirements. TriG GNSS EM is currently undergoing testing and is expected to complete full performance testing later this year.

  14. MediAgent: a WWW-based scalable and self-learning medical search engine.

    PubMed Central

    Tay, J.; Ke, S.; Lun, K. C.

    1998-01-01

    Searching for medical information on the Internet can be tedious and frustrating due to the number of irrelevant entries returned from generic search engines. We have developed MediAgent, a scalable search engine that aims to deliver a web-based medical search solution which is focused, exhaustive and able to keep improving its databases. The software package can run off a single low-end system and be scaled into a client-server, distributed computing architecture for high-end needs. This scalable architecture boosts MediAgent's handling capacity to tens of millions of web pages. In addition to large volume handling, MediAgent is designed to be manageable. All subsystems are not only highly configurable, but also support remote, interactive management and monitoring by the system administrator. PMID:9929289

  15. High-Power Zinc-Air Energy Storage: Enhanced Metal-Air Energy Storage System with Advanced Grid-Interoperable Power Electronics Enabling Scalability and Ultra-Low Cost

    SciTech Connect

    2010-10-01

    GRIDS Project: Fluidic is developing a low-cost, rechargeable, high-power module for Zinc-air batteries that will be used to store renewable energy. Zinc-air batteries are traditionally found in small, non-rechargeable devices like hearing aids because they are well-suited to delivering low levels of power for long periods of time. Historically, Zinc-air batteries have not been as useful for applications which require periodic bursts of power, like on the electrical grid. Fluidic hopes to fill this need by combining the high energy, low cost, and long run-time of a Zinc-air battery with new chemistry providing high power, high efficiency, and fast response. The battery module could allow large grid-storage batteries to provide much more power on very short demand—the most costly kind of power for utilities—and with much more versatile performance.

  16. Scalable tensor factorizations with missing data.

    SciTech Connect

    Morup, Morten; Dunlavy, Daniel M.; Acar, Evrim; Kolda, Tamara Gibson

    2010-04-01

    The problem of missing data is ubiquitous in domains such as biomedical signal processing, network traffic analysis, bibliometrics, social network analysis, chemometrics, computer vision, and communication networks|all domains in which data collection is subject to occasional errors. Moreover, these data sets can be quite large and have more than two axes of variation, e.g., sender, receiver, time. Many applications in those domains aim to capture the underlying latent structure of the data; in other words, they need to factorize data sets with missing entries. If we cannot address the problem of missing data, many important data sets will be discarded or improperly analyzed. Therefore, we need a robust and scalable approach for factorizing multi-way arrays (i.e., tensors) in the presence of missing data. We focus on one of the most well-known tensor factorizations, CANDECOMP/PARAFAC (CP), and formulate the CP model as a weighted least squares problem that models only the known entries. We develop an algorithm called CP-WOPT (CP Weighted OPTimization) using a first-order optimization approach to solve the weighted least squares problem. Based on extensive numerical experiments, our algorithm is shown to successfully factor tensors with noise and up to 70% missing data. Moreover, our approach is significantly faster than the leading alternative and scales to larger problems. To show the real-world usefulness of CP-WOPT, we illustrate its applicability on a novel EEG (electroencephalogram) application where missing data is frequently encountered due to disconnections of electrodes.

  17. Microscopic Characterization of Scalable Coherent Rydberg Superatoms

    NASA Astrophysics Data System (ADS)

    Zeiher, Johannes; Schauß, Peter; Hild, Sebastian; Macrı, Tommaso; Bloch, Immanuel; Gross, Christian

    2015-07-01

    Strong interactions can amplify quantum effects such that they become important on macroscopic scales. Controlling these coherently on a single-particle level is essential for the tailored preparation of strongly correlated quantum systems and opens up new prospects for quantum technologies. Rydberg atoms offer such strong interactions, which lead to extreme nonlinearities in laser-coupled atomic ensembles. As a result, multiple excitation of a micrometer-sized cloud can be blocked while the light-matter coupling becomes collectively enhanced. The resulting two-level system, often called a "superatom," is a valuable resource for quantum information, providing a collective qubit. Here, we report on the preparation of 2 orders of magnitude scalable superatoms utilizing the large interaction strength provided by Rydberg atoms combined with precise control of an ensemble of ultracold atoms in an optical lattice. The latter is achieved with sub-shot-noise precision by local manipulation of a two-dimensional Mott insulator. We microscopically confirm the superatom picture by in situ detection of the Rydberg excitations and observe the characteristic square-root scaling of the optical coupling with the number of atoms. Enabled by the full control over the atomic sample, including the motional degrees of freedom, we infer the overlap of the produced many-body state with a W state from the observed Rabi oscillations and deduce the presence of entanglement. Finally, we investigate the breakdown of the superatom picture when two Rydberg excitations are present in the system, which leads to dephasing and a loss of coherence.

  18. Myria: Scalable Analytics as a Service

    NASA Astrophysics Data System (ADS)

    Howe, B.; Halperin, D.; Whitaker, A.

    2014-12-01

    At the UW eScience Institute, we're working to empower non-experts, especially in the sciences, to write and use data-parallel algorithms. To this end, we are building Myria, a web-based platform for scalable analytics and data-parallel programming. Myria's internal model of computation is the relational algebra extended with iteration, such that every program is inherently data-parallel, just as every query in a database is inherently data-parallel. But unlike databases, iteration is a first class concept, allowing us to express machine learning tasks, graph traversal tasks, and more. Programs can be expressed in a number of languages and can be executed on a number of execution environments, but we emphasize a particular language called MyriaL that supports both imperative and declarative styles and a particular execution engine called MyriaX that uses an in-memory column-oriented representation and asynchronous iteration. We deliver Myria over the web as a service, providing an editor, performance analysis tools, and catalog browsing features in a single environment. We find that this web-based "delivery vector" is critical in reaching non-experts: they are insulated from irrelevant effort technical work associated with installation, configuration, and resource management. The MyriaX backend, one of several execution runtimes we support, is a main-memory, column-oriented, RDBMS-on-the-worker system that supports cyclic data flows as a first-class citizen and has been shown to outperform competitive systems on 100-machine cluster sizes. I will describe the Myria system, give a demo, and present some new results in large-scale oceanographic microbiology.

  19. Multimodal optogenetic neural interfacing device fabricated by scalable optical fiber drawing technique.

    PubMed

    Davey, Christopher J; Argyros, Alexander; Fleming, Simon C; Solomon, Samuel G

    2015-12-01

    We present a novel approach to the design and manufacture of optrodes for use in the biomedical research field of optogenetic neural interfacing. Using recently developed optical fiber drawing techniques that involve co-drawing metal/polymer composite fiber, we have assembled and characterized a novel optrode with promising optical and electrical functionality. The fabrication technique is flexible, scalable, and amenable to extension to implantable optrodes with high-density arrays of multiple electrodes, waveguides, and drug delivery channels. PMID:26836662

  20. GASPRNG: GPU accelerated scalable parallel random number generator library

    NASA Astrophysics Data System (ADS)

    Gao, Shuang; Peterson, Gregory D.

    2013-04-01

    Graphics processors represent a promising technology for accelerating computational science applications. Many computational science applications require fast and scalable random number generation with good statistical properties, so they use the Scalable Parallel Random Number Generators library (SPRNG). We present the GPU Accelerated SPRNG library (GASPRNG) to accelerate SPRNG in GPU-based high performance computing systems. GASPRNG includes code for a host CPU and CUDA code for execution on NVIDIA graphics processing units (GPUs) along with a programming interface to support various usage models for pseudorandom numbers and computational science applications executing on the CPU, GPU, or both. This paper describes the implementation approach used to produce high performance and also describes how to use the programming interface. The programming interface allows a user to be able to use GASPRNG the same way as SPRNG on traditional serial or parallel computers as well as to develop tightly coupled programs executing primarily on the GPU. We also describe how to install GASPRNG and use it. To help illustrate linking with GASPRNG, various demonstration codes are included for the different usage models. GASPRNG on a single GPU shows up to 280x speedup over SPRNG on a single CPU core and is able to scale for larger systems in the same manner as SPRNG. Because GASPRNG generates identical streams of pseudorandom numbers as SPRNG, users can be confident about the quality of GASPRNG for scalable computational science applications. Catalogue identifier: AEOI_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEOI_v1_0.html Program obtainable from: CPC Program Library, Queen’s University, Belfast, N. Ireland Licensing provisions: UTK license. No. of lines in distributed program, including test data, etc.: 167900 No. of bytes in distributed program, including test data, etc.: 1422058 Distribution format: tar.gz Programming language: C and CUDA. Computer: Any PC or

  1. Jumping-Droplet-Enhanced Condensation on Scalable Superhydrophobic Nanostructured Surfaces

    SciTech Connect

    Miljkovic, N; Enright, R; Nam, Y; Lopez, K; Dou, N; Sack, J; Wang, E

    2013-01-09

    When droplets coalesce on a superhydrophobic nanostructured surface, the resulting droplet can jump from the surface due to the release of excess surface energy. If designed properly, these superhydrophobic nanostructured surfaces can not only allow for easy droplet removal at micrometric length scales during condensation but also promise to enhance heat transfer performance. However, the rationale for the design of an ideal nanostructured surface as well as heat transfer experiments demonstrating the advantage of this jumping behavior are lacking. Here, we show that silanized copper oxide surfaces created via a simple fabrication method can achieve highly efficient jumping-droplet condensation heat transfer. We experimentally demonstrated a 25% higher overall heat flux and 30% higher condensation heat transfer coefficient compared to state-of-the-art hydrophobic condensing surfaces at low supersaturations (<1.12). This work not only shows significant condensation heat transfer enhancement but also promises a low cost and scalable approach to increase efficiency for applications such as atmospheric water harvesting and dehumidification. Furthermore, the results offer insights and an avenue to achieve high flux superhydrophobic condensation.

  2. Scalable purification of Bacillus anthracis protective antigen from Escherichia coli.

    PubMed

    Gwinn, William; Zhang, Mei; Mon, Sandii; Sampey, Darryl; Zukauskas, David; Kassebaum, Corby; Zmuda, Jonathan F; Tsai, Amos; Laird, Michael W

    2006-01-01

    The anthrax toxin consists of three proteins, protective antigen (PA), lethal factor, and edema factor that are produced by the Gram-positive bacterium, Bacillus anthracis. Current vaccines against anthrax use PA as their primary component. In this study, we developed a scalable process to produce and purify multi-gram quantities of highly pure, recombinant PA (rPA) from Escherichia coli. The rPA protein was produced in a 50-L fermentor and purified to >99% purity using anion-exchange, hydrophobic interaction, and hydroxyapatite chromatography. The final yield of purified rPA from medium cell density fermentations resulted in approximately 2.7 g of rPA per kg of cell paste (approximately 270 mg/L) of highly pure, biologically active rPA protein. The results presented here exhibit the ability to generate multi-gram quantities of rPA from E. coli that may be used for the development of new anthrax vaccines and anthrax therapeutics. PMID:15935696

  3. Superconductor digital electronics: Scalability and energy efficiency issues (Review Article)

    NASA Astrophysics Data System (ADS)

    Tolpygo, Sergey K.

    2016-05-01

    Superconductor digital electronics using Josephson junctions as ultrafast switches and magnetic-flux encoding of information was proposed over 30 years ago as a sub-terahertz clock frequency alternative to semiconductor electronics based on complementary metal-oxide-semiconductor (CMOS) transistors. Recently, interest in developing superconductor electronics has been renewed due to a search for energy saving solutions in applications related to high-performance computing. The current state of superconductor electronics and fabrication processes are reviewed in order to evaluate whether this electronics is scalable to a very large scale integration (VLSI) required to achieve computation complexities comparable to CMOS processors. A fully planarized process at MIT Lincoln Laboratory, perhaps the most advanced process developed so far for superconductor electronics, is used as an example. The process has nine superconducting layers: eight Nb wiring layers with the minimum feature size of 350 nm, and a thin superconducting layer for making compact high-kinetic-inductance bias inductors. All circuit layers are fully planarized using chemical mechanical planarization (CMP) of SiO2 interlayer dielectric. The physical limitations imposed on the circuit density by Josephson junctions, circuit inductors, shunt and bias resistors, etc., are discussed. Energy dissipation in superconducting circuits is also reviewed in order to estimate whether this technology, which requires cryogenic refrigeration, can be energy efficient. Fabrication process development required for increasing the density of superconductor digital circuits by a factor of ten and achieving densities above 107 Josephson junctions per cm2 is described.

  4. Scalable Parallel Density-based Clustering and Applications

    NASA Astrophysics Data System (ADS)

    Patwary, Mostofa Ali

    2014-04-01

    Recently, density-based clustering algorithms (DBSCAN and OPTICS) have gotten significant attention of the scientific community due to their unique capability of discovering arbitrary shaped clusters and eliminating noise data. These algorithms have several applications, which require high performance computing, including finding halos and subhalos (clusters) from massive cosmology data in astrophysics, analyzing satellite images, X-ray crystallography, and anomaly detection. However, parallelization of these algorithms are extremely challenging as they exhibit inherent sequential data access order, unbalanced workload resulting in low parallel efficiency. To break the data access sequentiality and to achieve high parallelism, we develop new parallel algorithms, both for DBSCAN and OPTICS, designed using graph algorithmic techniques. For example, our parallel DBSCAN algorithm exploits the similarities between DBSCAN and computing connected components. Using datasets containing up to a billion floating point numbers, we show that our parallel density-based clustering algorithms significantly outperform the existing algorithms, achieving speedups up to 27.5 on 40 cores on shared memory architecture and speedups up to 5,765 using 8,192 cores on distributed memory architecture. In our experiments, we found that while achieving the scalability, our algorithms produce clustering results with comparable quality to the classical algorithms.

  5. Phonon-based scalable quantum computing and sensing (Presentation Video)

    NASA Astrophysics Data System (ADS)

    El-Kady, Ihab

    2015-04-01

    Quantum computing fundamentally depends on the ability to concurrently entangle and individually address/control a large number of qubits. In general, the primary inhibitors of large scale entanglement are qubit dependent; for example inhomogeneity in quantum dots, spectral crowding brought about by proximity-based entanglement in ions, weak interactions of neutral atoms, and the fabrication tolerances in the case of Si-vacancies or SQUIDs. We propose an inherently scalable solid-state qubit system with individually addressable qubits based on the coupling of a phonon with an acceptor impurity in a high-Q Phononic Crystal resonant cavity. Due to their unique nonlinear properties, phonons enable new opportunities for quantum devices and physics. We present a phononic crystal-based platform for observing the phonon analogy of cavity quantum electrodynamics, called phonodynamics, in a solid-state system. Practical schemes involve selective placement of a single acceptor atom in the peak of the strain field in a high-Q phononic crystal cavity that enables strong coupling of the phonon modes to the energy levels of the atom. A qubit is then created by entangling a phonon at the resonance frequency of the cavity with the atomic acceptor states. We show theoretical optimization of the cavity design and excitation waveguides, along with estimated performance figures of the phoniton system. Qubits based on this half-sound, half-matter quasi-particle, may outcompete other quantum architectures in terms of combined emission rate, coherence lifetime, and fabrication demands.

  6. Scalability of Coral Rugosity From Microns to Centimeters

    NASA Astrophysics Data System (ADS)

    Zawada, D. G.; Rohwer, F. L.; Frank, L. R.

    2006-12-01

    Coral reefs represent one of the roughest structures in the marine environment. This roughness is a significant component of the high degree of habitat complexity associated with reefs. Various studies have investigated reef rugosity at discrete spatial scales, ranging from kilometers down to millimeters, and the associated biological (e.g. organism assemblages and recruitment) and physical (e.g. circulation and mass transfer) impacts. In this study, we devised a new technique for quantifying rugosity over a continuum of fine spatial scales from 200 μm to 1 cm. To achieve this high spatial resolution, a digital approach was developed, based on images collected with a functional magnetic resonance imaging (fMRI) system. Each fMRI image represents a 200 μm thick slice through a coral. Consecutive image slices were acquired as the specimen was translated through the fMRI system. These images were processed to create a three dimensional model representing the external surface of a coral. Analogous to the commonly used chain method for quantifying rugosity, digital "chains" with link sizes ranging from 200 μm to 1 cm were draped over the coral model. Here, we present results pertaining to the scalability of rugosity over these small spatial scales for three different coral species exhibiting different morphologies.

  7. A Scalable Gaussian Process Analysis Algorithm for Biomass Monitoring

    SciTech Connect

    Chandola, Varun; Vatsavai, Raju

    2011-01-01

    Biomass monitoring is vital for studying the carbon cycle of earth's ecosystem and has several significant implications, especially in the context of understanding climate change and its impacts. Recently, several change detection methods have been proposed to identify land cover changes in temporal profiles (time series) of vegetation collected using remote sensing instruments, but do not satisfy one or both of the two requirements of the biomass monitoring problem, i.e., {\\em operating in online mode} and {\\em handling periodic time series}. In this paper, we adapt Gaussian process regression to detect changes in such time series in an online fashion. While Gaussian process (GP) have been widely used as a kernel based learning method for regression and classification, their applicability to massive spatio-temporal data sets, such as remote sensing data, has been limited owing to the high computational costs involved. We focus on addressing the scalability issues associated with the proposed GP based change detection algorithm. This paper makes several significant contributions. First, we propose a GP based online time series change detection algorithm and demonstrate its effectiveness in detecting different types of changes in {\\em Normalized Difference Vegetation Index} (NDVI) data obtained from a study area in Iowa, USA. Second, we propose an efficient Toeplitz matrix based solution which significantly improves the computational complexity and memory requirements of the proposed GP based method. Specifically, the proposed solution can analyze a time series of length $t$ in $O(t^2)$ time while maintaining a $O(t)$ memory footprint, compared to the $O(t^3)$ time and $O(t^2)$ memory requirement of standard matrix manipulation based methods. Third, we describe a parallel version of the proposed solution which can be used to simultaneously analyze a large number of time series. We study three different parallel implementations: using threads, MPI, and a hybrid

  8. Bubble pump: scalable strategy for in-plane liquid routing.

    PubMed

    Oskooei, Ali; Günther, Axel

    2015-07-01

    We present an on-chip liquid routing technique intended for application in well-based microfluidic systems that require long-term active pumping at low to medium flowrates. Our technique requires only one fluidic feature layer, one pneumatic control line and does not rely on flexible membranes and mechanical or moving parts. The presented bubble pump is therefore compatible with both elastomeric and rigid substrate materials and the associated scalable manufacturing processes. Directed liquid flow was achieved in a microchannel by an in-series configuration of two previously described "bubble gates", i.e., by gas-bubble enabled miniature gate valves. Only one time-dependent pressure signal is required and initiates at the upstream (active) bubble gate a reciprocating bubble motion. Applied at the downstream (passive) gate a time-constant gas pressure level is applied. In its rest state, the passive gate remains closed and only temporarily opens while the liquid pressure rises due to the active gate's reciprocating bubble motion. We have designed, fabricated and consistently operated our bubble pump with a variety of working liquids for >72 hours. Flow rates of 0-5.5 μl min(-1), were obtained and depended on the selected geometric dimensions, working fluids and actuation frequencies. The maximum operational pressure was 2.9 kPa-9.1 kPa and depended on the interfacial tension of the working fluids. Attainable flow rates compared favorably with those of available micropumps. We achieved flow rate enhancements of 30-100% by operating two bubble pumps in tandem and demonstrated scalability of the concept in a multi-well format with 12 individually and uniformly perfused microchannels (variation in flow rate <7%). We envision the demonstrated concept to allow for the consistent on-chip delivery of a wide range of different liquids that may even include highly reactive or moisture sensitive solutions. The presented bubble pump may provide active flow control for

  9. GPU-based Scalable Volumetric Reconstruction for Multi-view Stereo

    SciTech Connect

    Kim, H; Duchaineau, M; Max, N

    2011-09-21

    We present a new scalable volumetric reconstruction algorithm for multi-view stereo using a graphics processing unit (GPU). It is an effectively parallelized GPU algorithm that simultaneously uses a large number of GPU threads, each of which performs voxel carving, in order to integrate depth maps with images from multiple views. Each depth map, triangulated from pair-wise semi-dense correspondences, represents a view-dependent surface of the scene. This algorithm also provides scalability for large-scale scene reconstruction in a high resolution voxel grid by utilizing streaming and parallel computation. The output is a photo-realistic 3D scene model in a volumetric or point-based representation. We demonstrate the effectiveness and the speed of our algorithm with a synthetic scene and real urban/outdoor scenes. Our method can also be integrated with existing multi-view stereo algorithms such as PMVS2 to fill holes or gaps in textureless regions.

  10. Generation of scalable terahertz radiation from cylindrically focused two-color laser pulses in air

    NASA Astrophysics Data System (ADS)

    Kuk, D.; Yoo, Y. J.; Rosenthal, E. W.; Jhajj, N.; Milchberg, H. M.; Kim, K. Y.

    2016-03-01

    We demonstrate scalable terahertz (THz) generation by focusing terawatt, two-color laser pulses in air with a cylindrical lens. This focusing geometry creates a two-dimensional air plasma sheet, which yields two diverging THz lobe profiles in the far field. This setup can avoid plasma-induced laser defocusing and subsequent THz saturation, previously observed with spherical lens focusing of high-power laser pulses. By expanding the plasma source into a two-dimensional sheet, cylindrical focusing can lead to scalable THz generation. This scheme provides an energy conversion efficiency of 7 × 10-4, ˜7 times better than spherical lens focusing. The diverging THz lobes are refocused with a combination of cylindrical and parabolic mirrors to produce strong THz fields (>21 MV/cm) at the focal point.

  11. Interface-Free Area-Scalable Self-Powered Electroluminescent System Driven by Triboelectric Generator

    PubMed Central

    Yan Wei, Xiao; Kuang, Shuang Yang; Yang Li, Hua; Pan, Caofeng; Zhu, Guang; Wang, Zhong Lin

    2015-01-01

    Self-powered system that is interface-free is greatly desired for area-scalable application. Here we report a self-powered electroluminescent system that consists of a triboelectric generator (TEG) and a thin-film electroluminescent (TFEL) lamp. The TEG provides high-voltage alternating electric output, which fits in well with the needs of the TFEL lamp. Induced charges pumped onto the lamp by the TEG generate an electric field that is sufficient to excite luminescence without an electrical interface circuit. Through rational serial connection of multiple TFEL lamps, effective and area-scalable luminescence is realized. It is demonstrated that multiple types of TEGs are applicable to the self-powered system, indicating that the system can make use of diverse mechanical sources and thus has potentially broad applications in illumination, display, entertainment, indication, surveillance and many others. PMID:26338365

  12. Toward Scalable Trustworthy Computing Using the Human-Physiology-Immunity Metaphor

    SciTech Connect

    Hively, Lee M; Sheldon, Frederick T

    2011-01-01

    The cybersecurity landscape consists of an ad hoc patchwork of solutions. Optimal cybersecurity is difficult for various reasons: complexity, immense data and processing requirements, resource-agnostic cloud computing, practical time-space-energy constraints, inherent flaws in 'Maginot Line' defenses, and the growing number and sophistication of cyberattacks. This article defines the high-priority problems and examines the potential solution space. In that space, achieving scalable trustworthy computing and communications is possible through real-time knowledge-based decisions about cyber trust. This vision is based on the human-physiology-immunity metaphor and the human brain's ability to extract knowledge from data and information. The article outlines future steps toward scalable trustworthy systems requiring a long-term commitment to solve the well-known challenges.

  13. Scalable Sensor Data Processor: A Multi-Core Payload Data Processor ASIC

    NASA Astrophysics Data System (ADS)

    Berrojo, L.; Moreno, R.; Regada, R.; Garcia, E.; Trautner, R.; Rauwerda, G.; Sunesen, K.; He, Y.; Redant, S.; Thys, G.; Andersson, J.; Habinc, S.

    2015-09-01

    The Scalable Sensor Data Processor (SSDP) project, under ESA contract and with TAS-E as prime contractor, targets the development of a multi-core ASIC for payload data processing to be used, among other terrestrial and space application areas, in future scientific and exploration missions with harsh radiation environments. The SSDP is a mixed-signal heterogeneous multi-core System-on-Chip (SoC). It combines GPP and NoC-based DSP subsystems with on-chip ADCs and several standard space I/Fs to make a flexible, configurable and scalable device. The NoC comprises two state-of-the-art fixed point Xentium® DSP processors, providing the device with high data processing capabilities.

  14. Accuracy, Speed, Scalability: the Challenges of Large-Scale DFT Simulations

    NASA Astrophysics Data System (ADS)

    Gygi, Francois

    2014-03-01

    First-Principles Molecular Dynamics (FPMD) simulations based on Density Functional Theory (DFT) have become popular in investigations of electronic and structural properties of liquids and solids. The current upsurge in available computing resources enables simulations of larger and more complex systems, such as solvated ions or defects in crystalline solids. The high cost of FPMD simulations however still strongly limits the size of feasible simulations, in particular when using hybrid-DFT approximations. In addition, the simulation times needed to extract statistically meaningful quantities also grows with system size, which puts a premium on scalable implementations. We discuss recent research in the design and implementation of scalable FPMD algorithms, with emphasis on controlled-accuracy approximations and accurate hybrid-DFT molecular dynamics simulations, using examples of applications to materials science and chemistry. Work supported by DOE-BES under grant DE-SC0008938.

  15. Architecture-Aware Algorithms for Scalable Performance and Resilience on Heterogeneous Architectures. Final Report

    SciTech Connect

    Gropp, William D.

    2014-06-23

    With the coming end of Moore's law, it has become essential to develop new algorithms and techniques that can provide the performance needed by demanding computational science applications, especially those that are part of the DOE science mission. This work was part of a multi-institution, multi-investigator project that explored several approaches to develop algorithms that would be effective at the extreme scales and with the complex processor architectures that are expected at the end of this decade. The work by this group developed new performance models that have already helped guide the development of highly scalable versions of an algebraic multigrid solver, new programming approaches designed to support numerical algorithms on heterogeneous architectures, and a new, more scalable version of conjugate gradient, an important algorithm in the solution of very large linear systems of equations.

  16. Lilith: A software framework for the rapid development of scalable tools for distributed computing

    SciTech Connect

    Gentile, A.C.; Evensky, D.A.; Armstrong, R.C.

    1997-12-31

    Lilith is a general purpose tool that provides a highly scalable, easy distribution of user code across a heterogeneous computing platform. By handling the details of code distribution and communication, such a framework allows for the rapid development of tools for the use and management of large distributed systems. This speed-up in development not only enables the easy creation of tools as needed but also facilitates the ultimate development of more refined, hard-coded tools as well. Lilith is written in Java, providing platform independence and further facilitating rapid tool development through Object reuse and ease of development. The authors present the user-involved objects in the Lilith Distributed Object System and the Lilith User API. They present an example of tool development, illustrating the user calls, and present results demonstrating Lilith`s scalability.

  17. Interface-Free Area-Scalable Self-Powered Electroluminescent System Driven by Triboelectric Generator

    NASA Astrophysics Data System (ADS)

    Yan Wei, Xiao; Kuang, Shuang Yang; Yang Li, Hua; Pan, Caofeng; Zhu, Guang; Wang, Zhong Lin

    2015-09-01

    Self-powered system that is interface-free is greatly desired for area-scalable application. Here we report a self-powered electroluminescent system that consists of a triboelectric generator (TEG) and a thin-film electroluminescent (TFEL) lamp. The TEG provides high-voltage alternating electric output, which fits in well with the needs of the TFEL lamp. Induced charges pumped onto the lamp by the TEG generate an electric field that is sufficient to excite luminescence without an electrical interface circuit. Through rational serial connection of multiple TFEL lamps, effective and area-scalable luminescence is realized. It is demonstrated that multiple types of TEGs are applicable to the self-powered system, indicating that the system can make use of diverse mechanical sources and thus has potentially broad applications in illumination, display, entertainment, indication, surveillance and many others.

  18. A scalable climate health justice assessment model

    PubMed Central

    McDonald, Yolanda J.; Grineski, Sara E.; Collins, Timothy W.; Kim, Young-An

    2014-01-01

    This paper introduces a scalable “climate health justice” model for assessing and projecting incidence, treatment costs, and sociospatial disparities for diseases with well-documented climate change linkages. The model is designed to employ low-cost secondary data, and it is rooted in a perspective that merges normative environmental justice concerns with theoretical grounding in health inequalities. Since the model employs International Classification of Diseases, Ninth Revision Clinical Modification (ICD-9-CM) disease codes, it is transferable to other contexts, appropriate for use across spatial scales, and suitable for comparative analyses. We demonstrate the utility of the model through analysis of 2008–2010 hospitalization discharge data at state and county levels in Texas (USA). We identified several disease categories (i.e., cardiovascular, gastrointestinal, heat-related, and respiratory) associated with climate change, and then selected corresponding ICD-9 codes with the highest hospitalization counts for further analyses. Selected diseases include ischemic heart disease, diarrhea, heat exhaustion/cramps/stroke/syncope, and asthma. Cardiovascular disease ranked first among the general categories of diseases for age-adjusted hospital admission rate (5286.37 per 100,000). In terms of specific selected diseases (per 100,000 population), asthma ranked first (517.51), followed by ischemic heart disease (195.20), diarrhea (75.35), and heat exhaustion/cramps/stroke/syncope (7.81). Charges associated with the selected diseases over the 3-year period amounted to US$5.6 billion. Blacks were disproportionately burdened by the selected diseases in comparison to non-Hispanic whites, while Hispanics were not. Spatial distributions of the selected disease rates revealed geographic zones of disproportionate risk. Based upon a downscaled regional climate-change projection model, we estimate a >5% increase in the incidence and treatment costs of asthma attributable to

  19. A scalable neuroinformatics data flow for electrophysiological signals using MapReduce.

    PubMed

    Jayapandian, Catherine; Wei, Annan; Ramesh, Priya; Zonjy, Bilal; Lhatoo, Samden D; Loparo, Kenneth; Zhang, Guo-Qiang; Sahoo, Satya S

    2015-01-01

    Data-driven neuroscience research is providing new insights in progression of neurological disorders and supporting the development of improved treatment approaches. However, the volume, velocity, and variety of neuroscience data generated from sophisticated recording instruments and acquisition methods have exacerbated the limited scalability of existing neuroinformatics tools. This makes it difficult for neuroscience researchers to effectively leverage the growing multi-modal neuroscience data to advance research in serious neurological disorders, such as epilepsy. We describe the development of the Cloudwave data flow that uses new data partitioning techniques to store and analyze electrophysiological signal in distributed computing infrastructure. The Cloudwave data flow uses MapReduce parallel programming algorithm to implement an integrated signal data processing pipeline that scales with large volume of data generated at high velocity. Using an epilepsy domain ontology together with an epilepsy focused extensible data representation format called Cloudwave Signal Format (CSF), the data flow addresses the challenge of data heterogeneity and is interoperable with existing neuroinformatics data representation formats, such as HDF5. The scalability of the Cloudwave data flow is evaluated using a 30-node cluster installed with the open source Hadoop software stack. The results demonstrate that the Cloudwave data flow can process increasing volume of signal data by leveraging Hadoop Data Nodes to reduce the total data processing time. The Cloudwave data flow is a template for developing highly scalable neuroscience data processing pipelines using MapReduce algorithms to support a variety of user applications. PMID:25852536

  20. Heat-treated stainless steel felt as scalable anode material for bioelectrochemical systems.

    PubMed

    Guo, Kun; Soeriyadi, Alexander H; Feng, Huajun; Prévoteau, Antonin; Patil, Sunil A; Gooding, J Justin; Rabaey, Korneel

    2015-11-01

    This work reports a simple and scalable method to convert stainless steel (SS) felt into an effective anode for bioelectrochemical systems (BESs) by means of heat treatment. X-ray photoelectron spectroscopy and cyclic voltammetry elucidated that the heat treatment generated an iron oxide rich layer on the SS felt surface. The iron oxide layer dramatically enhanced the electroactive biofilm formation on SS felt surface in BESs. Consequently, the sustained current densities achieved on the treated electrodes (1 cm(2)) were around 1.5±0.13 mA/cm(2), which was seven times higher than the untreated electrodes (0.22±0.04 mA/cm(2)). To test the scalability of this material, the heat-treated SS felt was scaled up to 150 cm(2) and similar current density (1.5 mA/cm(2)) was achieved on the larger electrode. The low cost, straightforwardness of the treatment, high conductivity and high bioelectrocatalytic performance make heat-treated SS felt a scalable anodic material for BESs. PMID:26112346

  1. A scalable neuroinformatics data flow for electrophysiological signals using MapReduce

    PubMed Central

    Jayapandian, Catherine; Wei, Annan; Ramesh, Priya; Zonjy, Bilal; Lhatoo, Samden D.; Loparo, Kenneth; Zhang, Guo-Qiang; Sahoo, Satya S.

    2015-01-01

    Data-driven neuroscience research is providing new insights in progression of neurological disorders and supporting the development of improved treatment approaches. However, the volume, velocity, and variety of neuroscience data generated from sophisticated recording instruments and acquisition methods have exacerbated the limited scalability of existing neuroinformatics tools. This makes it difficult for neuroscience researchers to effectively leverage the growing multi-modal neuroscience data to advance research in serious neurological disorders, such as epilepsy. We describe the development of the Cloudwave data flow that uses new data partitioning techniques to store and analyze electrophysiological signal in distributed computing infrastructure. The Cloudwave data flow uses MapReduce parallel programming algorithm to implement an integrated signal data processing pipeline that scales with large volume of data generated at high velocity. Using an epilepsy domain ontology together with an epilepsy focused extensible data representation format called Cloudwave Signal Format (CSF), the data flow addresses the challenge of data heterogeneity and is interoperable with existing neuroinformatics data representation formats, such as HDF5. The scalability of the Cloudwave data flow is evaluated using a 30-node cluster installed with the open source Hadoop software stack. The results demonstrate that the Cloudwave data flow can process increasing volume of signal data by leveraging Hadoop Data Nodes to reduce the total data processing time. The Cloudwave data flow is a template for developing highly scalable neuroscience data processing pipelines using MapReduce algorithms to support a variety of user applications. PMID:25852536

  2. Thermal and plasma treatments for improved (sub-)1 nm equivalent oxide thickness planar and FinFET-based replacement metal gate high-k last devices and enabling a simplified scalable CMOS integration scheme

    NASA Astrophysics Data System (ADS)

    Veloso, Anabela; Boccardi, Guillaume; Ragnarsson, Lars-Åke; Higuchi, Yuichi; Arimura, Hiroaki; Lee, Jae Woo; Simoen, Eddy; Cho, Moon Ju; Roussel, Philippe J.; Paraschiv, Vasile; Shi, Xiaoping; Schram, Tom; Aik Chew, Soon; Brus, Stephan; Dangol, Anish; Vecchio, Emma; Sebaai, Farid; Kellens, Kristof; Heylen, Nancy; Devriendt, Katia; Dekkers, Harold; Van Ammel, Annemie; Witters, Thomas; Conard, Thierry; Vaesen, Inge; Richard, Olivier; Bender, Hugo; Athimulam, Raja; Chiarella, Thomas; Thean, Aaron; Horiguchi, Naoto

    2014-01-01

    We report on aggressively scaled replacement metal gate, high-k last (RMG-HKL) planar and multi-gate fin field-effect transistor (FinFET) devices, systematically investigating the impact of post high-k deposition thermal (PDA) and plasma (SF6) treatments on device characteristics, and providing a deeper insight into underlying degradation mechanisms. We demonstrate that: 1) substantially reduced gate leakage (JG) and noise can be obtained for both type of devices with PDA and F incorporation in the gate stack by SF6, without equivalent oxide thickness (EOT) penalty; 2) SF6 enables improved mobility and reduced interface trapped charge density (Nit) down to narrower fin devices [fin width (WFin) ≥ 5 nm], mitigating the impact of fin patterning and fin sidewall crystal orientations, while allowing a simplified dual-effective work function (EWF) CMOS scheme suitable for both device architectures; 3) PDA yields smaller, in absolute values, PMOS threshold voltage |VT|, and substantially improved reliability behavior due to reduction of bulk defects.

  3. Dynamic superhydrophobic behavior in scalable random textured polymeric surfaces

    NASA Astrophysics Data System (ADS)

    Moreira, David; Park, Sung-hoon; Lee, Sangeui; Verma, Neil; Bandaru, Prabhakar R.

    2016-03-01

    Superhydrophobic (SH) surfaces, created from hydrophobic materials with micro- or nano- roughness, trap air pockets in the interstices of the roughness, leading, in fluid flow conditions, to shear-free regions with finite interfacial fluid velocity and reduced resistance to flow. Significant attention has been given to SH conditions on ordered, periodic surfaces. However, in practical terms, random surfaces are more applicable due to their relative ease of fabrication. We investigate SH behavior on a novel durable polymeric rough surface created through a scalable roll-coating process with varying micro-scale roughness through velocity and pressure drop measurements. We introduce a new method to construct the velocity profile over SH surfaces with significant roughness in microchannels. Slip length was measured as a function of differing roughness and interstitial air conditions, with roughness and air fraction parameters obtained through direct visualization. The slip length was matched to scaling laws with good agreement. Roughness at high air fractions led to a reduced pressure drop and higher velocities, demonstrating the effectiveness of the considered surface in terms of reduced resistance to flow. We conclude that the observed air fraction under flow conditions is the primary factor determining the response in fluid flow. Such behavior correlated well with the hydrophobic or superhydrophobic response, indicating significant potential for practical use in enhancing fluid flow efficiency.

  4. Efficient and scalable scaffolding using optical restriction maps

    PubMed Central

    2014-01-01

    In the next generation sequencing techniques millions of short reads are produced from a genomic sequence at a single run. The chances of low read coverage to some regions of the sequence are very high. The reads are short and very large in number. Due to erroneous base calling, there could be errors in the reads. As a consequence, sequence assemblers often fail to sequence an entire DNA molecule and instead output a set of overlapping segments that together represent a consensus region of the DNA. This set of overlapping segments are collectively called contigs in the literature. The final step of the sequencing process, called scaffolding, is to assemble the contigs into a correct order. Scaffolding techniques typically exploit additional information such as mate-pairs, pair-ends, or optical restriction maps. In this paper we introduce a series of novel algorithms for scaffolding that exploit optical restriction maps (ORMs). Simulation results show that our algorithms are indeed reliable, scalable, and efficient compared to the best known algorithms in the literature. PMID:25081913

  5. Advances in Patch-Based Adaptive Mesh Refinement Scalability

    SciTech Connect

    Gunney, Brian T.N.; Anderson, Robert W.

    2015-12-18

    Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extension of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.

  6. The Scalable Coherent Interface and related standards projects

    SciTech Connect

    Gustavson, D.B.

    1991-09-01

    The Scalable Coherent Interface (SCI) project (IEEE P1596) found a way to avoid the limits that are inherent in bus technology. SCI provides bus-like services by transmitting packets on a collection of point-to-point unidirectional links. The SCI protocols support cache coherence in a distributed-shared-memory multiprocessor model, message passing, I/O, and local-area-network-like communication over fiber optic or wire links. VLSI circuits that operate parallel links at 1000 MByte/s and serial links at 1000 Mbit/s will be available early in 1992. Several ongoing SCI-related projects are applying the SCI technology to new areas or extending it to more difficult problems. P1596.1 defines the architecture of a bridge between SCI and VME; P1596.2 compatibly extends the cache coherence mechanism for efficient operation with kiloprocessor systems; P1596.3 defines new low-voltage (about 0.25 V) differential signals suitable for low power interfaces for CMOS or GaAs VLSI implementations of SCI; P1596.4 defines a high performance memory chip interface using these signals; P1596.5 defines data transfer formats for efficient interprocessor communication in heterogeneous multiprocessor systems. This paper reports the current status of SCI, related standards, and new projects. 16 refs.

  7. Scalable, Low-Noise Architecture for Integrated Terahertz Imagers

    NASA Astrophysics Data System (ADS)

    Gergelyi, Domonkos; Földesy, Péter; Zarándy, Ákos

    2015-06-01

    We propose a scalable, low-noise imager architecture for terahertz recordings that helps to build large-scale integrated arrays from any field-effect transistor (FET)- or HEMT-based terahertz detector. It enhances the signal-to-noise ratio (SNR) by inherently enabling complex sampling schemes. The distinguishing feature of the architecture is the serially connected detectors with electronically controllable photoresponse. We show that this architecture facilitate room temperature imaging by decreasing the low-noise amplifier (LNA) noise to one-sixteenth of a non-serial sensor while also reducing the number of multiplexed signals in the same proportion. The serially coupled architecture can be combined with the existing read-out circuit organizations to create high-resolution, coarse-grain sensor arrays. Besides, it adds the capability to suppress overall noise with increasing array size. The theoretical considerations are proven on a 4 by 4 detector array manufactured on 180 nm feature sized standard CMOS technology. The detector array is integrated with a low-noise AC-coupled amplifier of 40 dB gain and has a resonant peak at 460 GHz with 200 kV/W overall sensitivity.

  8. Scalable Indoor Localization via Mobile Crowdsourcing and Gaussian Process

    PubMed Central

    Chang, Qiang; Li, Qun; Shi, Zesen; Chen, Wei; Wang, Weiping

    2016-01-01

    Indoor localization using Received Signal Strength Indication (RSSI) fingerprinting has been extensively studied for decades. The positioning accuracy is highly dependent on the density of the signal database. In areas without calibration data, however, this algorithm breaks down. Building and updating a dense signal database is labor intensive, expensive, and even impossible in some areas. Researchers are continually searching for better algorithms to create and update dense databases more efficiently. In this paper, we propose a scalable indoor positioning algorithm that works both in surveyed and unsurveyed areas. We first propose Minimum Inverse Distance (MID) algorithm to build a virtual database with uniformly distributed virtual Reference Points (RP). The area covered by the virtual RPs can be larger than the surveyed area. A Local Gaussian Process (LGP) is then applied to estimate the virtual RPs’ RSSI values based on the crowdsourced training data. Finally, we improve the Bayesian algorithm to estimate the user’s location using the virtual database. All the parameters are optimized by simulations, and the new algorithm is tested on real-case scenarios. The results show that the new algorithm improves the accuracy by 25.5% in the surveyed area, with an average positioning error below 2.2 m for 80% of the cases. Moreover, the proposed algorithm can localize the users in the neighboring unsurveyed area. PMID:26999139

  9. Scalable Indoor Localization via Mobile Crowdsourcing and Gaussian Process.

    PubMed

    Chang, Qiang; Li, Qun; Shi, Zesen; Chen, Wei; Wang, Weiping

    2016-01-01

    Indoor localization using Received Signal Strength Indication (RSSI) fingerprinting has been extensively studied for decades. The positioning accuracy is highly dependent on the density of the signal database. In areas without calibration data, however, this algorithm breaks down. Building and updating a dense signal database is labor intensive, expensive, and even impossible in some areas. Researchers are continually searching for better algorithms to create and update dense databases more efficiently. In this paper, we propose a scalable indoor positioning algorithm that works both in surveyed and unsurveyed areas. We first propose Minimum Inverse Distance (MID) algorithm to build a virtual database with uniformly distributed virtual Reference Points (RP). The area covered by the virtual RPs can be larger than the surveyed area. A Local Gaussian Process (LGP) is then applied to estimate the virtual RPs' RSSI values based on the crowdsourced training data. Finally, we improve the Bayesian algorithm to estimate the user's location using the virtual database. All the parameters are optimized by simulations, and the new algorithm is tested on real-case scenarios. The results show that the new algorithm improves the accuracy by 25.5% in the surveyed area, with an average positioning error below 2.2 m for 80% of the cases. Moreover, the proposed algorithm can localize the users in the neighboring unsurveyed area. PMID:26999139

  10. The design of a scalable, fixed-time computer benchmark

    SciTech Connect

    Gustafson, J.; Rover, D.; Elbert, S.; Carter, M.

    1990-10-01

    By using the principle of fixed time benchmarking, it is possible to compare a very wide range of computers, from a small personal computer to the most powerful parallel supercomputer, an a single scale. Fixed-time benchmarks promise far greater longevity than those based on a particular problem size, and are more appropriate for grand challenge'' capability comparison. We present the design of a benchmark, SLALOM{trademark}, that scales automatically to the computing power available, and corrects several deficiencies in various existing benchmarks: it is highly scalable, it solves a real problem, it includes input and output times, and it can be run on parallel machines of all kinds, using any convenient language. The benchmark provides a reasonable estimate of the size of problem solvable on scientific computers. Results are presented that span six orders of magnitude for contemporary computers of various architectures. The benchmarks also can be used to demonstrate a new source of superlinear speedup in parallel computers. 15 refs., 14 figs., 3 tabs.

  11. Design and performance of a scalable, parallel statistics toolkit.

    SciTech Connect

    Thompson, David C.; Bennett, Janine Camille; Pebay, Philippe Pierre

    2010-11-01

    Most statistical software packages implement a broad range of techniques but do so in an ad hoc fashion, leaving users who do not have a broad knowledge of statistics at a disadvantage since they may not understand all the implications of a given analysis or how to test the validity of results. These packages are also largely serial in nature, or target multicore architectures instead of distributed-memory systems, or provide only a small number of statistics in parallel. This paper surveys a collection of parallel implementations of statistics algorithm developed as part of a common framework over the last 3 years. The framework strategically groups modeling techniques with associated verification and validation techniques to make the underlying assumptions of the statistics more clear. Furthermore it employs a design pattern specifically targeted for distributed-memory parallelism, where architectural advances in large-scale high-performance computing have been focused. Moment-based statistics (which include descriptive, correlative, and multicorrelative statistics, principal component analysis (PCA), and k-means statistics) scale nearly linearly with the data set size and number of processes. Entropy-based statistics (which include order and contingency statistics) do not scale well when the data in question is continuous or quasi-diffuse but do scale well when the data is discrete and compact. We confirm and extend our earlier results by now establishing near-optimal scalability with up to 10,000 processes.

  12. Designing Scalable PGAS Communication Subsystems on Cray Gemini Interconnect

    SciTech Connect

    Vishnu, Abhinav; Daily, Jeffrey A.; Palmer, Bruce J.

    2012-12-26

    The Cray Gemini Interconnect has been recently introduced as a next generation network architecture for building multi-petaflop supercomputers. Cray XE6 systems including LANL Cielo, NERSC Hopper, ORNL Titan and proposed NCSA BlueWaters leverage the Gemini Interconnect as their primary Interconnection network. At the same time, programming models such as the Message Passing Interface (MPI) and Partitioned Global Address Space (PGAS) models such as Unified Parallel C (UPC) and Co-Array Fortran (CAF) have become available on these systems. Global Arrays is a popular PGAS model used in a variety of application domains including hydrodynamics, chemistry and visualization. Global Arrays uses Aggregate Re- mote Memory Copy Interface (ARMCI) as the communication runtime system for Remote Memory Access communication. This paper presents a design, implementation and performance evaluation of scalable and high performance communication subsystems on Cray Gemini Interconnect using ARMCI. The design space is explored and time-space complexities of commu- nication protocols for one-sided communication primitives such as contiguous and uniformly non-contiguous datatypes, atomic memory operations (AMOs) and memory synchronization is presented. An implementation of the proposed design (referred as ARMCI-Gemini) demonstrates the efficacy on communication primitives, application kernels such as LU decomposition and full applications such as Smooth Particle Hydrodynamics (SPH) application.

  13. The Simulation of Read-time Scalable Coherent Interface

    NASA Technical Reports Server (NTRS)

    Li, Qiang; Grant, Terry; Grover, Radhika S.

    1997-01-01

    Scalable Coherent Interface (SCI, IEEE/ANSI Std 1596-1992) (SCI1, SCI2) is a high performance interconnect for shared memory multiprocessor systems. In this project we investigate an SCI Real Time Protocols (RTSCI1) using Directed Flow Control Symbols. We studied the issues of efficient generation of control symbols, and created a simulation model of the protocol on a ring-based SCI system. This report presents the results of the study. The project has been implemented using SES/Workbench. The details that follow encompass aspects of both SCI and Flow Control Protocols, as well as the effect of realistic client/server processing delay. The report is organized as follows. Section 2 provides a description of the simulation model. Section 3 describes the protocol implementation details. The next three sections of the report elaborate on the workload, results and conclusions. Appended to the report is a description of the tool, SES/Workbench, used in our simulation, and internal details of our implementation of the protocol.

  14. Advances in Patch-Based Adaptive Mesh Refinement Scalability

    DOE PAGESBeta

    Gunney, Brian T.N.; Anderson, Robert W.

    2015-12-18

    Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extensionmore » of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.« less

  15. Scalability enhancement of AODV using local link repairing

    NASA Astrophysics Data System (ADS)

    Jain, Jyoti; Gupta, Roopam; Bandhopadhyay, T. K.

    2014-09-01

    Dynamic change in the topology of an ad hoc network makes it difficult to design an efficient routing protocol. Scalability of an ad hoc network is also one of the important criteria of research in this field. Most of the research works in ad hoc network focus on routing and medium access protocols and produce simulation results for limited-size networks. Ad hoc on-demand distance vector (AODV) is one of the best reactive routing protocols. In this article, modified routing protocols based on local link repairing of AODV are proposed. Method of finding alternate routes for next-to-next node is proposed in case of link failure. These protocols are beacon-less, means periodic hello message is removed from the basic AODV to improve scalability. Few control packet formats have been changed to accommodate suggested modification. Proposed protocols are simulated to investigate scalability performance and compared with basic AODV protocol. This also proves that local link repairing of proposed protocol improves scalability of the network. From simulation results, it is clear that scalability performance of routing protocol is improved because of link repairing method. We have tested protocols for different terrain area with approximate constant node densities and different traffic load.

  16. Scalable Fabrication of Nanoporous Carbon Fiber Films as Bifunctional Catalytic Electrodes for Flexible Zn-Air Batteries.

    PubMed

    Liu, Qin; Wang, Yaobing; Dai, Liming; Yao, Jiannian

    2016-04-01

    A flexible nanoporous carbon-fiber film for wearable electronics is prepared by a facile and scalable method through pyrolysis of electrospun polyimide. It exhibits excellent bifunctional electrocatalytic activities for oxygen reduction and oxygen evolution. Flexible rechargeable zinc-air batteries based on the carbon-fiber film show high round-trip efficiency and mechanical stability. PMID:26914270

  17. Scalable synthesis and energy applications of defect engineeered nano materials

    NASA Astrophysics Data System (ADS)

    Karakaya, Mehmet

    Nanomaterials and nanotechnologies have attracted a great deal of attention in a few decades due to their novel physical properties such as, high aspect ratio, surface morphology, impurities, etc. which lead to unique chemical, optical and electronic properties. The awareness of importance of nanomaterials has motivated researchers to develop nanomaterial growth techniques to further control nanostructures properties such as, size, surface morphology, etc. that may alter their fundamental behavior. Carbon nanotubes (CNTs) are one of the most promising materials with their rigidity, strength, elasticity and electric conductivity for future applications. Despite their excellent properties explored by the abundant research works, there is big challenge to introduce them into the macroscopic world for practical applications. This thesis first gives a brief overview of the CNTs, it will then go on mechanical and oil absorption properties of macro-scale CNT assemblies, then following CNT energy storage applications and finally fundamental studies of defect introduced graphene systems. Chapter Two focuses on helically coiled carbon nanotube (HCNT) foams in compression. Similarly to other foams, HCNT foams exhibit preconditioning effects in response to cyclic loading; however, their fundamental deformation mechanisms are unique. Bulk HCNT foams exhibit super-compressibility and recover more than 90% of large compressive strains (up to 80%). When subjected to striker impacts, HCNT foams mitigate impact stresses more effectively compared to other CNT foams comprised of non-helical CNTs (~50% improvement). The unique mechanical properties we revealed demonstrate that the HCNT foams are ideally suited for applications in packaging, impact protection, and vibration mitigation. The third chapter describes a simple method for the scalable synthesis of three-dimensional, elastic, and recyclable multi-walled carbon nanotube (MWCNT) based light weight bucky-aerogels (BAGs) that are

  18. Cooperative Server Clustering for a Scalable GAS Model on Petascale Cray XT5 Systems

    SciTech Connect

    Yu, Weikuan; Que, Xinyu; Tipparaju, Vinod; Graham, Richard L; Vetter, Jeffrey S

    2010-05-01

    Global Address Space (GAS) programming models are attractive because they retain the easy-to-use addressing model that is the characteristic of shared-memory style load and store operations. The scalability of GAS models depends directly on the design and implementation of runtime libraries on the targeted platforms. In this paper, we examine the memory requirement of a popular GAS run-time library, Aggregate Remote Memory Copy Interface (ARMCI) on petascale Cray XT 5 systems. Then we describe a new technique, cooperative server clustering, that enhances the memory scalability of ARMCI communication servers. In cooperative server clustering, ARMCI servers are organized into clusters, and cooperatively process incoming communication requests among them. A request intervention scheme is also designed to expedite the return of responses to the initiating processes. Our experimental results demonstrate that, with very little impact on ARMCI communication latency and bandwidth, cooperative server clustering is able to significantly reduce the memory requirement of ARMCI communication servers, thereby enabling highly scalable scientific applications. In particular, it dramatically reduces the total execution time of a scientific application, NWChem, by 45% on 2400 processes.

  19. Cooperative Server Clustering for a Scalable GAS Model on petascale cray XT5 Systems

    SciTech Connect

    Yu, Weikuan; Que, Xinyu; Graham, Richard L; Vetter, Jeffrey S

    2010-01-01

    Global Address Space (GAS) programming models are attractive because they retain the easy-to-use addressing model that is the characteristic of shared-memory style load and store operations. The scalability of GAS models depends directly on the design and implementation of runtime libraries on the targeted platforms. In this paper, we examine the memory requirement of a popular GAS runtime library, Aggregate Remote Memory Copy Interface (ARMCI) on petascale Cray XT 5 systems. Then we describe a new technique cooperative server clustering that enhances the memory scalability of ARMCI communication servers. In cooperative server clustering, ARMCI servers are organized into clusters, and cooperatively process incoming communication requests among them. A request intervention scheme is also designed to expedite the return of responses to the initiating processes. Our experimental results demonstrate that, with very little impact on ARMCI communication latency and bandwidth, cooperative server clustering is able to significantly reduce the memory requirement of ARMCI communication servers, thereby enabling highly scalable scientific applications. In particular, it dramatically reduces the total execution time of a scientific application, NWChem, by 45% on 2400 processes.

  20. Inter-layer motion field mapping for the scalable extension of HEVC

    NASA Astrophysics Data System (ADS)

    Xiu, Xiaoyu; Ye, Yan; He, Yong; He, Yuwen

    2013-02-01

    The next generation video coding standard, High Efficiency Video Coding (HEVC), is under development by the Joint Collaborative Team on Video Coding (JCT-VC) of the ITU-T VCEG and the ISO/IEC MPEG. As the first version of single-layer HEVC standard comes close to completion, there is a great interest to extend the standard with scalable capabilities. In this paper, an inter-layer Motion Field Mapping (MFM) algorithm is proposed for the scalable extension of HEVC to generate the motion field of inter-layer reference pictures, such that the correlation between the motion vectors (MVs) of base-layer and enhancement-layer can be exploited. Moreover, as the proposed method does not change any block-level operation, the existing single-layer encoder and decoder logic of HEVC can be directly applied without modification of motion vector prediction for the enhancement-layer. The experimental results show the effectiveness of the proposed MFM method in improving the performance of enhancement-layer motion prediction in scalable HEVC.

  1. Scalable Transparent Checkpoint-Restart of Global Address Space Applications on Virtual Machines over Infiniband

    SciTech Connect

    Villa, Oreste; Krishnamoorthy, Sriram; Nieplocha, Jaroslaw; Brown, David ML

    2009-05-18

    Checkpoint-Restart is one of the most used software approaches to achieve fault-tolerance in high-end clusters. While standard techniques typically focus on user-level solutions, the advent of virtualization software has enabled efficient and transparent system-level approaches. In this paper, we present a scalable transparent system-level solution to address fault-tolerance for applications based on global address space (GAS) programming models on Infiniband clusters. In addition to handling communication, the solution addresses transparent checkpoint of user-generated files. We exploit the support for the Infiniband network in the Xen virtual machine environment. We have developed a version of the Aggregate Remote Memory Copy Interface (ARMCI) one-sided communication library capable of suspending and resuming applications. We present efficient and scalable mechanisms to distribute checkpoint requests and to backup virtual machines memory images and file systems. We tested our approach in the context of NWChem, a popular computational chemistry suite. We demonstrated that NWChem can be executed, without any modification to the source code, on a virtualized 8-node cluster with very little overhead (below 3%). We observe that the total checkpoint time is limited by disk I/O. Finally, we measured system-size depended components of the checkpoint time on up to 1024 cores (128 nodes), demonstrating the scalability of our approach in medium/large-scale systems.

  2. A New, Scalable and Low Cost Multi-Channel Monitoring System for Polymer Electrolyte Fuel Cells.

    PubMed

    Calderón, Antonio José; González, Isaías; Calderón, Manuel; Segura, Francisca; Andújar, José Manuel

    2016-01-01

    In this work a new, scalable and low cost multi-channel monitoring system for Polymer Electrolyte Fuel Cells (PEFCs) has been designed, constructed and experimentally validated. This developed monitoring system performs non-intrusive voltage measurement of each individual cell of a PEFC stack and it is scalable, in the sense that it is capable to carry out measurements in stacks from 1 to 120 cells (from watts to kilowatts). The developed system comprises two main subsystems: hardware devoted to data acquisition (DAQ) and software devoted to real-time monitoring. The DAQ subsystem is based on the low-cost open-source platform Arduino and the real-time monitoring subsystem has been developed using the high-level graphical language NI LabVIEW. Such integration can be considered a novelty in scientific literature for PEFC monitoring systems. An original amplifying and multiplexing board has been designed to increase the Arduino input port availability. Data storage and real-time monitoring have been performed with an easy-to-use interface. Graphical and numerical visualization allows a continuous tracking of cell voltage. Scalability, flexibility, easy-to-use, versatility and low cost are the main features of the proposed approach. The system is described and experimental results are presented. These results demonstrate its suitability to monitor the voltage in a PEFC at cell level. PMID:27005630

  3. A New, Scalable and Low Cost Multi-Channel Monitoring System for Polymer Electrolyte Fuel Cells

    PubMed Central

    Calderón, Antonio José; González, Isaías; Calderón, Manuel; Segura, Francisca; Andújar, José Manuel

    2016-01-01

    In this work a new, scalable and low cost multi-channel monitoring system for Polymer Electrolyte Fuel Cells (PEFCs) has been designed, constructed and experimentally validated. This developed monitoring system performs non-intrusive voltage measurement of each individual cell of a PEFC stack and it is scalable, in the sense that it is capable to carry out measurements in stacks from 1 to 120 cells (from watts to kilowatts). The developed system comprises two main subsystems: hardware devoted to data acquisition (DAQ) and software devoted to real-time monitoring. The DAQ subsystem is based on the low-cost open-source platform Arduino and the real-time monitoring subsystem has been developed using the high-level graphical language NI LabVIEW. Such integration can be considered a novelty in scientific literature for PEFC monitoring systems. An original amplifying and multiplexing board has been designed to increase the Arduino input port availability. Data storage and real-time monitoring have been performed with an easy-to-use interface. Graphical and numerical visualization allows a continuous tracking of cell voltage. Scalability, flexibility, easy-to-use, versatility and low cost are the main features of the proposed approach. The system is described and experimental results are presented. These results demonstrate its suitability to monitor the voltage in a PEFC at cell level. PMID:27005630

  4. Instantly decodable network coding for real-time scalable video broadcast over wireless networks

    NASA Astrophysics Data System (ADS)

    Karim, Mohammad S.; Sadeghi, Parastoo; Sorour, Sameh; Aboutorab, Neda

    2016-01-01

    In this paper, we study real-time scalable video broadcast over wireless networks using instantly decodable network coding (IDNC). Such real-time scalable videos have hard deadline and impose a decoding order on the video layers. We first derive the upper bound on the probability that the individual completion times of all receivers meet the deadline. Using this probability, we design two prioritized IDNC algorithms, namely the expanding window IDNC (EW-IDNC) algorithm and the non-overlapping window IDNC (NOW-IDNC) algorithm. These algorithms provide a high level of protection to the most important video layer, namely the base layer, before considering additional video layers, namely the enhancement layers, in coding decisions. Moreover, in these algorithms, we select an appropriate packet combination over a given number of video layers so that these video layers are decoded by the maximum number of receivers before the deadline. We formulate this packet selection problem as a two-stage maximal clique selection problem over an IDNC graph. Simulation results over a real scalable video sequence show that our proposed EW-IDNC and NOW-IDNC algorithms improve the received video quality compared to the existing IDNC algorithms.

  5. Decomposable decoding and display structure for scalable media visualization over advanced collaborative environments

    NASA Astrophysics Data System (ADS)

    Kim, JaeYoun; Kim, JongWon

    2005-10-01

    In this paper, we propose a scalable visualization system to offer high-resolution visualization on multiparty collaborative environments. The proposed system treats with a coordination technique to employ large-scale high-resolution display system and to display multiple high-quality videos effectively on systems with limited resources. To handle these, the proposed system includes the distributed visualization application under generic structure to enable high-resolution video format, such as DV (digital video) and HDV (high definition video) streaming, and under decomposable decoding and display structure to assign the separated visualization task (decoding/display) to different system resources. The system is based on high-performance local area network and the high-performance network between decoding and display task is utilized as the system bus to transfer the decoded large pixel data. The main focus in this paper is the decoupling technique of decoding and display based on high-performance network to handle multiple high-resolution videos effectively. We explore the possibility of the proposed system by implementing a prototype and evaluating it over a high-performance network. Finally, the experiment results verify the improved scalable display system through the proposed structure.

  6. Network-aware scalable video monitoring system for emergency situations with operator-managed fidelity control

    NASA Astrophysics Data System (ADS)

    Al Hadhrami, Tawfik; Nightingale, James M.; Wang, Qi; Grecos, Christos

    2014-05-01

    In emergency situations, the ability to remotely monitor unfolding events using high-quality video feeds will significantly improve the incident commander's understanding of the situation and thereby aids effective decision making. This paper presents a novel, adaptive video monitoring system for emergency situations where the normal communications network infrastructure has been severely impaired or is no longer operational. The proposed scheme, operating over a rapidly deployable wireless mesh network, supports real-time video feeds between first responders, forward operating bases and primary command and control centers. Video feeds captured on portable devices carried by first responders and by static visual sensors are encoded in H.264/SVC, the scalable extension to H.264/AVC, allowing efficient, standard-based temporal, spatial, and quality scalability of the video. A three-tier video delivery system is proposed, which balances the need to avoid overuse of mesh nodes with the operational requirements of the emergency management team. In the first tier, the video feeds are delivered at a low spatial and temporal resolution employing only the base layer of the H.264/SVC video stream. Routing in this mode is designed to employ all nodes across the entire mesh network. In the second tier, whenever operational considerations require that commanders or operators focus on a particular video feed, a `fidelity control' mechanism at the monitoring station sends control messages to the routing and scheduling agents in the mesh network, which increase the quality of the received picture using SNR scalability while conserving bandwidth by maintaining a low frame rate. In this mode, routing decisions are based on reliable packet delivery with the most reliable routes being used to deliver the base and lower enhancement layers; as fidelity is increased and more scalable layers are transmitted they will be assigned to routes in descending order of reliability. The third tier

  7. Current parallel I/O limitations to scalable data analysis.

    SciTech Connect

    Mascarenhas, Ajith Arthur; Pebay, Philippe Pierre

    2011-07-01

    This report describes the limitations to parallel scalability which we have encountered when applying our otherwise optimally scalable parallel statistical analysis tool kit to large data sets distributed across the parallel file system of the current premier DOE computational facility. This report describes our study to evaluate the effect of parallel I/O on the overall scalability of a parallel data analysis pipeline using our scalable parallel statistics tool kit [PTBM11]. In this goal, we tested it using the Jaguar-pf DOE/ORNL peta-scale platform on a large combustion simulation data under a variety of process counts and domain decompositions scenarios. In this report we have recalled the foundations of the parallel statistical analysis tool kit which we have designed and implemented, with the specific double intent of reproducing typical data analysis workflows, and achieving optimal design for scalable parallel implementations. We have briefly reviewed those earlier results and publications which allow us to conclude that we have achieved both goals. However, in this report we have further established that, when used in conjuction with a state-of-the-art parallel I/O system, as can be found on the premier DOE peta-scale platform, the scaling properties of the overall analysis pipeline comprising parallel data access routines degrade rapidly. This finding is problematic and must be addressed if peta-scale data analysis is to be made scalable, or even possible. In order to attempt to address these parallel I/O limitations, we will investigate the use the Adaptable IO System (ADIOS) [LZL+10] to improve I/O performance, while maintaining flexibility for a variety of IO options, such MPI IO, POSIX IO. This system is developed at ORNL and other collaborating institutions, and is being tested extensively on Jaguar-pf. Simulation code being developed on these systems will also use ADIOS to output the data thereby making it easier for other systems, such as ours, to

  8. Comparison of scalable fast methods for long-range interactions.

    PubMed

    Arnold, Axel; Fahrenberger, Florian; Holm, Christian; Lenz, Olaf; Bolten, Matthias; Dachsel, Holger; Halver, Rene; Kabadshow, Ivo; Gähler, Franz; Heber, Frederik; Iseringhausen, Julian; Hofmann, Michael; Pippig, Michael; Potts, Daniel; Sutmann, Godehard

    2013-12-01

    Based on a parallel scalable library for Coulomb interactions in particle systems, a comparison between the fast multipole method (FMM), multigrid-based methods, fast Fourier transform (FFT)-based methods, and a Maxwell solver is provided for the case of three-dimensional periodic boundary conditions. These methods are directly compared with respect to complexity, scalability, performance, and accuracy. To ensure comparable conditions for all methods and to cover typical applications, we tested all methods on the same set of computers using identical benchmark systems. Our findings suggest that, depending on system size and desired accuracy, the FMM- and FFT-based methods are most efficient in performance and stability. PMID:24483585

  9. SSEL1.0. Sandia Scalable Encryption Software

    SciTech Connect

    Tarman, T.D.

    1996-08-29

    Sandia Scalable Encryption Library (SSEL) Version 1.0 is a library of functions that implement Sandia`s scalable encryption algorithm. This algorithm is used to encrypt Asynchronous Transfer Mode (ATM) data traffic, and is capable of operating on an arbitrary number of bits at a time (which permits scaling via parallel implementations), while being interoperable with differently scaled versions of this algorithm. The routines in this library implement 8 bit and 32 bit versions of a non-linear mixer which is compatible with Sandia`s hardware-based ATM encryptor.

  10. Scalable File Systems for High Performance Computing Final Report

    SciTech Connect

    Brandt, S A

    2007-10-03

    Simulations of mode I interlaminar fracture toughness tests of a carbon-reinforced composite material (BMS 8-212) were conducted with LSDYNA. The fracture toughness tests were performed by U.C. Berkeley. The simulations were performed to investigate the validity and practicality of employing decohesive elements to represent interlaminar bond failures that are prevalent in carbon-fiber composite structure penetration events. The simulations employed a decohesive element formulation that was verified on a simple two element model before being employed to perform the full model simulations. Care was required during the simulations to ensure that the explicit time integration of LSDYNA duplicate the near steady-state testing conditions. In general, this study validated the use of employing decohesive elements to represent the interlaminar bond failures seen in carbon-fiber composite structures, but the practicality of employing the elements to represent the bond failures seen in carbon-fiber composite structures during penetration events was not established.

  11. Highly scalable linear solvers on thousands of processors.

    SciTech Connect

    Domino, Stefan Paul; Karlin, Ian; Siefert, Christopher; Hu, Jonathan Joseph; Robinson, Allen Conrad; Tuminaro, Raymond Stephen

    2009-09-01

    In this report we summarize research into new parallel algebraic multigrid (AMG) methods. We first provide a introduction to parallel AMG. We then discuss our research in parallel AMG algorithms for very large scale platforms. We detail significant improvements in the AMG setup phase to a matrix-matrix multiplication kernel. We present a smoothed aggregation AMG algorithm with fewer communication synchronization points, and discuss its links to domain decomposition methods. Finally, we discuss a multigrid smoothing technique that utilizes two message passing layers for use on multicore processors.

  12. Improved performance of quantum cascade lasers through a scalable, manufacturable epitaxial-side-down mounting process

    PubMed Central

    Tsekoun, Alexei; Go, Rowel; Pushkarsky, Michael; Razeghi, Manijeh; Patel, C. Kumar N.

    2006-01-01

    We report substantially improved performance of high-power quantum cascade lasers (QCLs) by using epitaxial-side-down mounting that provides superior heat dissipation properties. We used aluminum nitride as the heatsink material and gold–tin eutectic solder. We have obtained continuous wave power output of 450 mW at 20°C from mid-IR QCLs. The improved thermal management achieved with epitaxial-side-down mounting combined with a highly manufacturable and scalable assembly process should permit incorporation of mid-IR QCLs in reliable instrumentation. PMID:16547130

  13. Complexity analysis of scalable motion-compensated wavelet video decoders

    NASA Astrophysics Data System (ADS)

    Landge, Gouri; van der Schaar, Mihaela; Akella, Venkatesh

    2004-11-01

    Scalable wavelet video coders based on Motion Compensated Temporal Filtering (MCTF) have been shown to exhibit good coding efficiency over a large range of bit-rates, in addition to providing spatial, temporal and SNR scalabilities. However, the complexity of these wavelet video coding schemes has not been thoroughly investigated. In this paper, we analyze the computational complexity of a fully-scalable MCTF-based wavelet video decoder that is likely to become part of the emerging MPEG-21 standard. We model the change in computational complexity of various components of the decoder as a function of bit-rate, encoding parameters such as filter types for spatial and temporal decomposition and the number of decomposition levels, and sequence characteristics. A key by-product of our analysis is the observation that fixed-function hardware accelerators are not appropriate for implementing these next generation fully scalable video decoders. The absolute complexity of the various functional units as well as their relative complexity varies depending on the transmission bit-rate, thereby requiring different hardware/software architecture support at different bit-rates. To cope with these variations, a preliminary architecture comprising of a reconfigurable co-processor and a general purpose processor is proposed as an implementation platform for these video decoders. We also propose an algorithm to utilize the co-processor efficiently.

  14. A Scalable Platform for Functional Nanomaterials via Bubble-Bursting.

    PubMed

    Feng, Jie; Nunes, Janine K; Shin, Sangwoo; Yan, Jing; Kong, Yong Lin; Prud'homme, Robert K; Arnaudov, Luben N; Stoyanov, Simeon D; Stone, Howard A

    2016-06-01

    A continuous and scalable bubbling system to generate functional nanodroplets dispersed in a continuous phase is proposed. Scaling up of this system can be achieved by simply tuning the bubbling parameters. This new and versatile system is capable of encapsulating various functional nanomaterials to form functional nanoemulsions and nanoparticles in one step. PMID:27007617

  15. PADMA: PArallel Data Mining Agents for scalable text classification

    SciTech Connect

    Kargupta, H.; Hamzaoglu, I.; Stafford, B.

    1997-03-01

    This paper introduces PADMA (PArallel Data Mining Agents), a parallel agent based system for scalable text classification. PADMA contains modules for (1) parallel data accessing operations, (2) parallel hierarchical clustering, and (3) web-based data visualization. This paper introduces the general architecture of PADMA and presents a detailed description of its different modules.

  16. : A Scalable and Transparent System for Simulating MPI Programs

    SciTech Connect

    Perumalla, Kalyan S

    2010-01-01

    is a scalable, transparent system for experimenting with the execution of parallel programs on simulated computing platforms. The level of simulated detail can be varied for application behavior as well as for machine characteristics. Unique features of are repeatability of execution, scalability to millions of simulated (virtual) MPI ranks, scalability to hundreds of thousands of host (real) MPI ranks, portability of the system to a variety of host supercomputing platforms, and the ability to experiment with scientific applications whose source-code is available. The set of source-code interfaces supported by is being expanded to support a wider set of applications, and MPI-based scientific computing benchmarks are being ported. In proof-of-concept experiments, has been successfully exercised to spawn and sustain very large-scale executions of an MPI test program given in source code form. Low slowdowns are observed, due to its use of purely discrete event style of execution, and due to the scalability and efficiency of the underlying parallel discrete event simulation engine, sik. In the largest runs, has been executed on up to 216,000 cores of a Cray XT5 supercomputer, successfully simulating over 27 million virtual MPI ranks, each virtual rank containing its own thread context, and all ranks fully synchronized by virtual time.

  17. Factor-Analytic Procedures for Assessing Response Pattern Scalability

    ERIC Educational Resources Information Center

    Ferrando, Pere J.

    2007-01-01

    This paper proposes procedures for assessing the fit of a psychometric model at the level of the individual respondent. The procedures are intended for personality measures made up of Likert-type items, which, in applied research, are usually analyzed by means of factor analysis. Two scalability indices are proposed, which can be considered as…

  18. Industry Scalable Commercial Lighting Solutions for the Mainstream Market

    SciTech Connect

    Jones, Carol C.; Puranik, Sucheta

    2008-08-17

    Inevitably the greatest obstacles to deep energy savings and mainstream market transformation include complexity and cost. Currently there is a tremendous marketplace gap between the need for widespread integrated lighting solutions and the capacity of the market to provide them. This paper will describe how a new USDOE commercial lighting program provides a multi-faceted strategy to provide the needed “how to” guidance in support of the numerous mandates and programs that are reaching far beyond codes and standards. The program provides lighting energy-efficiency solutions using high performance products, daylighting, and lighting controls. These lighting solutions are widely applicable to common spaces and are delivered via an interactive webtool, making them scalable to the mainstream market. Complexity is reduced by providing pre-designed vignettes and controls strategies that can be reviewed and selected by the end user or design team. The webtool provides analysis and documentation to show performance against energy goals in support of end-user applications for incentives, which addresses the cost obstacle. Utilities and Energy Effiency Program Sponsors (EEPS) benefit by having actionable guidance for customers and energy analysis sufficient to create programs designed around kWh rather than LPD or component-based rebates. The program is organized around the major commercial market sectors: retail, commercial real estate (e.g., offices, developers, lodging), and institutional (e.g., healthcare, education). This allows design solutions to be developed specifically for each sector with the input of the appropriate end users. The partnership model for the program is robust (including end users, design professionals, manufacturers, Non-Governmental Organizations (NGO’s), and EEPS) and provides the network by which feedback is gathered, lighting solutions are deployed, and performance is measured.

  19. Protocol design for scalable and reliable group rekeying

    NASA Astrophysics Data System (ADS)

    Zhang, Xincheng B.; Lam, Simon S.; Lee, Dong Y.; Yang, Yang R.

    2001-07-01

    We present the design and specification of a scalable and reliable protocol for group rekeying together with performance evaluation results. The protocol is based upon the use of key trees for secure groups and periodic batch rekeying. At the beginning of each rekey period, the key server sends a rekey message to all users consisting of encrypted new keys (encryptions, in short) carried in a sequence of packets. We present a simple strategy for identifying keys, encryptions, and users, and a key assignment algorithm which ensures that the encryptions needed by a user are in the same packet. Our protocol provides reliable delivery of new keys to all users eventually. It also attempts to deliver new keys to all users with a high probability by the end of the rekeying period. For each rekey message, the protocol runs in two steps: a multicast step followed by a unicast step. Proactive FEC multicast is used to control NACK implosion and reduce delivery latency. Our experiments show that a small FEC block size can be used to reduce encoding time at the server without increasing server bandwidth overhead. Early transition to unicast, after at most two multicast rounds, further reduces the worst-case delivery latency as well as user bandwidth requirement. The key server adaptively adjusts the proactivity factor based upon past feedback information; our experiments show that the number of NACKs after a multicast round can be effectively controlled around a target number. Throughout the protocol design, we strive to minimize processing and bandwidth requirements for both the key server and users.

  20. Analysis of prediction algorithms for residual compression in a lossy to lossless scalable video coding system based on HEVC

    NASA Astrophysics Data System (ADS)

    Heindel, Andreas; Wige, Eugen; Kaup, André

    2014-09-01

    Lossless image and video compression is required in many professional applications. However, lossless coding results in a high data rate, which leads to a long wait for the user when the channel capacity is limited. To overcome this problem, scalable lossless coding is an elegant solution. It provides a fast accessible preview by a lossy compressed base layer, which can be refined to a lossless output when the enhancement layer is received. Therefore, this paper presents a lossy to lossless scalable coding system where the enhancement layer is coded by means of intra prediction and entropy coding. Several algorithms are evaluated for the prediction step in this paper. It turned out that Sample-based Weighted Prediction is a reasonable choice for usual consumer video sequences and the Median Edge Detection algorithm is better suited for medical content from computed tomography. For both types of sequences the efficiency may be further improved by the much more complex Edge-Directed Prediction algorithm. In the best case, in total only about 2.7% additional data rate has to be invested for scalable coding compared to single-layer JPEG-LS compression for usual consumer video sequences. For the case of the medical sequences scalable coding is even more efficient than JPEG-LS compression for certain values of QP.

  1. A Novel Coarsening Method for Scalable and Efficient Mesh Generation

    SciTech Connect

    Yoo, A; Hysom, D; Gunney, B

    2010-12-02

    matrix-vector multiplication can be performed locally on each processor and hence to minimize communication. Furthermore, a good graph partitioning scheme ensures the equal amount of computation performed on each processor. Graph partitioning is a well known NP-complete problem, and thus the most commonly used graph partitioning algorithms employ some forms of heuristics. These algorithms vary in terms of their complexity, partition generation time, and the quality of partitions, and they tend to trade off these factors. A significant challenge we are currently facing at the Lawrence Livermore National Laboratory is how to partition very large meshes on massive-size distributed memory machines like IBM BlueGene/P, where scalability becomes a big issue. For example, we have found that the ParMetis, a very popular graph partitioning tool, can only scale to 16K processors. An ideal graph partitioning method on such an environment should be fast and scale to very large meshes, while producing high quality partitions. This is an extremely challenging task, as to scale to that level, the partitioning algorithm should be simple and be able to produce partitions that minimize inter-processor communications and balance the load imposed on the processors. Our goals in this work are two-fold: (1) To develop a new scalable graph partitioning method with good load balancing and communication reduction capability. (2) To study the performance of the proposed partitioning method on very large parallel machines using actual data sets and compare the performance to that of existing methods. The proposed method achieves the desired scalability by reducing the mesh size. For this, it coarsens an input mesh into a smaller size mesh by coalescing the vertices and edges of the original mesh into a set of mega-vertices and mega-edges. A new coarsening method called brick algorithm is developed in this research. In the brick algorithm, the zones in a given mesh are first grouped into fixed size

  2. More scalability, less pain : A simple programming model and its implementation for extreme computing.

    SciTech Connect

    Lusk, E. L.; Pieper, S. C.; Butler, R. M.; Middle Tennessee State Univ.

    2010-01-01

    This is the story of a simple programming model, its implementation for extreme computing, and a breakthrough in nuclear physics. A critical issue for the future of high-performance computing is the programming model to use on next-generation architectures. Described here is a promising approach: program very large machines by combining a simplified programming model with a scalable library implementation. The presentation takes the form of a case study in nuclear physics. The chosen application addresses fundamental issues in the origins of our Universe, while the library developed to enable this application on the largest computers may have applications beyond this one.

  3. Iterative methods for large scale static analysis of structures on a scalable multiprocessor supercomputer

    NASA Technical Reports Server (NTRS)

    Sobh, Nahil Atef

    1992-01-01

    A parallel Preconditioned Conjugate Gradient (PCG) iterative solver has been developed and implemented on the iPSC-860 scalable hypercube. This new implementation makes use of the Parallel Automated Runtime Toolkit at ICASE (PARTI) primitives to efficiently program irregular communications patterns that exist in general sparse matrices and in particular in the finite element sparse stiffness matrices. The iterative PCG has been used to solve the finite element equations that result from discretizing large scale aerospace structures. In particular, the static response of the High Speed Civil Transport (HSCT) finite element model is solved on the iPSC-860.

  4. NYU3T: teaching, technology, teamwork: a model for interprofessional education scalability and sustainability.

    PubMed

    Djukic, Maja; Fulmer, Terry; Adams, Jennifer G; Lee, Sabrina; Triola, Marc M

    2012-09-01

    Interprofessional education is a critical precursor to effective teamwork and the collaboration of health care professionals in clinical settings. Numerous barriers have been identified that preclude scalable and sustainable interprofessional education (IPE) efforts. This article describes NYU3T: Teaching, Technology, Teamwork, a model that uses novel technologies such as Web-based learning, virtual patients, and high-fidelity simulation to overcome some of the common barriers and drive implementation of evidence-based teamwork curricula. It outlines the program's curricular components, implementation strategy, evaluation methods, and lessons learned from the first year of delivery and describes implications for future large-scale IPE initiatives. PMID:22920424

  5. The NIDS Cluster: Scalable, Stateful Network Intrusion Detection on Commodity Hardware

    SciTech Connect

    Tierney, Brian L; Vallentin, Matthias; Sommer, Robin; Lee, Jason; Leres, Craig; Paxson, Vern; Tierney, Brian

    2007-09-19

    In this work we present a NIDS cluster as a scalable solution for realizing high-performance, stateful network intrusion detection on commodity hardware. The design addresses three challenges: (i) distributing traffic evenly across an extensible set of analysis nodes in a fashion that minimizes the communication required for coordination, (ii) adapting the NIDS's operation to support coordinating its low-level analysis rather than just aggregating alerts; and (iii) validating that the cluster produces sound results. Prototypes of our NIDS cluster now operate at the Lawrence Berkeley National Laboratory and the University of California at Berkeley. In both environments the clusters greatly enhance the power of the network security monitoring.

  6. Bioinspired superhydrophobic surfaces, fabricated through simple and scalable roll-to-roll processing

    NASA Astrophysics Data System (ADS)

    Park, Sung-Hoon; Lee, Sangeui; Moreira, David; Bandaru, Prabhakar R.; Han, Intaek; Yun, Dong-Jin

    2015-10-01

    A simple, scalable, non-lithographic, technique for fabricating durable superhydrophobic (SH) surfaces, based on the fingering instabilities associated with non-Newtonian flow and shear tearing, has been developed. The high viscosity of the nanotube/elastomer paste has been exploited for the fabrication. The fabricated SH surfaces had the appearance of bristled shark skin and were robust with respect to mechanical forces. While flow instability is regarded as adverse to roll-coating processes for fabricating uniform films, we especially use the effect to create the SH surface. Along with their durability and self-cleaning capabilities, we have demonstrated drag reduction effects of the fabricated films through dynamic flow measurements.

  7. Scalable dielectrophoresis of single walled carbon nanotubes

    NASA Astrophysics Data System (ADS)

    Fitzhugh, William A.

    Single Walled Carbon Nanotubes (SWNTs) have attracted much attention as a candidate material for future nano-scale 'beyond silicon' devices. However industrial scale operations have been impeded by difficulties in separating the metallic and semiconducting species. This paper addresses the use of highly inhomogeneous alternating electric fields, dielectrophoresis, to isolate SWNT species in scaled systems. Both numerical and experimental methods will be discussed.

  8. Horizon: The Portable, Scalable, and Reusable Framework for Developing Automated Data Management and Product Generation Systems

    NASA Astrophysics Data System (ADS)

    Huang, T.; Alarcon, C.; Quach, N. T.

    2014-12-01

    Capture, curate, and analysis are the typical activities performed at any given Earth Science data center. Modern data management systems must be adaptable to heterogeneous science data formats, scalable to meet the mission's quality of service requirements, and able to manage the life-cycle of any given science data product. Designing a scalable data management doesn't happen overnight. It takes countless hours of refining, refactoring, retesting, and re-architecting. The Horizon data management and workflow framework, developed at the Jet Propulsion Laboratory, is a portable, scalable, and reusable framework for developing high-performance data management and product generation workflow systems to automate data capturing, data curation, and data analysis activities. The NASA's Physical Oceanography Distributed Active Archive Center (PO.DAAC)'s Data Management and Archive System (DMAS) is its core data infrastructure that handles capturing and distribution of hundreds of thousands of satellite observations each day around the clock. DMAS is an application of the Horizon framework. The NASA Global Imagery Browse Services (GIBS) is NASA's Earth Observing System Data and Information System (EOSDIS)'s solution for making high-resolution global imageries available to the science communities. The Imagery Exchange (TIE), an application of the Horizon framework, is a core subsystem for GIBS responsible for data capturing and imagery generation automation to support the EOSDIS' 12 distributed active archive centers and 17 Science Investigator-led Processing Systems (SIPS). This presentation discusses our ongoing effort in refining, refactoring, retesting, and re-architecting the Horizon framework to enable data-intensive science and its applications.

  9. GPU-FS-kNN: A Software Tool for Fast and Scalable kNN Computation Using GPUs

    PubMed Central

    Arefin, Ahmed Shamsul; Riveros, Carlos; Berretta, Regina; Moscato, Pablo

    2012-01-01

    Background The analysis of biological networks has become a major challenge due to the recent development of high-throughput techniques that are rapidly producing very large data sets. The exploding volumes of biological data are craving for extreme computational power and special computing facilities (i.e. super-computers). An inexpensive solution, such as General Purpose computation based on Graphics Processing Units (GPGPU), can be adapted to tackle this challenge, but the limitation of the device internal memory can pose a new problem of scalability. An efficient data and computational parallelism with partitioning is required to provide a fast and scalable solution to this problem. Results We propose an efficient parallel formulation of the k-Nearest Neighbour (kNN) search problem, which is a popular method for classifying objects in several fields of research, such as pattern recognition, machine learning and bioinformatics. Being very simple and straightforward, the performance of the kNN search degrades dramatically for large data sets, since the task is computationally intensive. The proposed approach is not only fast but also scalable to large-scale instances. Based on our approach, we implemented a software tool GPU-FS-kNN (GPU-based Fast and Scalable k-Nearest Neighbour) for CUDA enabled GPUs. The basic approach is simple and adaptable to other available GPU architectures. We observed speed-ups of 50–60 times compared with CPU implementation on a well-known breast microarray study and its associated data sets. Conclusion Our GPU-based Fast and Scalable k-Nearest Neighbour search technique (GPU-FS-kNN) provides a significant performance improvement for nearest neighbour computation in large-scale networks. Source code and the software tool is available under GNU Public License (GPL) at https://sourceforge.net/p/gpufsknn/. PMID:22937144

  10. Scalable and portable visualization of large atomistic datasets

    NASA Astrophysics Data System (ADS)

    Sharma, Ashish; Kalia, Rajiv K.; Nakano, Aiichiro; Vashishta, Priya

    2004-10-01

    A scalable and portable code named Atomsviewer has been developed to interactively visualize a large atomistic dataset consisting of up to a billion atoms. The code uses a hierarchical view frustum-culling algorithm based on the octree data structure to efficiently remove atoms outside of the user's field-of-view. Probabilistic and depth-based occlusion-culling algorithms then select atoms, which have a high probability of being visible. Finally a multiresolution algorithm is used to render the selected subset of visible atoms at varying levels of detail. Atomsviewer is written in C++ and OpenGL, and it has been tested on a number of architectures including Windows, Macintosh, and SGI. Atomsviewer has been used to visualize tens of millions of atoms on a standard desktop computer and, in its parallel version, up to a billion atoms. Program summaryTitle of program: Atomsviewer Catalogue identifier: ADUM Program summary URL:http://cpc.cs.qub.ac.uk/summaries/ADUM Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Computer for which the program is designed and others on which it has been tested: 2.4 GHz Pentium 4/Xeon processor, professional graphics card; Apple G4 (867 MHz)/G5, professional graphics card Operating systems under which the program has been tested: Windows 2000/XP, Mac OS 10.2/10.3, SGI IRIX 6.5 Programming languages used: C++, C and OpenGL Memory required to execute with typical data: 1 gigabyte of RAM High speed storage required: 60 gigabytes No. of lines in the distributed program including test data, etc.: 550 241 No. of bytes in the distributed program including test data, etc.: 6 258 245 Number of bits in a word: Arbitrary Number of processors used: 1 Has the code been vectorized or parallelized: No Distribution format: tar gzip file Nature of physical problem: Scientific visualization of atomic systems Method of solution: Rendering of atoms using computer graphic techniques, culling algorithms for data

  11. Coupled Workflow Approach to Scalable Multi-Scale Subsurface Simulations

    NASA Astrophysics Data System (ADS)

    Schuchardt, K.; Agarwal, K.; Chase, J.; Scheibe, T. D.; Palmer, B. J.; Tartakovsky, A. M.

    2012-12-01

    Continuum scale models have been used to study subsurface flow, transport, and reactions for many years but lack the capability to resolve fine-grained processes. Recently, pore scale models, which operate at scales of individual soil grains, have been developed to more accurately model pore scale phenomena, such as mineral precipitation and dissolution reactions. Pore-scale methods to study microbially-mediated surface reactions are also under development. However, these finer-grained models are prohibitively expensive for modeling realistic domains. To apply these new techniques to realistic systems, we are developing a hybrid, multi-scale model which initially simulates the full domain at continuum scale and applies the pore scale model only to areas of high reactivity. Since the location and number of pore regions in the model varies as the reactions proceed, an adaptive Pore Generator code defines the number and location of pore regions at each continuum iteration. A fourth code provides data transformation from the pore scale back to the continuum scale. These components are loosely coupled into a single multi-scale model using the Swift workflow system for two reasons: we have several choices of simulators for the continuum and pore codes, and the mathematical model integration methods, data transformations, and adaptive behavior are all highly problem dependent at this time. Our approach provides a framework to solve new problems by replacing individual components. Our initial experiment simulates the parallel transport of two solutes with an irreversible mixing-controlled kinetic bimolecular reaction occurring at the interface between the two solutes. The system is filled with porous medium like sand. The sand is saturated with water, and two solutes (denoted A and B) are injected at the bottom and flow to the top at a specified rate. As the solutions flow upward through the flow cell, they mix along the centerline, leading to reaction and formation of a

  12. Scalable and Environmentally Benign Process for Smart Textile Nanofinishing.

    PubMed

    Feng, Jicheng; Hontañón, Esther; Blanes, Maria; Meyer, Jörg; Guo, Xiaoai; Santos, Laura; Paltrinieri, Laura; Ramlawi, Nabil; Smet, Louis C P M de; Nirschl, Hermann; Kruis, Frank Einar; Schmidt-Ott, Andreas; Biskos, George

    2016-06-15

    A major challenge in nanotechnology is that of determining how to introduce green and sustainable principles when assembling individual nanoscale elements to create working devices. For instance, textile nanofinishing is restricted by the many constraints of traditional pad-dry-cure processes, such as the use of costly chemical precursors to produce nanoparticles (NPs), the high liquid and energy consumption, the production of harmful liquid wastes, and multistep batch operations. By integrating low-cost, scalable, and environmentally benign aerosol processes of the type proposed here into textile nanofinishing, these constraints can be circumvented while leading to a new class of fabrics. The proposed one-step textile nanofinishing process relies on the diffusional deposition of aerosol NPs onto textile fibers. As proof of this concept, we deposit Ag NPs onto a range of textiles and assess their antimicrobial properties for two strains of bacteria (i.e., Staphylococcus aureus and Klebsiella pneumoniae). The measurements show that the logarithmic reduction in bacterial count can get as high as ca. 5.5 (corresponding to a reduction efficiency of 99.96%) when the Ag loading is 1 order of magnitude less (10 ppm; i.e., 10 mg Ag NPs per kg of textile) than that of textiles treated by traditional wet-routes. The antimicrobial activity does not increase in proportion to the Ag content above 10 ppm as a consequence of a "saturation" effect. Such low NP loadings on antimicrobial textiles minimizes the risk to human health (during textile use) and to the ecosystem (after textile disposal), as well as it reduces potential changes in color and texture of the resulting textile products. After three washes, the release of Ag is in the order of 1 wt %, which is comparable to textiles nanofinished with wet routes using binders. Interestingly, the washed textiles exhibit almost no reduction in antimicrobial activity, much as those of as-deposited samples. Considering that a realm

  13. Wearable energy-dense and power-dense supercapacitor yarns enabled by scalable graphene-metallic textile composite electrodes

    NASA Astrophysics Data System (ADS)

    Liu, Libin; Yu, You; Yan, Casey; Li, Kan; Zheng, Zijian

    2015-06-01

    One-dimensional flexible supercapacitor yarns are of considerable interest for future wearable electronics. The bottleneck in this field is how to develop devices of high energy and power density, by using economically viable materials and scalable fabrication technologies. Here we report a hierarchical graphene-metallic textile composite electrode concept to address this challenge. The hierarchical composite electrodes consist of low-cost graphene sheets immobilized on the surface of Ni-coated cotton yarns, which are fabricated by highly scalable electroless deposition of Ni and electrochemical deposition of graphene on commercial cotton yarns. Remarkably, the volumetric energy density and power density of the all solid-state supercapacitor yarn made of one pair of these composite electrodes are 6.1 mWh cm-3 and 1,400 mW cm-3, respectively. In addition, this SC yarn is lightweight, highly flexible, strong, durable in life cycle and bending fatigue tests, and integratable into various wearable electronic devices.

  14. Towards reproducible, scalable lateral molecular electronic devices

    NASA Astrophysics Data System (ADS)

    Durkan, Colm; Zhang, Qian

    2014-08-01

    An approach to reproducibly fabricate molecular electronic devices is presented. Lateral nanometer-scale gaps with high yield are formed in Au/Pd nanowires by a combination of electromigration and Joule-heating-induced thermomechanical stress. The resulting nanogap devices are used to measure the electrical properties of small numbers of two different molecular species with different end-groups, namely 1,4-butane dithiol and 1,5-diamino-2-methylpentane. Fluctuations in the current reveal that in the case of the dithiol molecule devices, individual molecules conduct intermittently, with the fluctuations becoming more pronounced at larger biases.

  15. Two Scalable Syntheses of (S)-2-Methylazetidine.

    PubMed

    Dowling, Matthew S; Fernando, Dilinie P; Hou, Jie; Liu, Bo; Smith, Aaron C

    2016-04-01

    Two orthogonal routes for preparing (S)-2-methylazetidine as a bench stable, crystalline (R)-(-)-CSA salt are presented. One route features the in situ generation and cyclization of a 1,3-bis-triflate to form the azetidine ring, while the second route involves chemoselective reduction of N-Boc azetidine-2-carboxylic acid. Both sequences afford the desired product in good overall yields (61% and 49%) and high enantiomeric excess (>99% ee), avoid column chromatography, and are suitable for the large-scale production of this material. PMID:26895201

  16. Scalable lidar technique for fire detection

    NASA Astrophysics Data System (ADS)

    Utkin, Andrei B.; Piedade, Fernando; Beixiga, Vasco; Mota, Pedro; Lousã, Pedro

    2014-08-01

    Lidar (light detection and ranging) presents better sensitivity than fire surveillance based on imaging. However, the price of conventional lidar equipment is often too high as compared to passive fire detection instruments. We describe possibilities to downscale the technology. First, a conventional lidar, capable of smoke-plume detection up to ~10 km, may be replaced by an industrially manufactured solid-state laser rangefinder. This reduces the detection range to about 5 km, but decreases the purchase price by one order of magnitude. Further downscaling is possible by constructing the lidar smoke sensor on the basis of a low-cost laser diode.

  17. Towards reproducible, scalable lateral molecular electronic devices

    SciTech Connect

    Durkan, Colm Zhang, Qian

    2014-08-25

    An approach to reproducibly fabricate molecular electronic devices is presented. Lateral nanometer-scale gaps with high yield are formed in Au/Pd nanowires by a combination of electromigration and Joule-heating-induced thermomechanical stress. The resulting nanogap devices are used to measure the electrical properties of small numbers of two different molecular species with different end-groups, namely 1,4-butane dithiol and 1,5-diamino-2-methylpentane. Fluctuations in the current reveal that in the case of the dithiol molecule devices, individual molecules conduct intermittently, with the fluctuations becoming more pronounced at larger biases.

  18. Scalable NMR spectroscopy with semiconductor chips

    PubMed Central

    Ha, Dongwan; Paulsen, Jeffrey; Sun, Nan; Song, Yi-Qiao; Ham, Donhee

    2014-01-01

    State-of-the-art NMR spectrometers using superconducting magnets have enabled, with their ultrafine spectral resolution, the determination of the structure of large molecules such as proteins, which is one of the most profound applications of modern NMR spectroscopy. Many chemical and biotechnological applications, however, involve only small-to-medium size molecules, for which the ultrafine resolution of the bulky, expensive, and high-maintenance NMR spectrometers is not required. For these applications, there is a critical need for portable, affordable, and low-maintenance NMR spectrometers to enable in-field, on-demand, or online applications (e.g., quality control, chemical reaction monitoring) and co-use of NMR with other analytical methods (e.g., chromatography, electrophoresis). As a critical step toward NMR spectrometer miniaturization, small permanent magnets with high field homogeneity have been developed. In contrast, NMR spectrometer electronics capable of modern multidimensional spectroscopy have thus far remained bulky. Complementing the magnet miniaturization, here we integrate the NMR spectrometer electronics into 4-mm2 silicon chips. Furthermore, we perform various multidimensional NMR spectroscopies by operating these spectrometer electronics chips together with a compact permanent magnet. This combination of the spectrometer-electronics-on-a-chip with a permanent magnet represents a useful step toward miniaturization of the overall NMR spectrometer into a portable platform. PMID:25092330

  19. Scalable NMR spectroscopy with semiconductor chips.

    PubMed

    Ha, Dongwan; Paulsen, Jeffrey; Sun, Nan; Song, Yi-Qiao; Ham, Donhee

    2014-08-19

    State-of-the-art NMR spectrometers using superconducting magnets have enabled, with their ultrafine spectral resolution, the determination of the structure of large molecules such as proteins, which is one of the most profound applications of modern NMR spectroscopy. Many chemical and biotechnological applications, however, involve only small-to-medium size molecules, for which the ultrafine resolution of the bulky, expensive, and high-maintenance NMR spectrometers is not required. For these applications, there is a critical need for portable, affordable, and low-maintenance NMR spectrometers to enable in-field, on-demand, or online applications (e.g., quality control, chemical reaction monitoring) and co-use of NMR with other analytical methods (e.g., chromatography, electrophoresis). As a critical step toward NMR spectrometer miniaturization, small permanent magnets with high field homogeneity have been developed. In contrast, NMR spectrometer electronics capable of modern multidimensional spectroscopy have thus far remained bulky. Complementing the magnet miniaturization, here we integrate the NMR spectrometer electronics into 4-mm(2) silicon chips. Furthermore, we perform various multidimensional NMR spectroscopies by operating these spectrometer electronics chips together with a compact permanent magnet. This combination of the spectrometer-electronics-on-a-chip with a permanent magnet represents a useful step toward miniaturization of the overall NMR spectrometer into a portable platform. PMID:25092330

  20. ParaText : scalable text analysis and visualization.

    SciTech Connect

    Dunlavy, Daniel M.; Stanton, Eric T.; Shead, Timothy M.

    2010-07-01

    Automated analysis of unstructured text documents (e.g., web pages, newswire articles, research publications, business reports) is a key capability for solving important problems in areas including decision making, risk assessment, social network analysis, intelligence analysis, scholarly research and others. However, as data sizes continue to grow in these areas, scalable processing, modeling, and semantic analysis of text collections becomes essential. In this paper, we present the ParaText text analysis engine, a distributed memory software framework for processing, modeling, and analyzing collections of unstructured text documents. Results on several document collections using hundreds of processors are presented to illustrate the exibility, extensibility, and scalability of the the entire process of text modeling from raw data ingestion to application analysis.

  1. A scalable quantum architecture using efficient non-local gates

    NASA Astrophysics Data System (ADS)

    Brennen, Gavin

    2003-03-01

    Many protocols for quantum information processing use a control sequence or circuit of interactions between qubits and control fields wherein arbitrary qubits can be made to interact with one another. The primary problem with many ``physically scalable" architectures is that the qubits are restricted to nearest neighbor interactions and quantum wires between distant qubits do not exist. Because of errors, nearest neighbor interactions often present difficulty with scalability. We describe a protocol that efficiently performs non-local gates between elements of separated static logical qubits using a bus of dynamic qubits as a refreshable entanglement resource. Imperfect resource preparation due to error propagation from noisy gates and measurement errors can purified within the bus channel. Because of the inherent parallelism of entanglement swapping, communication latency within the quantum computer can be significantly reduced.

  2. A look at scalable dense linear algebra libraries

    SciTech Connect

    Dongarra, J.J. |; van de Geijn, R.; Walker, D.W.

    1992-07-01

    We discuss the essential design features of a library of scalable software for performing dense linear algebra computations on distributed memory concurrent computers. The square block scattered decomposition is proposed as a flexible and general-purpose way of decomposing most, if not all, dense matrix problems. An object- oriented interface to the library permits more portable applications to be written, and is easy to learn and use, since details of the parallel implementation are hidden from the user. Experiments on the Intel Touchstone Delta system with a prototype code that uses the square block scattered decomposition to perform LU factorization are presented and analyzed. It was found that the code was both scalable and efficient, performing at about 14 Gflop/s (double precision) for the largest problem considered.

  3. ATLAS Grid Data Processing: system evolution and scalability

    NASA Astrophysics Data System (ADS)

    Golubkov, D.; Kersevan, B.; Klimentov, A.; Minaenko, A.; Nevski, P.; Vaniachine, A.; Walker, R.

    2012-12-01

    The production system for Grid Data Processing handles petascale ATLAS data reprocessing and Monte Carlo activities. The production system empowered further data processing steps on the Grid performed by dozens of ATLAS physics groups with coordinated access to computing resources worldwide, including additional resources sponsored by regional facilities. The system provides knowledge management of configuration parameters for massive data processing tasks, reproducibility of results, scalable database access, orchestrated workflow and performance monitoring, dynamic workload sharing, automated fault tolerance and petascale data integrity control. The system evolves to accommodate a growing number of users and new requirements from our contacts in ATLAS main areas: Trigger, Physics, Data Preparation and Software & Computing. To assure scalability, the next generation production system architecture development is in progress. We report on scaling up the production system for a growing number of users providing data for physics analysis and other ATLAS main activities.

  4. Scalable fabrication of triboelectric nanogenerators for commercial applications

    NASA Astrophysics Data System (ADS)

    Dhakar, Lokesh; Shan, Xuechuan; Wang, Zhiping; Yang, Bin; Eng Hock Tay, Francis; Heng, Chun-Huat; Lee, Chengkuo

    2015-12-01

    Harvesting mechanical energy from irregular sources is a potential way to charge batteries for devices and sensor nodes. Triboelectric effect has been extensively utilized in energy harvesting devices as a method to convert mechanical energy into electrical energy. As triboelectric nanogenerators have immense potential to be commercialized, it is important to develop scalable fabrication methods to manufacture these devices. This paper presents scalable fabrication steps to realize large scale triboelectric nanogenerators. Roll-to-roll UV embossing and lamination techniques are used to fabricate different components of large scale triboelectric nanogenerators. The device generated a peak-to-peak voltage and current of 486 V and 21.2 μA, respectively at a frequency of 5 Hz.

  5. Scalable Network Emulator Architecture for IP Optical Network Management

    NASA Astrophysics Data System (ADS)

    Oki, Eiji; Kitsuwan, Nattapong; Tsunoda, Shunichi; Miyamura, Takashi; Masuda, Akeo; Shiomoto, Kohei

    This letter proposes a scalable network emulator architecture to support IP optical network management. The network emulator uses the same router interfaces to communicate with the IP optical TE server as the actual IP optical network, and behaves as an actual IP optical network between the interfaces. The network emulator mainly consists of databases and three modules: interface module, resource simulator module, and traffic generator module. To make the network emulator scalable in terms of network size, we employ TCP/IP socket communications between the modules. The proposed network emulator has the benefit that its implementation is not strongly dependent on hardware limitations. We develop a prototype of the network emulator based on the proposed architecture. Our design and experiments show that the proposed architecture is effective.

  6. Scalable Computation of Streamlines on Very Large Datasets

    SciTech Connect

    Pugmire, David; Childs, Hank; Garth, Christoph; Ahern, Sean; Weber, Gunther H.

    2009-09-01

    Understanding vector fields resulting from large scientific simulations is an important and often difficult task. Streamlines, curves that are tangential to a vector field at each point, are a powerful visualization method in this context. Application of streamline-based visualization to very large vector field data represents a significant challenge due to the non-local and data-dependent nature of streamline computation, and requires careful balancing of computational demands placed on I/O, memory, communication, and processors. In this paper we review two parallelization approaches based on established parallelization paradigms (static decomposition and on-demand loading) and present a novel hybrid algorithm for computing streamlines. Our algorithm is aimed at good scalability and performance across the widely varying computational characteristics of streamline-based problems. We perform performance and scalability studies of all three algorithms on a number of prototypical application problems and demonstrate that our hybrid scheme is able to perform well in different settings.

  7. Scalable digital hardware for a trapped ion quantum computer

    NASA Astrophysics Data System (ADS)

    Mount, Emily; Gaultney, Daniel; Vrijsen, Geert; Adams, Michael; Baek, So-Young; Hudek, Kai; Isabella, Louis; Crain, Stephen; van Rynbach, Andre; Maunz, Peter; Kim, Jungsang

    2015-09-01

    Many of the challenges of scaling quantum computer hardware lie at the interface between the qubits and the classical control signals used to manipulate them. Modular ion trap quantum computer architectures address scalability by constructing individual quantum processors interconnected via a network of quantum communication channels. Successful operation of such quantum hardware requires a fully programmable classical control system capable of frequency stabilizing the continuous wave lasers necessary for loading, cooling, initialization, and detection of the ion qubits, stabilizing the optical frequency combs used to drive logic gate operations on the ion qubits, providing a large number of analog voltage sources to drive the trap electrodes, and a scheme for maintaining phase coherence among all the controllers that manipulate the qubits. In this work, we describe scalable solutions to these hardware development challenges.

  8. Scalable quantum memory in the ultrastrong coupling regime.

    PubMed

    Kyaw, T H; Felicetti, S; Romero, G; Solano, E; Kwek, L-C

    2015-01-01

    Circuit quantum electrodynamics, consisting of superconducting artificial atoms coupled to on-chip resonators, represents a prime candidate to implement the scalable quantum computing architecture because of the presence of good tunability and controllability. Furthermore, recent advances have pushed the technology towards the ultrastrong coupling regime of light-matter interaction, where the qubit-resonator coupling strength reaches a considerable fraction of the resonator frequency. Here, we propose a qubit-resonator system operating in that regime, as a quantum memory device and study the storage and retrieval of quantum information in and from the Z2 parity-protected quantum memory, within experimentally feasible schemes. We are also convinced that our proposal might pave a way to realize a scalable quantum random-access memory due to its fast storage and readout performances. PMID:25727251

  9. Scalable synthesis and energy applications of defect engineeered nano materials

    NASA Astrophysics Data System (ADS)

    Karakaya, Mehmet

    Nanomaterials and nanotechnologies have attracted a great deal of attention in a few decades due to their novel physical properties such as, high aspect ratio, surface morphology, impurities, etc. which lead to unique chemical, optical and electronic properties. The awareness of importance of nanomaterials has motivated researchers to develop nanomaterial growth techniques to further control nanostructures properties such as, size, surface morphology, etc. that may alter their fundamental behavior. Carbon nanotubes (CNTs) are one of the most promising materials with their rigidity, strength, elasticity and electric conductivity for future applications. Despite their excellent properties explored by the abundant research works, there is big challenge to introduce them into the macroscopic world for practical applications. This thesis first gives a brief overview of the CNTs, it will then go on mechanical and oil absorption properties of macro-scale CNT assemblies, then following CNT energy storage applications and finally fundamental studies of defect introduced graphene systems. Chapter Two focuses on helically coiled carbon nanotube (HCNT) foams in compression. Similarly to other foams, HCNT foams exhibit preconditioning effects in response to cyclic loading; however, their fundamental deformation mechanisms are unique. Bulk HCNT foams exhibit super-compressibility and recover more than 90% of large compressive strains (up to 80%). When subjected to striker impacts, HCNT foams mitigate impact stresses more effectively compared to other CNT foams comprised of non-helical CNTs (~50% improvement). The unique mechanical properties we revealed demonstrate that the HCNT foams are ideally suited for applications in packaging, impact protection, and vibration mitigation. The third chapter describes a simple method for the scalable synthesis of three-dimensional, elastic, and recyclable multi-walled carbon nanotube (MWCNT) based light weight bucky-aerogels (BAGs) that are

  10. Performance and Scalability of the NAS Parallel Benchmarks in Java

    NASA Technical Reports Server (NTRS)

    Frumkin, Michael A.; Schultz, Matthew; Jin, Haoqiang; Yan, Jerry; Biegel, Bryan A. (Technical Monitor)

    2002-01-01

    Several features make Java an attractive choice for scientific applications. In order to gauge the applicability of Java to Computational Fluid Dynamics (CFD), we have implemented the NAS (NASA Advanced Supercomputing) Parallel Benchmarks in Java. The performance and scalability of the benchmarks point out the areas where improvement in Java compiler technology and in Java thread implementation would position Java closer to Fortran in the competition for scientific applications.

  11. HDB-Subdue: A Scalable Approach to Graph Mining

    NASA Astrophysics Data System (ADS)

    Padmanabhan, Srihari; Chakravarthy, Sharma

    Transactional data mining (association rules, decision trees etc.) has been effectively used to find non-trivial patterns in categorical and unstructured data. For applications that have an inherent structure (e.g., social networks, proteins), graph mining is useful since mapping the structured data into a transactional representation will lead to loss of information. Graph mining is used for identifying interesting or frequent subgraphs. Database mining uses SQL and relational representation to overcome limitations of main memory algorithms and to achieve scalability.

  12. Scalable Real Time Data Management for Smart Grid

    SciTech Connect

    Yin, Jian; Kulkarni, Anand V.; Purohit, Sumit; Gorton, Ian; Akyol, Bora A.

    2011-12-16

    This paper presents GridMW, a scalable and reliable data middleware for smart grids. Smart grids promise to improve the efficiency of power grid systems and reduce green house emissions through incorporating power generation from renewable sources and shaping demand to match the supply. As a result, power grid systems will become much more dynamic and require constant adjustments, which requires analysis and decision making applications to improve the efficiency and reliability of smart grid systems.

  13. Scalable wide-field optical coherence tomography-based angiography for in vivo imaging applications

    PubMed Central

    Xu, Jingjiang; Wei, Wei; Song, Shaozhen; Qi, Xiaoli; Wang, Ruikang K.

    2016-01-01

    Recent advances in optical coherence tomography (OCT)-based angiography have demonstrated a variety of biomedical applications in the diagnosis and therapeutic monitoring of diseases with vascular involvement. While promising, its imaging field of view (FOV) is however still limited (typically less than 9 mm2), which somehow slows down its clinical acceptance. In this paper, we report a high-speed spectral-domain OCT operating at 1310 nm to enable wide FOV up to 750 mm2. Using optical microangiography (OMAG) algorithm, we are able to map vascular networks within living biological tissues. Thanks to 2,048 pixel-array line scan InGaAs camera operating at 147 kHz scan rate, the system delivers a ranging depth of ~7.5 mm and provides wide-field OCT-based angiography at a single data acquisition. We implement two imaging modes (i.e., wide-field mode and high-resolution mode) in the OCT system, which gives highly scalable FOV with flexible lateral resolution. We demonstrate scalable wide-field vascular imaging for multiple finger nail beds in human and whole brain in mice with skull left intact at a single 3D scan, promising new opportunities for wide-field OCT-based angiography for many clinical applications. PMID:27231630

  14. Scalable and Direct Growth of Graphene Micro Ribbons on Dielectric Substrates

    PubMed Central

    Wang, Debin; Tian, He; Yang, Yi; Xie, Dan; Ren, Tian-Ling; Zhang, Yuegang

    2013-01-01

    Here we report on a scalable and direct growth of graphene micro ribbons on SiO2 dielectric substrates using a low temperature chemical vapor deposition. Due to the fast annealing at low temperature and dewetting of Ni, continuous few-layer graphene micro ribbons grow directly on bare dielectric substrates through Ni assisted catalytic decomposition of hydrocarbon precursors. These high quality graphene micro ribbons exhibit low sheet resistance of ~700 Ω −2100 Ω, high on/off current ratio of ~3, and high carrier mobility of ~655 cm2V−1s−1 at room temperature, all of which have shown significant improvement over other lithography patterned CVD graphene micro ribbons. This direct approach can in principle form graphene ribbons of any arbitrary sizes and geometries. It allows for a feasible methodology towards better integration with semiconductor materials for interconnect electronics and scalable production for graphene based electronic and optoelectronic applications where the electrical gating is the key enabling factor. PMID:23443152

  15. Institute for Scalable Application Development Software

    SciTech Connect

    Miller, Barton P

    2012-11-14

    Work by the University of Wisconsin as part of the DOE SciDAC CScADS includes the following accomplishments: Research on tool componentization, with concentration on the: InstructionAPI and InstructionSemanticsAPI ParseAPI DataflowAPI Co-organized a series of high successful workshops with Prof. John Mellor-Crummey, Rice University, on Performance Tools for Petascale Computing, held in Snowbird, Utah and Lake Tahoe, California in July or August of 2007 through 2012. Investigated the use of multicore in numerical libraries Dyninst porting to 32- and 64bit Power/PowerPC (including BlueGene) and 32- and 64-bit Pentium platforms. Applying our toolkits to advanced problems in binary code parsing associated with dealing with legacy and malicious code.

  16. A Systems Approach to Scalable Transportation Network Modeling

    SciTech Connect

    Perumalla, Kalyan S

    2006-01-01

    Emerging needs in transportation network modeling and simulation are raising new challenges with respect to scal-ability of network size and vehicular traffic intensity, speed of simulation for simulation-based optimization, and fidel-ity of vehicular behavior for accurate capture of event phe-nomena. Parallel execution is warranted to sustain the re-quired detail, size and speed. However, few parallel simulators exist for such applications, partly due to the challenges underlying their development. Moreover, many simulators are based on time-stepped models, which can be computationally inefficient for the purposes of modeling evacuation traffic. Here an approach is presented to de-signing a simulator with memory and speed efficiency as the goals from the outset, and, specifically, scalability via parallel execution. The design makes use of discrete event modeling techniques as well as parallel simulation meth-ods. Our simulator, called SCATTER, is being developed, incorporating such design considerations. Preliminary per-formance results are presented on benchmark road net-works, showing scalability to one million vehicles simu-lated on one processor.

  17. Scalability, Timing, and System Design Issues for Intrinsic Evolvable Hardware

    NASA Technical Reports Server (NTRS)

    Hereford, James; Gwaltney, David

    2004-01-01

    In this paper we address several issues pertinent to intrinsic evolvable hardware (EHW). The first issue is scalability; namely, how the design space scales as the programming string for the programmable device gets longer. We develop a model for population size and the number of generations as a function of the programming string length, L, and show that the number of circuit evaluations is an O(L2) process. We compare our model to several successful intrinsic EHW experiments and discuss the many implications of our model. The second issue that we address is the timing of intrinsic EHW experiments. We show that the processing time is a small part of the overall time to derive or evolve a circuit and that major improvements in processor speed alone will have only a minimal impact on improving the scalability of intrinsic EHW. The third issue we consider is the system-level design of intrinsic EHW experiments. We review what other researchers have done to break the scalability barrier and contend that the type of reconfigurable platform and the evolutionary algorithm are tied together and impose limits on each other.

  18. Evolving scalable and modular adaptive networks with Developmental Symbolic Encoding.

    PubMed

    Suchorzewski, Marcin

    2011-09-01

    Evolutionary neural networks, or neuroevolution, appear to be a promising way to build versatile adaptive systems, combining evolution and learning. One of the most challenging problems of neuroevolution is finding a scalable and robust genetic representation, which would allow to effectively grow increasingly complex networks for increasingly complex tasks. In this paper we propose a novel developmental encoding for networks, featuring scalability, modularity, regularity and hierarchy. The encoding allows to represent structural regularities of networks and build them from encapsulated and possibly reused subnetworks. These capabilities are demonstrated on several test problems. In particular for parity and symmetry problems we evolve solutions, which are fully general with respect to the number of inputs. We also evolve scalable and modular weightless recurrent networks capable of autonomous learning in a simple generic classification task. The encoding is very flexible and we demonstrate this by evolving networks capable of learning via neuromodulation. Finally, we evolve modular solutions to the retina problem, for which another well known neuroevolution method-HyperNEAT-was previously shown to fail. The proposed encoding outperformed HyperNEAT and Cellular Encoding also in another experiment, in which certain connectivity patterns must be discovered between layers. Therefore we conclude the proposed encoding is an interesting and competitive approach to evolve networks. PMID:21957432

  19. The Scalable HeterOgeneous Computing (SHOC) Benchmark Suite

    SciTech Connect

    Danalis, Antonios; Marin, Gabriel; McCurdy, Collin B; Meredith, Jeremy S; Roth, Philip C; Spafford, Kyle L; Tipparaju, Vinod; Vetter, Jeffrey S

    2010-01-01

    Scalable heterogeneous computing systems, which are composed of a mix of compute devices, such as commodity multicore processors, graphics processors, reconfigurable processors, and others, are gaining attention as one approach to continuing performance improvement while managing the new challenge of energy efficiency. As these systems become more common, it is important to be able to compare and contrast architectural designs and programming systems in a fair and open forum. To this end, we have designed the Scalable HeterOgeneous Computing benchmark suite (SHOC). SHOC's initial focus is on systems containing graphics processing units (GPUs) and multi-core processors, and on the new OpenCL programming standard. SHOC is a spectrum of programs that test the performance and stability of these scalable heterogeneous computing systems. At the lowest level, SHOC uses microbenchmarks to assess architectural features of the system. At higher levels, SHOC uses application kernels to determine system-wide performance including many system features such as intranode and internode communication among devices. SHOC includes benchmark implementations in both OpenCL and CUDA in order to provide a comparison of these programming models.

  20. Design and Implementation of Ceph: A Scalable Distributed File System

    SciTech Connect

    Weil, S A; Brandt, S A; Miller, E L; Long, D E; Maltzahn, C

    2006-04-19

    File system designers continue to look to new architectures to improve scalability. Object-based storage diverges from server-based (e.g. NFS) and SAN-based storage systems by coupling processors and memory with disk drives, delegating low-level allocation to object storage devices (OSDs) and decoupling I/O (read/write) from metadata (file open/close) operations. Even recent object-based systems inherit decades-old architectural choices going back to early UNIX file systems, however, limiting their ability to effectively scale to hundreds of petabytes. We present Ceph, a distributed file system that provides excellent performance and reliability with unprecedented scalability. Ceph maximizes the separation between data and metadata management by replacing allocation tables with a pseudo-random data distribution function (CRUSH) designed for heterogeneous and dynamic clusters of unreliable OSDs. We leverage OSD intelligence to distribute data replication, failure detection and recovery with semi-autonomous OSDs running a specialized local object storage file system (EBOFS). Finally, Ceph is built around a dynamic distributed metadata management cluster that provides extremely efficient metadata management that seamlessly adapts to a wide range of general purpose and scientific computing file system workloads. We present performance measurements under a variety of workloads that show superior I/O performance and scalable metadata management (more than a quarter million metadata ops/sec).

  1. S2HAT: Scalable Spherical Harmonic Transform Library

    NASA Astrophysics Data System (ADS)

    Stompor, Radek

    2011-10-01

    Many problems in astronomy and astrophysics require a computation of the spherical harmonic transforms. This is in particular the case whenever data to be analyzed are distributed over the sphere or a set of corresponding mock data sets has to be generated. In many of those contexts, rapidly improving resolutions of both the data and simulations puts increasingly bigger emphasis on our ability to calculate the transforms quickly and reliably. The scalable spherical harmonic transform library S2HAT consists of a set of flexible, massively parallel, and scalable routines for calculating diverse (scalar, spin-weighted, etc) spherical harmonic transforms for a class of isolatitude sky grids or pixelizations. The library routines implement the standard algorithm with the complexity of O(n^3/2), where n is a number of pixels/grid points on the sphere, however, owing to their efficient parallelization and advanced numerical implementation, they achieve very competitive performance and near perfect scalability. S2HAT is written in Fortran 90 with a C interface. This software is a derivative of the spherical harmonic transforms included in the HEALPix package and is based on both serial and MPI routines of its version 2.01, however, since version 2.5 this software is fully autonomous of HEALPix and can be compiled and run without the HEALPix library.

  2. Building a Community Infrastructure for Scalable On-Line Performance Analysis Tools around Open|Speedshop

    SciTech Connect

    Miller, Barton

    2014-06-30

    Peta-scale computing environments pose significant challenges for both system and application developers and addressing them required more than simply scaling up existing tera-scale solutions. Performance analysis tools play an important role in gaining this understanding, but previous monolithic tools with fixed feature sets have not sufficed. Instead, this project worked on the design, implementation, and evaluation of a general, flexible tool infrastructure supporting the construction of performance tools as “pipelines” of high-quality tool building blocks. These tool building blocks provide common performance tool functionality, and are designed for scalability, lightweight data acquisition and analysis, and interoperability. For this project, we built on Open|SpeedShop, a modular and extensible open source performance analysis tool set. The design and implementation of such a general and reusable infrastructure targeted for petascale systems required us to address several challenging research issues. All components needed to be designed for scale, a task made more difficult by the need to provide general modules. The infrastructure needed to support online data aggregation to cope with the large amounts of performance and debugging data. We needed to be able to map any combination of tool components to each target architecture. And we needed to design interoperable tool APIs and workflows that were concrete enough to support the required functionality, yet provide the necessary flexibility to address a wide range of tools. A major result of this project is the ability to use this scalable infrastructure to quickly create tools that match with a machine architecture and a performance problem that needs to be understood. Another benefit is the ability for application engineers to use the highly scalable, interoperable version of Open|SpeedShop, which are reassembled from the tool building blocks into a flexible, multi-user interface set of tools. This set of

  3. Scalable Computational Chemistry: New Developments and Applications

    SciTech Connect

    Yuri Alexeev

    2002-12-31

    The computational part of the thesis is the investigation of titanium chloride (II) as a potential catalyst for the bis-silylation reaction of ethylene with hexaclorodisilane at different levels of theory. Bis-silylation is an important reaction for producing bis(silyl) compounds and new C-Si bonds, which can serve as monomers for silicon containing polymers and silicon carbides. Ab initio calculations on the steps involved in a proposed mechanism are presented. This choice of reactants allows them to study this reaction at reliable levels of theory without compromising accuracy. The calculations indicate that this is a highly exothermic barrierless reaction. The TiCl{sub 2} catalyst removes a 50 kcal/mol activation energy barrier required for the reaction without the catalyst. The first step is interaction of TiCl{sub 2} with ethylene to form an intermediate that is 60 kcal/mol below the energy of the reactants. This is the driving force for the entire reaction. Dynamic correlation plays a significant role because RHF calculations indicate that the net barrier for the catalyzed reaction is 50 kcal/mol. They conclude that divalent Ti has the potential to become an important industrial catalyst for silylation reactions. In the programming part of the thesis, parallelization of different quantum chemistry methods is presented. The parallelization of code is becoming important aspects of quantum chemistry code development. Two trends contribute to it: the overall desire to study large chemical systems and the desire to employ highly correlated methods which are usually computationally and memory expensive. In the presented distributed data algorithms computation is parallelized and the largest arrays are evenly distributed among CPUs. First, the parallelization of the Hartree-Fock self-consistent field (SCF) method is considered. SCF method is the most common starting point for more accurate calculations. The Fock build (sub step of SCF) from AO integrals is also

  4. Scalable Optical Architectures for Electronic Holography.

    NASA Astrophysics Data System (ADS)

    St.-Hilaire, Pierre

    galvanometric scanners as the horizontal scanning element, two 18-channel acousto-optic Bragg cells working in tandem, and a bank of custom-designed high-bandwidth framebuffers. The application of the concept of parallelism has allowed a six-fold scale-up of the display, which now produces high quality images 150mm times 75 mm in frontal dimensions, with a 30 degrees view zone. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge MA 02139-4307. Ph. 617-253-5668; Fax 617-253-1690.).

  5. Scalable Data Mining and Archiving for the Square Kilometre Array

    NASA Astrophysics Data System (ADS)

    Jones, D. L.; Mattmann, C. A.; Hart, A. F.; Lazio, J.; Bennett, T.; Wagstaff, K. L.; Thompson, D. R.; Preston, R.

    2011-12-01

    As the technologies for remote observation improve, the rapid increase in the frequency and fidelity of those observations translates into an avalanche of data that is already beginning to eclipse the resources, both human and technical, of the institutions and facilities charged with managing the information. Common data management tasks like cataloging both data itself and contextual meta-data, creating and maintaining scalable permanent archive, and making data available on-demand for research present significant software engineering challenges when considered at the scales of modern multi-national scientific enterprises such as the upcoming Square Kilometre Array project. The NASA Jet Propulsion Laboratory (JPL), leveraging internal research and technology development funding, has begun to explore ways to address the data archiving and distribution challenges with a number of parallel activities involving collaborations with the EVLA and ALMA teams at the National Radio Astronomy Observatory (NRAO), and members of the Square Kilometre Array South Africa team. To date, we have leveraged the Apache OODT Process Control System framework and its catalog and archive service components that provide file management, workflow management, resource management as core web services. A client crawler framework ingests upstream data (e.g., EVLA raw directory output), identifies its MIME type and automatically extracts relevant metadata including temporal bounds, and job-relevant/processing information. A remote content acquisition (pushpull) service is responsible for staging remote content and handing it off to the crawler framework. A science algorithm wrapper (called CAS-PGE) wraps underlying code including CASApy programs for the EVLA, such as Continuum Imaging and Spectral Line Cube generation, executes the algorithm, and ingests its output (along with relevant extracted metadata). In addition to processing, the Process Control System has been leveraged to provide data

  6. Efficient scalable algorithms for hierarchically semiseparable matrices

    SciTech Connect

    Wang, Shen; Xia, Jianlin; Situ, Yingchong; Hoop, Maarten V. de

    2011-09-14

    Hierarchically semiseparable (HSS) matrix algorithms are emerging techniques in constructing the superfast direct solvers for both dense and sparse linear systems. Here, we develope a set of novel parallel algorithms for the key HSS operations that are used for solving large linear systems. These include the parallel rank-revealing QR factorization, the HSS constructions with hierarchical compression, the ULV HSS factorization, and the HSS solutions. The HSS tree based parallelism is fully exploited at the coarse level. The BLACS and ScaLAPACK libraries are used to facilitate the parallel dense kernel operations at the ne-grained level. We have appplied our new parallel HSS-embedded multifrontal solver to the anisotropic Helmholtz equations for seismic imaging, and were able to solve a linear system with 6.4 billion unknowns using 4096 processors, in about 20 minutes. The classical multifrontal solver simply failed due to high demand of memory. To our knowledge, this is the first successful demonstration of employing the HSS algorithms in solving the truly large-scale real-world problems. Our parallel strategies can be easily adapted to the parallelization of the other rank structured methods.

  7. On the Scalability of Parallel UCT

    NASA Astrophysics Data System (ADS)

    Segal, Richard B.

    The parallelization of MCTS across multiple-machines has proven surprisingly difficult. The limitations of existing algorithms were evident in the 2009 Computer Olympiad where Zen using a single four-core machine defeated both Fuego with ten eight-core machines, and Mogo with twenty thirty-two core machines. This paper investigates the limits of parallel MCTS in order to understand why distributed parallelism has proven so difficult and to pave the way towards future distributed algorithms with better scaling. We first analyze the single-threaded scaling of Fuego and find that there is an upper bound on the play-quality improvements which can come from additional search. We then analyze the scaling of an idealized N-core shared memory machine to determine the maximum amount of parallelism supported by MCTS. We show that parallel speedup depends critically on how much time is given to each player. We use this relationship to predict parallel scaling for time scales beyond what can be empirically evaluated due to the immense computation required. Our results show that MCTS can scale nearly perfectly to at least 64 threads when combined with virtual loss, but without virtual loss scaling is limited to just eight threads. We also find that for competition time controls scaling to thousands of threads is impossible not necessarily due to MCTS not scaling, but because high levels of parallelism can start to bump up against the upper performance bound of Fuego itself.

  8. Algorithmic cooling and scalable NMR quantum computers

    PubMed Central

    Boykin, P. Oscar; Mor, Tal; Roychowdhury, Vwani; Vatan, Farrokh; Vrijen, Rutger

    2002-01-01

    We present here algorithmic cooling (via polarization heat bath)—a powerful method for obtaining a large number of highly polarized spins in liquid nuclear-spin systems at finite temperature. Given that spin-half states represent (quantum) bits, algorithmic cooling cleans dirty bits beyond the Shannon's bound on data compression, by using a set of rapidly thermal-relaxing bits. Such auxiliary bits could be implemented by using spins that rapidly get into thermal equilibrium with the environment, e.g., electron spins. Interestingly, the interaction with the environment, usually a most undesired interaction, is used here to our benefit, allowing a cooling mechanism. Cooling spins to a very low temperature without cooling the environment could lead to a breakthrough in NMR experiments, and our “spin-refrigerating” method suggests that this is possible. The scaling of NMR ensemble computers is currently one of the main obstacles to building larger-scale quantum computing devices, and our spin-refrigerating method suggests that this problem can be resolved. PMID:11904402

  9. Advanced technologies for scalable ATLAS conditions database access on the grid

    NASA Astrophysics Data System (ADS)

    Basset, R.; Canali, L.; Dimitrov, G.; Girone, M.; Hawkings, R.; Nevski, P.; Valassi, A.; Vaniachine, A.; Viegas, F.; Walker, R.; Wong, A.

    2010-04-01

    During massive data reprocessing operations an ATLAS Conditions Database application must support concurrent access from numerous ATLAS data processing jobs running on the Grid. By simulating realistic work-flow, ATLAS database scalability tests provided feedback for Conditions Db software optimization and allowed precise determination of required distributed database resources. In distributed data processing one must take into account the chaotic nature of Grid computing characterized by peak loads, which can be much higher than average access rates. To validate database performance at peak loads, we tested database scalability at very high concurrent jobs rates. This has been achieved through coordinated database stress tests performed in series of ATLAS reprocessing exercises at the Tier-1 sites. The goal of database stress tests is to detect scalability limits of the hardware deployed at the Tier-1 sites, so that the server overload conditions can be safely avoided in a production environment. Our analysis of server performance under stress tests indicates that Conditions Db data access is limited by the disk I/O throughput. An unacceptable side-effect of the disk I/O saturation is a degradation of the WLCG 3D Services that update Conditions Db data at all ten ATLAS Tier-1 sites using the technology of Oracle Streams. To avoid such bottlenecks we prototyped and tested a novel approach for database peak load avoidance in Grid computing. Our approach is based upon the proven idea of pilot job submission on the Grid: instead of the actual query, an ATLAS utility library sends to the database server a pilot query first.

  10. A low complexity prioritized bit-plane coding for SNR scalability in MPEG-21 scalable video coding

    NASA Astrophysics Data System (ADS)

    Peng, Wen-Hsiao; Chiang, Tihao; Hang, Hsueh-Ming

    2005-07-01

    In this paper, we propose a low complexity prioritized bit-plane coding scheme to improve the rate-distortion performance of cyclical block coding in MPEG-21 scalable video coding. Specifically, we use a block priority assignment algorithm to firstly transmit the symbols and the blocks with potentially better rate-distortion performance. Different blocks are allowed to be coded unequally in a coding cycle. To avoid transmitting priority overhead, the encoder and the decoder refer to the same context to assign priority. Furthermore, to reduce the complexity, the priority assignment is done by a look-up-table and the coding of each block is controlled by a simple threshold comparison mechanism. Experimental results show that our prioritized bit-plane coding scheme can offer up to 0.5dB PSNR improvement over the cyclical block coding described in the joint scalable verification model (JSVM).

  11. Scalable quantum computing based on stationary spin qubits in coupled quantum dots inside double-sided optical microcavities

    PubMed Central

    Wei, Hai-Rui; Deng, Fu-Guo

    2014-01-01

    Quantum logic gates are the key elements in quantum computing. Here we investigate the possibility of achieving a scalable and compact quantum computing based on stationary electron-spin qubits, by using the giant optical circular birefringence induced by quantum-dot spins in double-sided optical microcavities as a result of cavity quantum electrodynamics. We design the compact quantum circuits for implementing universal and deterministic quantum gates for electron-spin systems, including the two-qubit CNOT gate and the three-qubit Toffoli gate. They are compact and economic, and they do not require additional electron-spin qubits. Moreover, our devices have good scalability and are attractive as they both are based on solid-state quantum systems and the qubits are stationary. They are feasible with the current experimental technology, and both high fidelity and high efficiency can be achieved when the ratio of the side leakage to the cavity decay is low. PMID:25518899

  12. Fast fiber Bragg grating interrogation system with scalability to support monitoring of large structures in harsh environments

    NASA Astrophysics Data System (ADS)

    Moslehi, Behzad; Black, Richard J.; Costa, Joannes M.; Edwards, Elizabeth H.; Faridian, Fereydoun; Sotoudeh, Vahid

    2014-04-01

    Fiber optic sensor systems can alleviate certain challenges faced by electronics sensors faced when monitoring structures subject to marine and other harsh environments. Challenges in implementation of such systems include scalability, interconnection and cabling. We describe a fiber Bragg grating (FBG) sensor system architecture based that is scalable to support over 1000 electromagnetic interference immune sensors at high sampling rates for harsh environment applications. A key enabler is a high performance FBG interrogator supporting subsection sampling rates ranging from kHz to MHz. Results are presented for fast dynamic switching between multiple structural sections and the use of this sensing system for dynamic load monitoring as well as the potential for acoustic emission and ultrasonic monitoring on materials ranging from aluminum and composites to concrete subject to severe environments.

  13. Scalable quantum computing based on stationary spin qubits in coupled quantum dots inside double-sided optical microcavities.

    PubMed

    Wei, Hai-Rui; Deng, Fu-Guo

    2014-01-01

    Quantum logic gates are the key elements in quantum computing. Here we investigate the possibility of achieving a scalable and compact quantum computing based on stationary electron-spin qubits, by using the giant optical circular birefringence induced by quantum-dot spins in double-sided optical microcavities as a result of cavity quantum electrodynamics. We design the compact quantum circuits for implementing universal and deterministic quantum gates for electron-spin systems, including the two-qubit CNOT gate and the three-qubit Toffoli gate. They are compact and economic, and they do not require additional electron-spin qubits. Moreover, our devices have good scalability and are attractive as they both are based on solid-state quantum systems and the qubits are stationary. They are feasible with the current experimental technology, and both high fidelity and high efficiency can be achieved when the ratio of the side leakage to the cavity decay is low. PMID:25518899

  14. Scalable I/O Tracing and Analysis

    SciTech Connect

    Vijayakumar, Karthik; Mueller, Frank; Ma, Xiaosong; Roth, Philip C

    2009-01-01

    As supercomputer performance approached and then surpassed the petaflop level, I/O performance has become a major performance bottleneck for many scientific applications. Several tools exist to collect I/O traces to assist in the analysis of I/O performance problems. However, these tools either produce extremely large trace files that complicate performance analysis, or sacrifice accuracy to collect high-level statistical information. We propose a multi-level trace generator tool, ScalaIOTrace, that collects traces at several levels in the HPC I/O stack. ScalaIOTrace features aggressive trace compression that generates trace files of near constant size for regular I/O patterns and orders of magnitudes smaller for less regular ones. This enables the collection of I/O and communication traces of applications running on thousands of processors. Our contributions also include automated trace analysis to collect selected statistical information of I/O calls by parsing the compressed trace on-the-fly and time-accurate replay of communication events with MPI-IO calls. We evaluated our approach with the Parallel Ocean Program (POP) climate simulation and the FLASH parallel I/O benchmark. POP uses NetCDF as an I/O library while FLASH I/O uses the parallel HDF5 I/O library, which internally maps onto MPI-IO. We collected MPI-IO and low-level POSIX I/O traces to study application I/O behavior. Our results show constant size trace files of only 145KB irrespective of the number of nodes for FLASH I/O benchmark, which exhibits regular I/O and communication pattern. For POP, we observe up to two orders of magnitude reduction in trace file sizes compared to flat traces. Statistical information gathered reveals insight on the number of I/O and communication calls issued in the POP and FLASH I/O. Such concise traces are unprecedented for isolated I/O and combined I/O plus communication tracing.

  15. Efficiency and Scalability of Barrier Synchronization on NoC Based Many-core Architectures

    SciTech Connect

    Villa, Oreste; Palermo, Gianluca; Silvano, Cristina

    2008-10-18

    Interconnects based on Networks-on-Chip are an appealing solution to address future microprocessor designs where, very likely, hundreds of cores will be connected on a single chip. A fundamental role in highly parallelized applications running on many-core architectures will be played by barrier primitives used to synchronize the execution of parallel processes. This paper focuses on the analysis of the efficiency and scalability of different barrier implementations in many-core architectures based on NoCs. Several message passing barrier implementations based on four algorithms (all-to-all, master-slave, butterfly and tree) have been implemented and evaluated for a single-chip target architecture composed of a variable number of cores (from 4 to 128) and different network topologies (mesh, torus, ring, clustered-ring and fat-tree). Using a cycle-accurate simulator, we show the scalability of each barrier for every NoC topology, analyzing and comparing theoretical with real behaviors. We observed that some barrier algorithms, when implemented in hardware or software, show a different scaling behavior with respect to those theoretically expected. We evaluate the efficiency of each combination topology-barrier, demonstrating that, in many cases, simple network topologies can be more efficient than complex and highly connected topologies.

  16. Fabrication of scalable and structured tissue engineering scaffolds using water dissolvable sacrificial 3D printed moulds.

    PubMed

    Mohanty, Soumyaranjan; Larsen, Layla Bashir; Trifol, Jon; Szabo, Peter; Burri, Harsha Vardhan Reddy; Canali, Chiara; Dufva, Marin; Emnéus, Jenny; Wolff, Anders

    2015-10-01

    One of the major challenges in producing large scale engineered tissue is the lack of ability to create large highly perfused scaffolds in which cells can grow at a high cell density and viability. Here, we explore 3D printed polyvinyl alcohol (PVA) as a sacrificial mould in a polymer casting process. The PVA mould network defines the channels and is dissolved after curing the polymer casted around it. The printing parameters determined the PVA filament density in the sacrificial structure and this density resulted in different stiffness of the corresponding elastomer replica. It was possible to achieve 80% porosity corresponding to about 150 cm(2)/cm(3) surface to volume ratio. The process is easily scalable as demonstrated by fabricating a 75 cm(3) scaffold with about 16,000 interconnected channels (about 1m(2) surface area) and with a channel to channel distance of only 78 μm. To our knowledge this is the largest scaffold ever to be produced with such small feature sizes and with so many structured channels. The fabricated scaffolds were applied for in-vitro culturing of hepatocytes over a 12-day culture period. Smaller scaffolds (6×4 mm) were tested for cell culturing and could support homogeneous cell growth throughout the scaffold. Presumably, the diffusion of oxygen and nutrient throughout the channel network is rapid enough to support cell growth. In conclusion, the described process is scalable, compatible with cell culture, rapid, and inexpensive. PMID:26117791

  17. Robust, Scalable, and Fast Bootstrap Method for Analyzing Large Scale Data

    NASA Astrophysics Data System (ADS)

    Basiri, Shahab; Ollila, Esa; Koivunen, Visa

    2016-02-01

    In this paper we address the problem of performing statistical inference for large scale data sets i.e., Big Data. The volume and dimensionality of the data may be so high that it cannot be processed or stored in a single computing node. We propose a scalable, statistically robust and computationally efficient bootstrap method, compatible with distributed processing and storage systems. Bootstrap resamples are constructed with smaller number of distinct data points on multiple disjoint subsets of data, similarly to the bag of little bootstrap method (BLB) [1]. Then significant savings in computation is achieved by avoiding the re-computation of the estimator for each bootstrap sample. Instead, a computationally efficient fixed-point estimation equation is analytically solved via a smart approximation following the Fast and Robust Bootstrap method (FRB) [2]. Our proposed bootstrap method facilitates the use of highly robust statistical methods in analyzing large scale data sets. The favorable statistical properties of the method are established analytically. Numerical examples demonstrate scalability, low complexity and robust statistical performance of the method in analyzing large data sets.

  18. CAM-SE: A scalable spectral element dynamical core for the Community Atmosphere Model.

    SciTech Connect

    Dennis, John; Edwards, Jim; Evans, Kate J; Guba, O; Lauritzen, Peter; Mirin, Art; St.-Cyr, Amik; Taylor, Mark; Worley, Patrick H

    2012-01-01

    The Community Atmosphere Model (CAM) version 5 includes a spectral element dynamical core option from NCAR's High-Order Method Modeling Environment. It is a continuous Galerkin spectral finite element method designed for fully unstructured quadrilateral meshes. The current configurations in CAM are based on the cubed-sphere grid. The main motivation for including a spectral element dynamical core is to improve the scalability of CAM by allowing quasi-uniform grids for the sphere that do not require polar filters. In addition, the approach provides other state-of-the-art capabilities such as improved conservation properties. Spectral elements are used for the horizontal discretization, while most other aspects of the dynamical core are a hybrid of well tested techniques from CAM's finite volume and global spectral dynamical core options. Here we first give a overview of the spectral element dynamical core as used in CAM. We then give scalability and performance results from CAM running with three different dynamical core options within the Community Earth System Model, using a pre-industrial time-slice configuration. We focus on high resolution simulations of 1/4 degree, 1/8 degree, and T340 spectral truncation.

  19. Magnetically anisotropic additive for scalable manufacturing of polymer nanocomposite: iron-coated carbon nanotubes

    NASA Astrophysics Data System (ADS)

    Yamamoto, Namiko; Manohara, Harish; Platzman, Ellen

    2016-02-01

    Novel nanoparticles additives for polymer nanocomposites were prepared by coating carbon nanotubes (CNTs) with ferromagnetic iron (Fe) layers, so that their micro-structures can be bulk-controlled by external magnetic field application. Application of magnetic fields is a promising, scalable method to deliver bulk amount of nanocomposites while maintaining organized nanoparticle assembly throughout the uncured polymer matrix. In this work, Fe layers (˜18 nm thick) were deposited on CNTs (˜38 nm diameter and ˜50 μm length) to form thin films with high aspect ratio, resulting in a dominance of shape anisotropy and thus high coercivity of ˜50-100 Oe. The Fe-coated CNTs were suspended in water and applied with a weak magnetic field of ˜75 G, and yet preliminary magnetic assembly was confirmed. Our results demonstrate that the fabricated Fe-coated CNTs are magnetically anisotropic and effectively respond to magnetic fields that are ˜103 times smaller than other existing work (˜105 G). We anticipate this work will pave the way for effective property enhancement and bulk application of CNT-polymer nanocomposites, through controlled micro-structure and scalable manufacturing.

  20. Scalable Production of Si Nanoparticles Directly from Low Grade Sources for Lithium-Ion Battery Anode.

    PubMed

    Zhu, Bin; Jin, Yan; Tan, Yingling; Zong, Linqi; Hu, Yue; Chen, Lei; Chen, Yanbin; Zhang, Qiao; Zhu, Jia

    2015-09-01

    Silicon, one of the most promising candidates as lithium-ion battery anode, has attracted much attention due to its high theoretical capacity, abundant existence, and mature infrastructure. Recently, Si nanostructures-based lithium-ion battery anode, with sophisticated structure designs and process development, has made significant progress. However, low cost and scalable processes to produce these Si nanostructures remained as a challenge, which limits the widespread applications. Herein, we demonstrate that Si nanoparticles with controlled size can be massively produced directly from low grade Si sources through a scalable high energy mechanical milling process. In addition, we systematically studied Si nanoparticles produced from two major low grade Si sources, metallurgical silicon (∼99 wt % Si, $1/kg) and ferrosilicon (∼83 wt % Si, $0.6/kg). It is found that nanoparticles produced from ferrosilicon sources contain FeSi2, which can serve as a buffer layer to alleviate the mechanical fractures of volume expansion, whereas nanoparticles from metallurgical Si sources have higher capacity and better kinetic properties because of higher purity and better electronic transport properties. Ferrosilicon nanoparticles and metallurgical Si nanoparticles demonstrate over 100 stable deep cycling after carbon coating with the reversible capacities of 1360 mAh g(-1) and 1205 mAh g(-1), respectively. Therefore, our approach provides a new strategy for cost-effective, energy-efficient, large scale synthesis of functional Si electrode materials. PMID:26258439

  1. Scalable Fabrication of Integrated Nanophotonic Circuits on Arrays of Thin Single Crystal Diamond Membrane Windows.

    PubMed

    Piracha, Afaq H; Rath, Patrik; Ganesan, Kumaravelu; Kühn, Stefan; Pernice, Wolfram H P; Prawer, Steven

    2016-05-11

    Diamond has emerged as a promising platform for nanophotonic, optical, and quantum technologies. High-quality, single crystalline substrates of acceptable size are a prerequisite to meet the demanding requirements on low-level impurities and low absorption loss when targeting large photonic circuits. Here, we describe a scalable fabrication method for single crystal diamond membrane windows that achieves three major goals with one fabrication method: providing high quality diamond, as confirmed by Raman spectroscopy; achieving homogeneously thin membranes, enabled by ion implantation; and providing compatibility with established planar fabrication via lithography and vertical etching. On such suspended diamond membranes we demonstrate a suite of photonic components as building blocks for nanophotonic circuits. Monolithic grating couplers are used to efficiently couple light between photonic circuits and optical fibers. In waveguide coupled optical ring resonators, we find loaded quality factors up to 66 000 at a wavelength of 1560 nm, corresponding to propagation loss below 7.2 dB/cm. Our approach holds promise for the scalable implementation of future diamond quantum photonic technologies and all-diamond photonic metrology tools. PMID:27111636

  2. Scalable shape-controlled fabrication of curved microstructures using a femtosecond laser wet-etching process.

    PubMed

    Bian, Hao; Yang, Qing; Chen, Feng; Liu, Hewei; Du, Guangqing; Deng, Zefang; Si, Jinhai; Yun, Feng; Hou, Xun

    2013-07-01

    Materials with curvilinear surface microstructures are highly desirable for micro-optical and biomedical devices. However, realization of such devices efficiently remains technically challenging. This paper demonstrates a facile and flexible method to fabricate curvilinear microstructures with controllable shapes and dimensions. The method composes of femtosecond laser exposures and chemical etching process with the hydrofluoric acid solutions. By fixed-point and step-in laser irradiations followed by the chemical treatments, concave microstructures with different profiles such as spherical, conical, bell-like and parabola were fabricated on silica glasses. The convex structures were replicated on polymers by the casting replication process. In this work, we used this technique to fabricate high-quality microlens arrays and high-aspect-ratio microwells which can be used in 3D cell culture. This approach offers several advantages such as high-efficient, scalable shape-controllable and easy manipulations. PMID:23623098

  3. Efficient cross polarized wave generation for compact, energy-scalable, ultrashort laser sources.

    PubMed

    Ramirez, L P; Papadopoulos, D N; Pellegrina, A; Georges, P; Druon, F; Monot, P; Ricci, A; Jullien, A; Chen, X; Rousseau, J P; Lopez-Martens, R

    2011-01-01

    The generation of high contrast and ultrashort laser pulses via a compact and energy-scalable cross polarized wave filter is presented. The setup incorporates a waveguide spatial filter into a single crystal XPW configuration, enabling high energy and high intensity transmission, efficient contrast enhancement and pulse shortening at the multi-mJ level. Excellent XPW conversion of up to 33% (global efficiency: 20%, intensity transmission: 40%) led to an output energy of 650 µJ for an input of 3.3 mJ. Additionally, efficient conversion under specific input phase conditions, allowed pulse shortening from 25 fs to 9.6 fs, indicating the prospective application of this setup as a high energy, ultrabroad laser source. PMID:21263545

  4. On the scalability of the Albany/FELIX first-order Stokes approximation ice sheet solver for large-scale simulations of the Greenland and Antarctic ice sheets

    DOE PAGESBeta

    Tezaur, Irina K.; Tuminaro, Raymond S.; Perego, Mauro; Salinger, Andrew G.; Price, Stephen F.

    2015-01-01

    We examine the scalability of the recently developed Albany/FELIX finite-element based code for the first-order Stokes momentum balance equations for ice flow. We focus our analysis on the performance of two possible preconditioners for the iterative solution of the sparse linear systems that arise from the discretization of the governing equations: (1) a preconditioner based on the incomplete LU (ILU) factorization, and (2) a recently-developed algebraic multigrid (AMG) preconditioner, constructed using the idea of semi-coarsening. A strong scalability study on a realistic, high resolution Greenland ice sheet problem reveals that, for a given number of processor cores, the AMG preconditionermore » results in faster linear solve times but the ILU preconditioner exhibits better scalability. A weak scalability study is performed on a realistic, moderate resolution Antarctic ice sheet problem, a substantial fraction of which contains floating ice shelves, making it fundamentally different from the Greenland ice sheet problem. Here, we show that as the problem size increases, the performance of the ILU preconditioner deteriorates whereas the AMG preconditioner maintains scalability. This is because the linear systems are extremely ill-conditioned in the presence of floating ice shelves, and the ill-conditioning has a greater negative effect on the ILU preconditioner than on the AMG preconditioner.« less

  5. Generation of scalable terahertz radiation from cylindrically focused laser pulses in air

    NASA Astrophysics Data System (ADS)

    Kuk, Donghoon; Yoo, Yungjun; Rosenthal, Eric; Jhajj, Nihal; Milchberg, Howard; Kim, Ki-Yong

    We have demonstrated scalable terahertz (THz) generation via cylindrical focusing of two-color laser pulses in air. In this experiment, we have used a terawatt (TW) laser system which can deliver >50 mJ, 800 nm, 50 fs pulses at a 10 Hz repetition rate. A 800 nm pulse passing through a nonlinear crystal (BBO) generates its second harmonic pulse (400 nm). Both pulses pass through a cylindrical lens and are focused together to generate a 2-dimensional plasma sheet in air. This yields two diverging THz lobes, characterized by an uncooled microbolometer. This observed radiation angle and pattern is explained by the optical-Cherenkov radiation theory. The diverging THz radiation is re-focused to yield strong THz field strengths (>20 MV/cm) at the focus. At laser energy of 40 mJ, cylindrical focusing provides THz energy of >30 microjoules, far exceeding the output produced by spherical focusing. This shows that cylindrical focusing can effectively minimize ionization-induced defocusing, previously observed in spherical focusing, and can allow scalable THz generation with relatively high laser energies (>20 mJ). Work supported by DOE, Office of Basic Energy Sciences, Division of Materials Sciences and Engineering under Award No. 014216-001.

  6. A rapid and scalable system for studying gene function in mice using conditional RNA interference

    PubMed Central

    Premsrirut, Prem K.; Dow, Lukas E.; Kim, Sang Yong; Camiolo, Matthew; Malone, Colin D.; Miething, Cornelius; Scuoppo, Claudio; Zuber, Johannes; Dickins, Ross A.; Kogan, Scott C.; Shroyer, Kenneth R.; Sordella, Raffaella; Hannon, Gregory J.; Lowe, Scott W.

    2011-01-01

    Summary RNA interference is a powerful tool for studying gene function, however, the reproducible generation of RNAi transgenic mice remains a significant limitation. By combining optimized fluorescence-coupled miR30-based shRNAs with high efficiency ES cell targeting, we developed a fast, scalable pipeline for the production of shRNA transgenic mice. Using this system, we generated eight tet-regulated shRNA transgenic lines targeting Firefly and Renilla luciferases, Oct4 and tumor suppressors p53, p16INK4a, p19ARF and APC and demonstrate potent gene silencing and GFP-tracked knockdown in a broad range of tissues in vivo. Further, using an shRNA targeting APC, we illustrate how this approach can identify predicted phenotypes and also unknown functions for a well-studied gene. In addition, through regulated gene silencing we validate APC/Wnt and p19ARF as potential therapeutic targets in T cell acute lymphoblastic leukemia/lymphoma and lung adenocarcinoma, respectively. This system provides a cost-effective and scalable platform for the production of RNAi transgenic mice targeting any mammalian gene. PMID:21458673

  7. Tier-scalable reconnaissance: the challenge of sensor optimization, sensor deployment, sensor fusion, and sensor interoperability

    NASA Astrophysics Data System (ADS)

    Fink, Wolfgang; George, Thomas; Tarbell, Mark A.

    2007-04-01

    Robotic reconnaissance operations are called for in extreme environments, not only those such as space, including planetary atmospheres, surfaces, and subsurfaces, but also in potentially hazardous or inaccessible operational areas on Earth, such as mine fields, battlefield environments, enemy occupied territories, terrorist infiltrated environments, or areas that have been exposed to biochemical agents or radiation. Real time reconnaissance enables the identification and characterization of transient events. A fundamentally new mission concept for tier-scalable reconnaissance of operational areas, originated by Fink et al., is aimed at replacing the engineering and safety constrained mission designs of the past. The tier-scalable paradigm integrates multi-tier (orbit atmosphere surface/subsurface) and multi-agent (satellite UAV/blimp surface/subsurface sensing platforms) hierarchical mission architectures, introducing not only mission redundancy and safety, but also enabling and optimizing intelligent, less constrained, and distributed reconnaissance in real time. Given the mass, size, and power constraints faced by such a multi-platform approach, this is an ideal application scenario for a diverse set of MEMS sensors. To support such mission architectures, a high degree of operational autonomy is required. Essential elements of such operational autonomy are: (1) automatic mapping of an operational area from different vantage points (including vehicle health monitoring); (2) automatic feature extraction and target/region-of-interest identification within the mapped operational area; and (3) automatic target prioritization for close-up examination. These requirements imply the optimal deployment of MEMS sensors and sensor platforms, sensor fusion, and sensor interoperability.

  8. FMOE-MR: content-driven multiresolution MPEG-4 fine grained scalable layered video encoding

    NASA Astrophysics Data System (ADS)

    Chattopadhyay, S.; Luo, X.; Bhandarkar, S. M.; Li, K.

    2007-01-01

    The MPEG-4 Fine Grained Scalability (FGS) profile aims at scalable layered video encoding, in order to ensure efficient video streaming in networks with fluctuating bandwidths. In this paper, we propose a novel technique, termed as FMOEMR, which delivers significantly improved rate distortion performance compared to existing MPEG-4 Base Layer encoding techniques. The video frames are re-encoded at high resolution at semantically and visually important regions of the video (termed as Features, Motion and Objects) that are defined using a mask (FMO-Mask) and at low resolution in the remaining regions. The multiple-resolution re-rendering step is implemented such that further MPEG-4 compression leads to low bit rate Base Layer video encoding. The Features, Motion and Objects Encoded-Multi- Resolution (FMOE-MR) scheme is an integrated approach that requires only encoder-side modifications, and is transparent to the decoder. Further, since the FMOE-MR scheme incorporates "smart" video preprocessing, it requires no change in existing MPEG-4 codecs. As a result, it is straightforward to use the proposed FMOE-MR scheme with any existing MPEG codec, thus allowing great flexibility in implementation. In this paper, we have described, and implemented, unsupervised and semi-supervised algorithms to create the FMO-Mask from a given video sequence, using state-of-the-art computer vision algorithms.

  9. NEXUS Scalable and Distributed Next-Generation Avionics Bus for Space Missions

    NASA Technical Reports Server (NTRS)

    He, Yutao; Shalom, Eddy; Chau, Savio N.; Some, Raphael R.; Bolotin, Gary S.

    2011-01-01

    A paper discusses NEXUS, a common, next-generation avionics interconnect that is transparently compatible with wired, fiber-optic, and RF physical layers; provides a flexible, scalable, packet switched topology; is fault-tolerant with sub-microsecond detection/recovery latency; has scalable bandwidth from 1 Kbps to 10 Gbps; has guaranteed real-time determinism with sub-microsecond latency/jitter; has built-in testability; features low power consumption (< 100 mW per Gbps); is lightweight with about a 5,000-logic-gate footprint; and is implemented in a small Bus Interface Unit (BIU) with reconfigurable back-end providing interface to legacy subsystems. NEXUS enhances a commercial interconnect standard, Serial RapidIO, to meet avionics interconnect requirements without breaking the standard. This unified interconnect technology can be used to meet performance, power, size, and reliability requirements of all ranges of equipment, sensors, and actuators at chip-to-chip, board-to-board, or box-to-box boundary. Early results from in-house modeling activity of Serial RapidIO using VisualSim indicate that the use of a switched, high-performance avionics network will provide a quantum leap in spacecraft onboard science and autonomy capability for science and exploration missions.

  10. A scalable and adaptable solution framework within components of the CCSM

    SciTech Connect

    Evans, Katherine J; Rouson, Damian; Salinger, Andy; Taylor, Mark; White III, James B; Weijer, Wilbert

    2009-01-01

    A framework for a fully implicit solution method is implemented into (1) the High Order Methods Modeling Environment (HOMME), which is a spectral element dynamical core option in the Community Atmosphere Model (CAM), and (2) the Parallel Ocean Program (POP) model of the global ocean. Both of these models are components of the Community Climate System Model (CCSM). HOMME is a development version of CAM and provides a scalable alternative when run with an explicit time integrator. However, it suffers the typical time step size limit to maintain stability. POP uses a time-split semi-implicit time integrator that allows larger time steps but less accuracy when used with scale interacting physics. A fully implicit solution framework allows larger time step sizes and additional climate analysis capability such as model steady state and spin-up efficiency gains without a loss in scalability. This framework is implemented into HOMME and POP using a new Fortran interface to the Trilinos solver library, ForTrilinos, which leverages several new capabilities in the current Fortran standard to maximize robustness and speed. The ForTrilinos solution template was also designed for interchangeability; other solution methods and capability improvements can be more easily implemented into the models as they are developed without severely interacting with the code structure. The utility of this approach is illustrated with a test case for each of the climate component models.

  11. Hydra: a scalable proteomic search engine which utilizes the Hadoop distributed computing framework

    PubMed Central

    2012-01-01

    Background For shotgun mass spectrometry based proteomics the most computationally expensive step is in matching the spectra against an increasingly large database of sequences and their post-translational modifications with known masses. Each mass spectrometer can generate data at an astonishingly high rate, and the scope of what is searched for is continually increasing. Therefore solutions for improving our ability to perform these searches are needed. Results We present a sequence database search engine that is specifically designed to run efficiently on the Hadoop MapReduce distributed computing framework. The search engine implements the K-score algorithm, generating comparable output for the same input files as the original implementation. The scalability of the system is shown, and the architecture required for the development of such distributed processing is discussed. Conclusion The software is scalable in its ability to handle a large peptide database, numerous modifications and large numbers of spectra. Performance scales with the number of processors in the cluster, allowing throughput to expand with the available resources. PMID:23216909

  12. Identifying unproven cancer treatments on the health web: addressing accuracy, generalizability and scalability.

    PubMed

    Aphinyanaphongs, Yin; Fu, Lawrence D; Aliferis, Constantin F

    2013-01-01

    Building machine learning models that identify unproven cancer treatments on the Health Web is a promising approach for dealing with the dissemination of false and dangerous information to vulnerable health consumers. Aside from the obvious requirement of accuracy, two issues are of practical importance in deploying these models in real world applications. (a) Generalizability: The models must generalize to all treatments (not just the ones used in the training of the models). (b) Scalability: The models can be applied efficiently to billions of documents on the Health Web. First, we provide methods and related empirical data demonstrating strong accuracy and generalizability. Second, by combining the MapReduce distributed architecture and high dimensionality compression via Markov Boundary feature selection, we show how to scale the application of the models to WWW-scale corpora. The present work provides evidence that (a) a very small subset of unproven cancer treatments is sufficient to build a model to identify unproven treatments on the web; (b) unproven treatments use distinct language to market their claims and this language is learnable; (c) through distributed parallelization and state of the art feature selection, it is possible to prepare the corpora and build and apply models with large scalability. PMID:23920640

  13. Scalable Method to Produce Biodegradable Nanoparticles that Rapidly Penetrate Human Mucus

    PubMed Central

    Xu, Qingguo; Boylan, Nicholas J.; Cai, Shutian; Miao, Bolong; Patel, Himatkumar; Hanes, Justin

    2013-01-01

    Mucus typically traps and rapidly removes foreign particles from the airways, gastrointestinal tract, nasopharynx, female reproductive tract and the surface of the eye. Nanoparticles capable of rapid penetration through mucus can potentially avoid rapid clearance, and open significant opportunities for controlled drug delivery at mucosal surfaces. Here, we report an industrially scalable emulsification method to produce biodegradable mucus-penetrating particles (MPP). The emulsification of diblock copolymers of poly(lactic-co-glycolic acid) and polyethylene glycol (PLGA-PEG) using low molecular weight (MW) emulsifiers forms dense brush PEG coatings on nanoparticles that allow rapid nanoparticle penetration through fresh undiluted human mucus. In comparison, conventional high MW emulsifiers, such as polyvinyl alcohol (PVA), interrupts the PEG coating on nanoparticles, resulting in their immobilization in mucus owing to adhesive interactions with mucus mesh elements. PLGA-PEG nanoparticles with a wide range of PEG MW (1, 2, 5, and 10 kDa), prepared by the emulsification method using low MW emulsifiers, all rapidly penetrated mucus. A range of drugs, from hydrophobic small molecules to hydrohilic large biologics, can be efficiently loaded into biodegradable MPP using the method described. This readily scalable method should facilitate the production of MPP products for mucosal drug delivery, as well as potentially longer-circulating particles following intravenous administration. PMID:23751567

  14. Identifying Unproven Cancer Treatments on the Health Web: Addressing Accuracy, Generalizability and Scalability

    PubMed Central

    Aphinyanaphongs, Yin; Fu, Lawrence D.; Aliferis, Constantin F.

    2014-01-01

    Building machine learning models that identify unproven cancer treatments on the Health Web is a promising approach for dealing with the dissemination of false and dangerous information to vulnerable health consumers. Aside from the obvious requirement of accuracy, two issues are of practical importance in deploying these models in real world applications. (a) Generalizability: The models must generalize to all treatments (not just the ones used in the training of the models). (b) Scalability: The models can be applied efficiently to billions of documents on the Health Web. First, we provide methods and related empirical data demonstrating strong accuracy and generalizability. Second, by combining the MapReduce distributed architecture and high dimensionality compression via Markov Boundary feature selection, we show how to scale the application of the models to WWW-scale corpora. The present work provides evidence that (a) a very small subset of unproven cancer treatments is sufficient to build a model to identify unproven treatments on the web; (b) unproven treatments use distinct language to market their claims and this language is learnable; (c) through distributed parallelization and state of the art feature selection, it is possible to prepare the corpora and build and apply models with large scalability. PMID:23920640

  15. A scalable graphene sulfur composite synthesis for rechargeable lithium batteries with good capacity and excellent columbic efficiency.

    PubMed

    Gao, Xianfeng; Li, Jianyang; Guan, Dongsheng; Yuan, Chris

    2014-03-26

    Sulfur nanoparticles wrapped with a conductive graphene framework was synthesized with a high sulfur loading through a scalable one-step process. The graphene-coated sulfur nanostructured composite, when used as cathode for lithium sulfur battery, shows a reversible capacity of 808 mAh g(-1) at a rate of 210 mA g(-1) and an average columbic efficiency of ∼98.3% over 100 cycles. It is found that graphene oxide (GO) with a porous structure offers flexible confinement function that helps prevent the loss of active materials, thus extending the cycling life of the electrode. Moreover, reduced graphene oxide provides a conductive network surrounding the sulfur particles, which facilitates both electron transport and ion transportation. This novel one-step, all-solution-based process is scalable and provides a promising approach for potential industrial applications. PMID:24555988

  16. Tunable electrophoretic separations using a scalable, fabric-based platform.

    PubMed

    Narahari, Tanya; Dendukuri, Dhananjaya; Murthy, Shashi K

    2015-02-17

    There is a rising need for low-cost and scalable platforms for sensitive medical diagnostic testing. Fabric weaving is a mature, scalable manufacturing technology and can be used as a platform to manufacture microfluidic diagnostic tests with controlled, tunable flow. Given its scalability, low manufacturing cost (<$0.25 per device), and potential for patterning multiplexed channel geometries, fabric is a viable platform for the development of analytical devices. In this paper, we describe a fabric-based electrophoretic platform for protein separation. Appropriate yarns were selected for each region of the device and weaved into straight channel electrophoretic chips in a single step. A wide dynamic range of analyte molecules ranging from small molecule dyes (<1 kDa) to macromolecule proteins (67-150 kDa) were separated in the device. Individual yarns behave as a chromatographic medium for electrophoresis. We therefore explored the effect of yarn and fabric parameters on separation resolution. Separation speed and resolution were enhanced by increasing the number of yarns per unit area of fabric and decreasing yarn hydrophilicity. However, for protein analytes that often require hydrophilic, passivated surfaces, these effects need to be properly tuned to achieve well-resolved separations. A fabric device tuned for protein separations was built and demonstrated. As an analytical output parameter for this device, the electrophoretic mobility of a sedimentation marker, Naphthol Blue Black bovine albumin in glycine-NaOH buffer, pH 8.58 was estimated and found to be -2.7 × 10(-8) m(2) V(-1) s(-1). The ability to tune separation may be used to predefine regions in the fabric for successive preconcentrations and separations. The device may then be applied for the multiplexed detection of low abundance proteins from complex biological samples such as serum and cell lysate. PMID:25582166

  17. A domain decomposition method of stochastic PDEs: An iterative solution techniques using a two-level scalable preconditioner

    SciTech Connect

    Subber, Waad Sarkar, Abhijit

    2014-01-15

    Recent advances in high performance computing systems and sensing technologies motivate computational simulations with extremely high resolution models with capabilities to quantify uncertainties for credible numerical predictions. A two-level domain decomposition method is reported in this investigation to devise a linear solver for the large-scale system in the Galerkin spectral stochastic finite element method (SSFEM). In particular, a two-level scalable preconditioner is introduced in order to iteratively solve the large-scale linear system in the intrusive SSFEM using an iterative substructuring based domain decomposition solver. The implementation of the algorithm involves solving a local problem on each subdomain that constructs the local part of the preconditioner and a coarse problem that propagates information globally among the subdomains. The numerical and parallel scalabilities of the two-level preconditioner are contrasted with the previously developed one-level preconditioner for two-dimensional flow through porous media and elasticity problems with spatially varying non-Gaussian material properties. A distributed implementation of the parallel algorithm is carried out using MPI and PETSc parallel libraries. The scalabilities of the algorithm are investigated in a Linux cluster.

  18. Flexible quantum circuits using scalable continuous-variable cluster states

    NASA Astrophysics Data System (ADS)

    Alexander, Rafael N.; Menicucci, Nicolas C.

    2016-06-01

    We show that measurement-based quantum computation on scalable continuous-variable (CV) cluster states admits more quantum-circuit flexibility and compactness than similar protocols for standard square-lattice CV cluster states. This advantage is a direct result of the macronode structure of these states—that is, a lattice structure in which each graph node actually consists of several physical modes. These extra modes provide additional measurement degrees of freedom at each graph location, which can be used to manipulate the flow and processing of quantum information more robustly and with additional flexibility that is not available on an ordinary lattice.

  19. Scalability and Performance of a Large Linux Cluster

    SciTech Connect

    BRIGHTWELL,RONALD B.; PLIMPTON,STEVEN J.

    2000-01-20

    In this paper the authors present performance results from several parallel benchmarks and applications on a 400-node Linux cluster at Sandia National Laboratories. They compare the results on the Linux cluster to performance obtained on a traditional distributed-memory massively parallel processing machine, the Intel TeraFLOPS. They discuss the characteristics of these machines that influence the performance results and identify the key components of the system software that they feel are important to allow for scalability of commodity-based PC clusters to hundreds and possibly thousands of processors.

  20. Scalable web services for the PSIPRED Protein Analysis Workbench.

    PubMed

    Buchan, Daniel W A; Minneci, Federico; Nugent, Tim C O; Bryson, Kevin; Jones, David T

    2013-07-01

    Here, we present the new UCL Bioinformatics Group's PSIPRED Protein Analysis Workbench. The Workbench unites all of our previously available analysis methods into a single web-based framework. The new web portal provides a greatly streamlined user interface with a number of new features to allow users to better explore their results. We offer a number of additional services to enable computationally scalable execution of our prediction methods; these include SOAP and XML-RPC web server access and new HADOOP packages. All software and services are available via the UCL Bioinformatics Group website at http://bioinf.cs.ucl.ac.uk/. PMID:23748958