On the Feasibility of Prefetching and Caching for Online TV Services: A Measurement Study on Hulu
NASA Astrophysics Data System (ADS)
Krishnappa, Dilip Kumar; Khemmarat, Samamon; Gao, Lixin; Zink, Michael
Lately researchers are looking at ways to reduce the delay on video playback through mechanisms like prefetching and caching for Video-on-Demand (VoD) services. The usage of prefetching and caching also has the potential to reduce the amount of network bandwidth usage, as most popular requests are served from a local cache rather than the server containing the original content. In this paper, we investigate the advantages of having such a prefetching and caching scheme for a free hosting service of professionally created video (movies and TV shows) named "hulu". We look into the advantages of using a prefetching scheme where the most popular videos of the week, as provided by the hulu website, are prefetched and compare this approach with a conventional LRU caching scheme with limited storage space and a combined scheme of prefetching and caching. Results from our measurement and analysis shows that employing a basic caching scheme at the proxy yields a hit ratio of up to 77.69%, but requires storage of about 236GB. Further analysis shows that a prefetching scheme where the top-100 popular videos of the week are downloaded to the proxy yields a hit ratio of 44% with a storage requirement of 10GB. A LRU caching scheme with a storage limitation of 20GB can achieve a hit ratio of 55% but downloads 4713 videos to achieve such high hit ratio compared to 100 videos in prefetching scheme, whereas a scheme with both prefetching and caching with the same storage yields a hit ratio of 59% with download requirement of 4439 videos. We find that employing a scheme of prefetching along with caching with trade-off on the storage will yield a better hit ratio and bandwidth saving than individual caching or prefetching schemes.
Evaluating the effect of online data compression on the disk cache of a mass storage system
NASA Technical Reports Server (NTRS)
Pentakalos, Odysseas I.; Yesha, Yelena
1994-01-01
A trace driven simulation of the disk cache of a mass storage system was used to evaluate the effect of an online compression algorithm on various performance measures. Traces from the system at NASA's Center for Computational Sciences were used to run the simulation and disk cache hit ratios, number of files and bytes migrating to tertiary storage were measured. The measurements were performed for both an LRU and a size based migration algorithm. In addition to seeing the effect of online data compression on the disk cache performance measure, the simulation provided insight into the characteristics of the interactive references, suggesting that hint based prefetching algorithms are the only alternative for any future improvements to the disk cache hit ratio.
Distributed Name Servers: Naming and Caching in Large Distributed Computing Environments
1985-12-01
transmission rate of the communication medium1, transmission over a 56K bps line costs approx- imately 54r, and similarly, communication over a 9.6K...memories for modem computer systems attempt to maximize the hit ratio for a fixed-size cache by utilizing intelligent cache replacement algorithms
Communications Patterns in a Symbolic Multiprocessor.
1987-06-01
instruction references that Multilisp programs make. The cache hit ratio is greatest when instruction references have a high degree of -- locality. Another...future touches hit an undetermined future. N, The only exception is Consim, in which one third of future touches hit unde- termined futures. Task...Cambridge, MA, June 1985. [52] S. Sugimoto, K. Agusa, K. Tabata , and Y. Ohno. A multi-microprocessor system for concurrent Lisp. In Proceedings of
A trace-driven analysis of name and attribute caching in a distributed system
NASA Technical Reports Server (NTRS)
Shirriff, Ken W.; Ousterhout, John K.
1992-01-01
This paper presents the results of simulating file name and attribute caching on client machines in a distributed file system. The simulation used trace data gathered on a network of about 40 workstations. Caching was found to be advantageous: a cache on each client containing just 10 directories had a 91 percent hit rate on name look ups. Entry-based name caches (holding individual directory entries) had poorer performance for several reasons, resulting in a maximum hit rate of about 83 percent. File attribute caching obtained a 90 percent hit rate with a cache on each machine of the attributes for 30 files. The simulations show that maintaining cache consistency between machines is not a significant problem; only 1 in 400 name component look ups required invalidation of a remotely cached entry. Process migration to remote machines had little effect on caching. Caching was less successful in heavily shared and modified directories such as /tmp, but there weren't enough references to /tmp overall to affect the results significantly. We estimate that adding name and attribute caching to the Sprite operating system could reduce server load by 36 percent and the number of network packets by 30 percent.
Prefetching in file systems for MIMD multiprocessors
NASA Technical Reports Server (NTRS)
Kotz, David F.; Ellis, Carla Schlatter
1990-01-01
The question of whether prefetching blocks on the file into the block cache can effectively reduce overall execution time of a parallel computation, even under favorable assumptions, is considered. Experiments have been conducted with an interleaved file system testbed on the Butterfly Plus multiprocessor. Results of these experiments suggest that (1) the hit ratio, the accepted measure in traditional caching studies, may not be an adequate measure of performance when the workload consists of parallel computations and parallel file access patterns, (2) caching with prefetching can significantly improve the hit ratio and the average time to perform an I/O (input/output) operation, and (3) an improvement in overall execution time has been observed in most cases. In spite of these gains, prefetching sometimes results in increased execution times (a negative result, given the optimistic nature of the study). The authors explore why it is not trivial to translate savings on individual I/O requests into consistently better overall performance and identify the key problems that need to be addressed in order to improve the potential of prefetching techniques in the environment.
An effective write policy for software coherence schemes
NASA Technical Reports Server (NTRS)
Chen, Yung-Chin; Veidenbaum, Alexander V.
1992-01-01
The authors study the write behavior and evaluate the performance of various write strategies and buffering techniques for a MIN-based multiprocessor system using the simple software coherence scheme. Hit ratios, memory latencies, total execution time, and total write traffic are used as the performance indices. The write-through write-allocate no-fetch cache using a write-back write buffer is shown to have a better performance than both write-through and write-back caches. This type of write buffer is effective in reducing the volume as well as bursts of write traffic. On average, the use of a write-back cache reduces by 60 percent the total write traffic generated by a write-through cache.
NASA Astrophysics Data System (ADS)
Li, Hao; Xie, Lunguo
2013-03-01
The design of cache system for Chip Multiprocessor (CMP) face many challenges because future CMPs will have more cores and greater on-chip cache capacity. There are two base design schemes about L2 cache: private scheme in which each L2 slice is treated as a private L2 cache and shared scheme in which all L2 slices are treated as a large L2 cache shared by all cores. Private caches provide the lowest hit latency but reduce the total effective cache capacity. A shared L2 cache increases the effective cache capacity but has long hit latencies when data is on a remote tile. This paper present a new Controlled Replication (CR) policy to reduce the capacities occupied by redundant shared replicas. the new CR policy increases the effective capacity than victim replication scheme and has lower hit latency than shared scheme. We evaluate the various schemes using full-system simulation of parallel applications. Results show that CR reduces the average memory access latency of shared scheme by an average of 13%, providing better overall performance than victim replication and shared schemes.
Visits, Hits, Caching and Counting on the World Wide Web: Old Wine in New Bottles?
ERIC Educational Resources Information Center
Berthon, Pierre; Pitt, Leyland; Prendergast, Gerard
1997-01-01
Although web browser caching speeds up retrieval, reduces network traffic, and decreases the load on servers and browser's computers, an unintended consequence for marketing research is that Web servers undercount hits. This article explores counting problems, caching, proxy servers, trawler software and presents a series of correction factors…
Load Balancing in Distributed Web Caching: A Novel Clustering Approach
NASA Astrophysics Data System (ADS)
Tiwari, R.; Kumar, K.; Khan, G.
2010-11-01
The World Wide Web suffers from scaling and reliability problems due to overloaded and congested proxy servers. Caching at local proxy servers helps, but cannot satisfy more than a third to half of requests; more requests are still sent to original remote origin servers. In this paper we have developed an algorithm for Distributed Web Cache, which incorporates cooperation among proxy servers of one cluster. This algorithm uses Distributed Web Cache concepts along with static hierarchies with geographical based clusters of level one proxy server with dynamic mechanism of proxy server during the congestion of one cluster. Congestion and scalability problems are being dealt by clustering concept used in our approach. This results in higher hit ratio of caches, with lesser latency delay for requested pages. This algorithm also guarantees data consistency between the original server objects and the proxy cache objects.
Cache write generate for parallel image processing on shared memory architectures.
Wittenbrink, C M; Somani, A K; Chen, C H
1996-01-01
We investigate cache write generate, our cache mode invention. We demonstrate that for parallel image processing applications, the new mode improves main memory bandwidth, CPU efficiency, cache hits, and cache latency. We use register level simulations validated by the UW-Proteus system. Many memory, cache, and processor configurations are evaluated.
Toward Millions of File System IOPS on Low-Cost, Commodity Hardware
Zheng, Da; Burns, Randal; Szalay, Alexander S.
2013-01-01
We describe a storage system that removes I/O bottlenecks to achieve more than one million IOPS based on a user-space file abstraction for arrays of commodity SSDs. The file abstraction refactors I/O scheduling and placement for extreme parallelism and non-uniform memory and I/O. The system includes a set-associative, parallel page cache in the user space. We redesign page caching to eliminate CPU overhead and lock-contention in non-uniform memory architecture machines. We evaluate our design on a 32 core NUMA machine with four, eight-core processors. Experiments show that our design delivers 1.23 million 512-byte read IOPS. The page cache realizes the scalable IOPS of Linux asynchronous I/O (AIO) and increases user-perceived I/O performance linearly with cache hit rates. The parallel, set-associative cache matches the cache hit rates of the global Linux page cache under real workloads. PMID:24402052
Toward Millions of File System IOPS on Low-Cost, Commodity Hardware.
Zheng, Da; Burns, Randal; Szalay, Alexander S
2013-01-01
We describe a storage system that removes I/O bottlenecks to achieve more than one million IOPS based on a user-space file abstraction for arrays of commodity SSDs. The file abstraction refactors I/O scheduling and placement for extreme parallelism and non-uniform memory and I/O. The system includes a set-associative, parallel page cache in the user space. We redesign page caching to eliminate CPU overhead and lock-contention in non-uniform memory architecture machines. We evaluate our design on a 32 core NUMA machine with four, eight-core processors. Experiments show that our design delivers 1.23 million 512-byte read IOPS. The page cache realizes the scalable IOPS of Linux asynchronous I/O (AIO) and increases user-perceived I/O performance linearly with cache hit rates. The parallel, set-associative cache matches the cache hit rates of the global Linux page cache under real workloads.
A Global User-Driven Model for Tile Prefetching in Web Geographical Information Systems.
Pan, Shaoming; Chong, Yanwen; Zhang, Hang; Tan, Xicheng
2017-01-01
A web geographical information system is a typical service-intensive application. Tile prefetching and cache replacement can improve cache hit ratios by proactively fetching tiles from storage and replacing the appropriate tiles from the high-speed cache buffer without waiting for a client's requests, which reduces disk latency and improves system access performance. Most popular prefetching strategies consider only the relative tile popularities to predict which tile should be prefetched or consider only a single individual user's access behavior to determine which neighbor tiles need to be prefetched. Some studies show that comprehensively considering all users' access behaviors and all tiles' relationships in the prediction process can achieve more significant improvements. Thus, this work proposes a new global user-driven model for tile prefetching and cache replacement. First, based on all users' access behaviors, a type of expression method for tile correlation is designed and implemented. Then, a conditional prefetching probability can be computed based on the proposed correlation expression mode. Thus, some tiles to be prefetched can be found by computing and comparing the conditional prefetching probability from the uncached tiles set and, similarly, some replacement tiles can be found in the cache buffer according to multi-step prefetching. Finally, some experiments are provided comparing the proposed model with other global user-driven models, other single user-driven models, and other client-side prefetching strategies. The results show that the proposed model can achieve a prefetching hit rate in approximately 10.6% ~ 110.5% higher than the compared methods.
A Distributed Cache Update Deployment Strategy in CDN
NASA Astrophysics Data System (ADS)
E, Xinhua; Zhu, Binjie
2018-04-01
The CDN management system distributes content objects to the edge of the internet to achieve the user's near access. Cache strategy is an important problem in network content distribution. A cache strategy was designed in which the content effective diffusion in the cache group, so more content was storage in the cache, and it improved the group hit rate.
A Global User-Driven Model for Tile Prefetching in Web Geographical Information Systems
Pan, Shaoming; Chong, Yanwen; Zhang, Hang; Tan, Xicheng
2017-01-01
A web geographical information system is a typical service-intensive application. Tile prefetching and cache replacement can improve cache hit ratios by proactively fetching tiles from storage and replacing the appropriate tiles from the high-speed cache buffer without waiting for a client’s requests, which reduces disk latency and improves system access performance. Most popular prefetching strategies consider only the relative tile popularities to predict which tile should be prefetched or consider only a single individual user's access behavior to determine which neighbor tiles need to be prefetched. Some studies show that comprehensively considering all users’ access behaviors and all tiles’ relationships in the prediction process can achieve more significant improvements. Thus, this work proposes a new global user-driven model for tile prefetching and cache replacement. First, based on all users’ access behaviors, a type of expression method for tile correlation is designed and implemented. Then, a conditional prefetching probability can be computed based on the proposed correlation expression mode. Thus, some tiles to be prefetched can be found by computing and comparing the conditional prefetching probability from the uncached tiles set and, similarly, some replacement tiles can be found in the cache buffer according to multi-step prefetching. Finally, some experiments are provided comparing the proposed model with other global user-driven models, other single user-driven models, and other client-side prefetching strategies. The results show that the proposed model can achieve a prefetching hit rate in approximately 10.6% ~ 110.5% higher than the compared methods. PMID:28085937
A two-level cache for distributed information retrieval in search engines.
Zhang, Weizhe; He, Hui; Ye, Jianwei
2013-01-01
To improve the performance of distributed information retrieval in search engines, we propose a two-level cache structure based on the queries of the users' logs. We extract the highest rank queries of users from the static cache, in which the queries are the most popular. We adopt the dynamic cache as an auxiliary to optimize the distribution of the cache data. We propose a distribution strategy of the cache data. The experiments prove that the hit rate, the efficiency, and the time consumption of the two-level cache have advantages compared with other structures of cache.
A Two-Level Cache for Distributed Information Retrieval in Search Engines
Zhang, Weizhe; He, Hui; Ye, Jianwei
2013-01-01
To improve the performance of distributed information retrieval in search engines, we propose a two-level cache structure based on the queries of the users' logs. We extract the highest rank queries of users from the static cache, in which the queries are the most popular. We adopt the dynamic cache as an auxiliary to optimize the distribution of the cache data. We propose a distribution strategy of the cache data. The experiments prove that the hit rate, the efficiency, and the time consumption of the two-level cache have advantages compared with other structures of cache. PMID:24363621
Improving Internet Archive Service through Proxy Cache.
ERIC Educational Resources Information Center
Yu, Hsiang-Fu; Chen, Yi-Ming; Wang, Shih-Yong; Tseng, Li-Ming
2003-01-01
Discusses file transfer protocol (FTP) servers for downloading archives (files with particular file extensions), and the change to HTTP (Hypertext transfer protocol) with increased Web use. Topics include the Archie server; proxy cache servers; and how to improve the hit rate of archives by a combination of caching and better searching mechanisms.…
NASA Technical Reports Server (NTRS)
Gunawardena, J. A.
1992-01-01
This cache mechanism is transparent but does not contain associative circuits. It does not rely on locality of reference of instructions or data. No redundant instructions or data are encached. Items in the cache are accessed without address arithmetic. A cache miss is detected by the simplest test; compare two bits. These features would result in faster access, higher hit rate, reduced chip area, and less power dissipation in comparison with associative systems of similar size.
Study of cache performance in distributed environment for data processing
NASA Astrophysics Data System (ADS)
Makatun, Dzmitry; Lauret, Jérôme; Šumbera, Michal
2014-06-01
Processing data in distributed environment has found its application in many fields of science (Nuclear and Particle Physics (NPP), astronomy, biology to name only those). Efficiently transferring data between sites is an essential part of such processing. The implementation of caching strategies in data transfer software and tools, such as the Reasoner for Intelligent File Transfer (RIFT) being developed in the STAR collaboration, can significantly decrease network load and waiting time by reusing the knowledge of data provenance as well as data placed in transfer cache to further expand on the availability of sources for files and data-sets. Though, a great variety of caching algorithms is known, a study is needed to evaluate which one can deliver the best performance in data access considering the realistic demand patterns. Records of access to the complete data-sets of NPP experiments were analyzed and used as input for computer simulations. Series of simulations were done in order to estimate the possible cache hits and cache hits per byte for known caching algorithms. The simulations were done for cache of different sizes within interval 0.001 - 90% of complete data-set and low-watermark within 0-90%. Records of data access were taken from several experiments and within different time intervals in order to validate the results. In this paper, we will discuss the different data caching strategies from canonical algorithms to hybrid cache strategies, present the results of our simulations for the diverse algorithms, debate and identify the choice for the best algorithm in the context of Physics Data analysis in NPP. While the results of those studies have been implemented in RIFT, they can also be used when setting up cache in any other computational work-flow (Cloud processing for example) or managing data storages with partial replicas of the entire data-set.
A performance study of the time-varying cache behavior: a study on APEX, Mantevo, NAS, and PARSEC
Siddique, Nafiul A.; Grubel, Patricia A.; Badawy, Abdel-Hameed A.; ...
2017-09-20
Cache has long been used to minimize the latency of main memory accesses by storing frequently used data near the processor. Processor performance depends on the underlying cache performance. Therefore, significant research has been done to identify the most crucial metrics of cache performance. Although the majority of research focuses on measuring cache hit rates and data movement as the primary cache performance metrics, cache utilization is significantly important. We investigate the application’s locality using cache utilization metrics. In addition, we present cache utilization and traditional cache performance metrics as the program progresses providing detailed insights into the dynamic applicationmore » behavior on parallel applications from four benchmark suites running on multiple cores. We explore cache utilization for APEX, Mantevo, NAS, and PARSEC, mostly scientific benchmark suites. Our results indicate that 40% of the data bytes in a cache line are accessed at least once before line eviction. Also, on average a byte is accessed two times before the cache line is evicted for these applications. Moreover, we present runtime cache utilization, as well as, conventional performance metrics that illustrate a holistic understanding of cache behavior. To facilitate this research, we build a memory simulator incorporated into the Structural Simulation Toolkit (Rodrigues et al. in SIGMETRICS Perform Eval Rev 38(4):37–42, 2011). Finally, our results suggest that variable cache line size can result in better performance and can also conserve power.« less
A performance study of the time-varying cache behavior: a study on APEX, Mantevo, NAS, and PARSEC
DOE Office of Scientific and Technical Information (OSTI.GOV)
Siddique, Nafiul A.; Grubel, Patricia A.; Badawy, Abdel-Hameed A.
Cache has long been used to minimize the latency of main memory accesses by storing frequently used data near the processor. Processor performance depends on the underlying cache performance. Therefore, significant research has been done to identify the most crucial metrics of cache performance. Although the majority of research focuses on measuring cache hit rates and data movement as the primary cache performance metrics, cache utilization is significantly important. We investigate the application’s locality using cache utilization metrics. In addition, we present cache utilization and traditional cache performance metrics as the program progresses providing detailed insights into the dynamic applicationmore » behavior on parallel applications from four benchmark suites running on multiple cores. We explore cache utilization for APEX, Mantevo, NAS, and PARSEC, mostly scientific benchmark suites. Our results indicate that 40% of the data bytes in a cache line are accessed at least once before line eviction. Also, on average a byte is accessed two times before the cache line is evicted for these applications. Moreover, we present runtime cache utilization, as well as, conventional performance metrics that illustrate a holistic understanding of cache behavior. To facilitate this research, we build a memory simulator incorporated into the Structural Simulation Toolkit (Rodrigues et al. in SIGMETRICS Perform Eval Rev 38(4):37–42, 2011). Finally, our results suggest that variable cache line size can result in better performance and can also conserve power.« less
An Intelligent Cloud Storage Gateway for Medical Imaging.
Viana-Ferreira, Carlos; Guerra, António; Silva, João F; Matos, Sérgio; Costa, Carlos
2017-09-01
Historically, medical imaging repositories have been supported by indoor infrastructures. However, the amount of diagnostic imaging procedures has continuously increased over the last decades, imposing several challenges associated with the storage volume, data redundancy and availability. Cloud platforms are focused on delivering hardware and software services over the Internet, becoming an appealing solution for repository outsourcing. Although this option may bring financial and technological benefits, it also presents new challenges. In medical imaging scenarios, communication latency is a critical issue that still hinders the adoption of this paradigm. This paper proposes an intelligent Cloud storage gateway that optimizes data access times. This is achieved through a new cache architecture that combines static rules and pattern recognition for eviction and prefetching. The evaluation results, obtained from experiments over a real-world dataset, show that cache hit ratios can reach around 80%, leading to reductions of image retrieval times by over 60%. The combined use of eviction and prefetching policies proposed can significantly reduce communication latency, even when using a small cache in comparison to the total size of the repository. Apart from the performance gains, the proposed system is capable of adjusting to specific workflows of different institutions.
Collaborative video caching scheme over OFDM-based long-reach passive optical networks
NASA Astrophysics Data System (ADS)
Li, Yan; Dai, Shifang; Chang, Xiangmao
2018-07-01
Long-reach passive optical networks (LR-PONs) are now considered as a desirable access solution for cost-efficiently delivering broadband services by integrating metro network with access network, among which orthogonal frequency division multiplexing (OFDM)-based LR-PONs gain greater research interests due to their good robustness and high spectrum efficiency. In such attractive OFDM-based LR-PONs, however, it is still challenging to effectively provide video service, which is one of the most popular and profitable broadband services, for end users. Given that more video requesters (i.e., end users) far away from optical line terminal (OLT) are served in OFDM-based LR-PONs, it is efficiency-prohibitive to use traditional video delivery model, which relies on the OLT to transmit videos to requesters, for providing video service, due to the model will incur not only larger video playback delay but also higher downstream bandwidth consumption. In this paper, we propose a novel video caching scheme that to collaboratively cache videos on distributed optical network units (ONUs) which are closer to end users, and thus to timely and cost-efficiently provide videos for requesters by ONUs over OFDM-based LR-PONs. We firstly construct an OFDM-based LR-PON architecture to enable the cooperation among ONUs while caching videos. Given a limited storage capacity of each ONU, we then propose collaborative approaches to cache videos on ONUs with the aim to maximize the local video hit ratio (LVHR), i.e., the proportion of video requests that can be directly satisfied by ONUs, under diverse resources requirements and requests distributions of videos. Simulations are finally conducted to evaluate the efficiency of our proposed scheme.
Performance Prediction Toolkit
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chennupati, Gopinath; Santhi, Nanadakishore; Eidenbenz, Stephen
The Performance Prediction Toolkit (PPT), is a scalable co-design tool that contains the hardware and middle-ware models, which accept proxy applications as input in runtime prediction. PPT relies on Simian, a parallel discrete event simulation engine in Python or Lua, that uses the process concept, where each computing unit (host, node, core) is a Simian entity. Processes perform their task through message exchanges to remain active, sleep, wake-up, begin and end. The PPT hardware model of a compute core (such as a Haswell core) consists of a set of parameters, such as clock speed, memory hierarchy levels, their respective sizes,more » cache-lines, access times for different cache levels, average cycle counts of ALU operations, etc. These parameters are ideally read off a spec sheet or are learned using regression models learned from hardware counters (PAPI) data. The compute core model offers an API to the software model, a function called time_compute(), which takes as input a tasklist. A tasklist is an unordered set of ALU, and other CPU-type operations (in particular virtual memory loads and stores). The PPT application model mimics the loop structure of the application and replaces the computational kernels with a call to the hardware model's time_compute() function giving tasklists as input that model the compute kernel. A PPT application model thus consists of tasklists representing kernels and the high-er level loop structure that we like to think of as pseudo code. The key challenge for the hardware model's time_compute-function is to translate virtual memory accesses into actual cache hierarchy level hits and misses.PPT also contains another CPU core level hardware model, Analytical Memory Model (AMM). The AMM solves this challenge soundly, where our previous alternatives explicitly include the L1,L2,L3 hit-rates as inputs to the tasklists. Explicit hit-rates inevitably only reflect the application modeler's best guess, perhaps informed by a few small test problems using hardware counters; also, hard-coded hit-rates make the hardware model insensitive to changes in cache sizes. Alternatively, we use reuse distance distributions in the tasklists. In general, reuse profiles require the application modeler to run a very expensive trace analysis on the real code that realistically can be done at best for small examples.« less
The effect of code expanding optimizations on instruction cache design
NASA Technical Reports Server (NTRS)
Chen, William Y.; Chang, Pohua P.; Conte, Thomas M.; Hwu, Wen-Mei W.
1991-01-01
It is shown that code expanding optimizations have strong and non-intuitive implications on instruction cache design. Three types of code expanding optimizations are studied: instruction placement, function inline expansion, and superscalar optimizations. Overall, instruction placement reduces the miss ratio of small caches. Function inline expansion improves the performance for small cache sizes, but degrades the performance of medium caches. Superscalar optimizations increases the cache size required for a given miss ratio. On the other hand, they also increase the sequentiality of instruction access so that a simple load-forward scheme effectively cancels the negative effects. Overall, it is shown that with load forwarding, the three types of code expanding optimizations jointly improve the performance of small caches and have little effect on large caches.
An area model for on-chip memories and its application
NASA Technical Reports Server (NTRS)
Mulder, Johannes M.; Quach, Nhon T.; Flynn, Michael J.
1991-01-01
An area model suitable for comparing data buffers of different organizations and arbitrary sizes is described. The area model considers the supplied bandwidth of a memory cell and includes such buffer overhead as control logic, driver logic, and tag storage. The model gave less than 10 percent error when verified against real caches and register files. It is shown that, comparing caches and register files in terms of area for the same storage capacity, caches generally occupy more area per bit than register files for small caches because the overhead dominates the cache area at these sizes. For larger caches, the smaller storage cells in the cache provide a smaller total cache area per bit than the register set. Studying cache performance (traffic ratio) as a function of area, it is shown that, for small caches, direct-mapped caches perform significantly better than four-way set-associative caches and, for caches of medium areas, both direct-mapped and set-associative caches perform better than fully associative caches.
Multiprocessor architectural study
NASA Technical Reports Server (NTRS)
Kosmala, A. L.; Stanten, S. F.; Vandever, W. H.
1972-01-01
An architectural design study was made of a multiprocessor computing system intended to meet functional and performance specifications appropriate to a manned space station application. Intermetrics, previous experience, and accumulated knowledge of the multiprocessor field is used to generate a baseline philosophy for the design of a future SUMC* multiprocessor. Interrupts are defined and the crucial questions of interrupt structure, such as processor selection and response time, are discussed. Memory hierarchy and performance is discussed extensively with particular attention to the design approach which utilizes a cache memory associated with each processor. The ability of an individual processor to approach its theoretical maximum performance is then analyzed in terms of a hit ratio. Memory management is envisioned as a virtual memory system implemented either through segmentation or paging. Addressing is discussed in terms of various register design adopted by current computers and those of advanced design.
Using Minimum-Surface Bodies for Iteration Space Partitioning
NASA Technical Reports Server (NTRS)
Frumlin, Michael; VanderWijngaart, Rob F.; Biegel, Bryan (Technical Monitor)
2001-01-01
A number of known techniques for improving cache performance in scientific computations involve the reordering of the iteration space. Some of these reorderings can be considered as coverings of the iteration space with the sets having good surface-to-volume ratio. Use of such sets reduces the number of cache misses in computations of local operators having the iteration space as a domain. We study coverings of iteration spaces represented by structured and unstructured grids. For structured grids we introduce a covering based on successive minima tiles of the interference lattice of the grid. We show that the covering has good surface-to-volume ratio and present a computer experiment showing actual reduction of the cache misses achieved by using these tiles. For unstructured grids no cache efficient covering can be guaranteed. We present a triangulation of a 3-dimensional cube such that any local operator on the corresponding grid has significantly larger number of cache misses than a similar operator on a structured grid.
Minimizing Cache Misses Using Minimum-Surface Bodies
NASA Technical Reports Server (NTRS)
Frumkin, Michael; VanderWijngaart, Rob; Biegel, Bryan (Technical Monitor)
2002-01-01
A number of known techniques for improving cache performance in scientific computations involve the reordering of the iteration space. Some of these reorderings can be considered as coverings of the iteration space with the sets having good surface-to-volume ratio. Use of such sets reduces the number of cache misses in computations of local operators having the iteration space as a domain. First, we derive lower bounds which any algorithm must suffer while computing a local operator on a grid. Then we explore coverings of iteration spaces represented by structured and unstructured grids which allow us to approach these lower bounds. For structured grids we introduce a covering by successive minima tiles of the interference lattice of the grid. We show that the covering has low surface-to-volume ratio and present a computer experiment showing actual reduction of the cache misses achieved by using these tiles. For planar unstructured grids we show existence of a covering which reduces the number of cache misses to the level of structured grids. On the other hand, we present a triangulation of a 3-dimensional cube such that any local operator on the corresponding grid has significantly larger number of cache misses than a similar operator on a structured grid.
Caching Joint Shortcut Routing to Improve Quality of Service for Information-Centric Networking.
Huang, Baixiang; Liu, Anfeng; Zhang, Chengyuan; Xiong, Naixue; Zeng, Zhiwen; Cai, Zhiping
2018-05-29
Hundreds of thousands of ubiquitous sensing (US) devices have provided an enormous number of data for Information-Centric Networking (ICN), which is an emerging network architecture that has the potential to solve a great variety of issues faced by the traditional network. A Caching Joint Shortcut Routing (CJSR) scheme is proposed in this paper to improve the Quality of service (QoS) for ICN. The CJSR scheme mainly has two innovations which are different from other in-network caching schemes: (1) Two routing shortcuts are set up to reduce the length of routing paths. Because of some inconvenient transmission processes, the routing paths of previous schemes are prolonged, and users can only request data from Data Centers (DCs) until the data have been uploaded from Data Producers (DPs) to DCs. Hence, the first kind of shortcut is built from DPs to users directly. This shortcut could release the burden of whole network and reduce delay. Moreover, in the second shortcut routing method, a Content Router (CR) which could yield shorter length of uploading routing path from DPs to DCs is chosen, and then data packets are uploaded through this chosen CR. In this method, the uploading path shares some segments with the pre-caching path, thus the overall length of routing paths is reduced. (2) The second innovation of the CJSR scheme is that a cooperative pre-caching mechanism is proposed so that QoS could have a further increase. Besides being used in downloading routing, the pre-caching mechanism can also be used when data packets are uploaded towards DCs. Combining uploading and downloading pre-caching, the cooperative pre-caching mechanism exhibits high performance in different situations. Furthermore, to address the scarcity of storage size, an algorithm that could make use of storage from idle CRs is proposed. After comparing the proposed scheme with five existing schemes via simulations, experiments results reveal that the CJSR scheme could reduce the total number of processed interest packets by 54.8%, enhance the cache hits of each CR and reduce the number of total hop counts by 51.6% and cut down the length of routing path for users to obtain their interested data by 28.6⁻85.7% compared with the traditional NDN scheme. Moreover, the length of uploading routing path could be decreased by 8.3⁻33.3%.
Cache Hardware Approaches to Multiple Independent Levels of Security (MILS)
2012-10-01
systems that require that several multicore processors be connected together in a single system. However, no such boards were available on the market ...available concerning each module. However, the availability of modules seems to significantly lag the time when the corresponding hardware hits the market ...version of real mode often referred to as “Unreal mode” can be entered by loading a Local Descriptor Table (LDT) and Global Descriptor Table (GDT
Considering User's Access Pattern in Multimedia File Systems
NASA Astrophysics Data System (ADS)
Cho, KyoungWoon; Ryu, YeonSeung; Won, Youjip; Koh, Kern
2002-12-01
Legacy buffer cache management schemes for multimedia server are grounded at the assumption that the application sequentially accesses the multimedia file. However, user access pattern may not be sequential in some circumstances, for example, in distance learning application, where the user may exploit the VCR-like function(rewind and play) of the system and accesses the particular segments of video repeatedly in the middle of sequential playback. Such a looping reference can cause a significant performance degradation of interval-based caching algorithms. And thus an appropriate buffer cache management scheme is required in order to deliver desirable performance even under the workload that exhibits looping reference behavior. We propose Adaptive Buffer cache Management(ABM) scheme which intelligently adapts to the file access characteristics. For each opened file, ABM applies either the LRU replacement or the interval-based caching depending on the Looping Reference Indicator, which indicates that how strong temporally localized access pattern is. According to our experiment, ABM exhibits better buffer cache miss ratio than interval-based caching or LRU, especially when the workload exhibits not only sequential but also looping reference property.
Efficient Cache use for Stencil Operations on Structured Discretization Grids
NASA Technical Reports Server (NTRS)
Frumkin, Michael; VanderWijngaart, Rob F.
2001-01-01
We derive tight bounds on the cache misses for evaluation of explicit stencil operators on structured grids. Our lower bound is based on the isoperimetrical property of the discrete octahedron. Our upper bound is based on a good surface to volume ratio of a parallelepiped spanned by a reduced basis of the interference lattice of a grid. Measurements show that our algorithm typically reduces the number of cache misses by a factor of three, relative to a compiler optimized code. We show that stencil calculations on grids whose interference lattice have a short vector feature abnormally high numbers of cache misses. We call such grids unfavorable and suggest to avoid these in computations by appropriate padding. By direct measurements on a MIPS R10000 processor we show a good correlation between abnormally high numbers of cache misses and unfavorable three-dimensional grids.
Single-pass memory system evaluation for multiprogramming workloads
NASA Technical Reports Server (NTRS)
Conte, Thomas M.; Hwu, Wen-Mei W.
1990-01-01
Modern memory systems are composed of levels of cache memories, a virtual memory system, and a backing store. Varying more than a few design parameters and measuring the performance of such systems has traditionally be constrained by the high cost of simulation. Models of cache performance recently introduced reduce the cost simulation but at the expense of accuracy of performance prediction. Stack-based methods predict performance accurately using one pass over the trace for all cache sizes, but these techniques have been limited to fully-associative organizations. This paper presents a stack-based method of evaluating the performance of cache memories using a recurrence/conflict model for the miss ratio. Unlike previous work, the performance of realistic cache designs, such as direct-mapped caches, are predicted by the method. The method also includes a new approach to the problem of the effects of multiprogramming. This new technique separates the characteristics of the individual program from that of the workload. The recurrence/conflict method is shown to be practical, general, and powerful by comparing its performance to that of a popular traditional cache simulator. The authors expect that the availability of such a tool will have a large impact on future architectural studies of memory systems.
Samelius, Gustaf; Alisauskas, Ray T; Hobson, Keith A; Larivière, Serge
2007-09-01
1. Many ecosystems are characterized by pulses of dramatically higher than normal levels of foods (pulsed resources) to which animals often respond by caching foods for future use. However, the extent to which animals use cached foods and how this varies in relation to fluctuations in other foods is poorly understood in most animals. 2. Arctic foxes Alopex lagopus (L.) cache thousands of eggs annually at large goose colonies where eggs are often superabundant during the nesting period by geese. We estimated the contribution of cached eggs to arctic fox diets in spring and autumn, when geese were not present in the study area, by comparing stable isotope ratios (delta(13)C and delta(15)N) of fox tissues with those of their foods using a multisource mixing model in Program IsoSource. 3. The contribution of cached eggs to arctic fox diets was inversely related to collared lemming Dicrostonyx groenlandicus (Traill) abundance; the contribution of cached eggs to overall fox diets increased from < 28% in years when collared lemmings were abundant to 30-74% in years when collared lemmings were scarce. 4. Further, arctic foxes used cached eggs well into the following spring (almost 1 year after eggs were acquired) - a pattern that differs from that of carnivores generally storing foods for only a few days before consumption. 5. This study showed that long-term use of eggs that were cached when geese were superabundant at the colony in summer varied with fluctuations in collared lemming abundance (a key component in arctic fox diets throughout most of their range) and suggests that cached eggs functioned as a buffer when collared lemmings were scarce.
Gonzalez Murcia, Josue D; Schmutz, Cameron; Munger, Caitlin; Perkes, Ammon; Gustin, Aaron; Peterson, Michael; Ebbert, Mark T W; Norton, Maria C; Tschanz, Joann T; Munger, Ronald G; Corcoran, Christopher D; Kauwe, John S K
2013-12-01
Recent studies have identified the rs75932628 (R47H) variant in TREM2 as an Alzheimer's disease risk factor with estimated odds ratio ranging from 2.9 to 5.1. The Cache County Memory Study is a large, population-based sample designed for the study of memory and aging. We genotyped R47H in 2974 samples (427 cases and 2540 control subjects) from the Cache County study using a custom TaqMan assay. We observed 7 heterozygous cases and 12 heterozygous control subjects with an odds ratio of 3.5 (95% confidence interval, 1.3-8.8; p = 0.0076). The minor allele frequency and population attributable fraction for R47H were 0.0029 and 0.004, respectively. This study replicates the association between R47H and Alzheimer's disease risk in a large, population-based sample, and estimates the population frequency and attributable risk of this rare variant. Copyright © 2013 Elsevier Inc. All rights reserved.
Cache-enabled small cell networks: modeling and tradeoffs.
Baştuǧ, Ejder; Bennis, Mehdi; Kountouris, Marios; Debbah, Mérouane
We consider a network model where small base stations (SBSs) have caching capabilities as a means to alleviate the backhaul load and satisfy users' demand. The SBSs are stochastically distributed over the plane according to a Poisson point process (PPP) and serve their users either (i) by bringing the content from the Internet through a finite rate backhaul or (ii) by serving them from the local caches. We derive closed-form expressions for the outage probability and the average delivery rate as a function of the signal-to-interference-plus-noise ratio (SINR), SBS density, target file bitrate, storage size, file length, and file popularity. We then analyze the impact of key operating parameters on the system performance. It is shown that a certain outage probability can be achieved either by increasing the number of base stations or the total storage size. Our results and analysis provide key insights into the deployment of cache-enabled small cell networks (SCNs), which are seen as a promising solution for future heterogeneous cellular networks.
NASA Astrophysics Data System (ADS)
Poat, M. D.; Lauret, J.
2017-10-01
As demand for widely accessible storage capacity increases and usage is on the rise, steady IO performance is desired but tends to suffer within multi-user environments. Typical deployments use standard hard drives as the cost per/GB is quite low. On the other hand, HDD based solutions for storage is not known to scale well with process concurrency and soon enough, high rate of IOPs create a “random access” pattern killing performance. Though not all SSDs are alike, SSDs are an established technology often used to address this exact “random access” problem. In this contribution, we will first discuss the IO performance of many different SSD drives (tested in a comparable and standalone manner). We will then be discussing the performance and integrity of at least three low-level disk caching techniques (Flashcache, dm-cache, and bcache) including individual policies, procedures, and IO performance. Furthermore, the STAR online computing infrastructure currently hosts a POSIX-compliant Ceph distributed storage cluster - while caching is not a native feature of CephFS (only exists in the Ceph Object store), we will show how one can implement a caching mechanism profiting from an implementation at a lower level. As our illustration, we will present our CephFS setup, IO performance tests, and overall experience from such configuration. We hope this work will service the community’s interest for using disk-caching mechanisms with applicable uses such as distributed storage systems and seeking an overall IO performance gain.
Wienert, Stephan; Beil, Michael; Saeger, Kai; Hufnagl, Peter; Schrader, Thomas
2009-01-09
The virtual microscopy is widely accepted in Pathology for educational purposes and teleconsultation but is far from the routine use in surgical pathology due to the technical requirements and some limitations. A technical problem is the limited bandwidth of a usual network and the delayed transmission rate and presentation time on the screen. In this study the process of secondary diagnostic was evaluated using the "T.Konsult Pathologie" service of the Professional Association of German Pathologists within the German breast cancer screening program. The characteristics of the access to the WSI (Whole Slide Images) have been analyzed to explore the possibilities of prefetching and caching to reduce the presentation and transfer time with the goal to increase user acceptance. The log files of the web server were analyzed to reconstruct the movements of the pathologist on the WSI and to create the observation path. Using a specialized tool the observation paths were extracted automatically from the log files. The attributes linearity, 3-point-linearity, changes per request, and number of consecutive requests were calculated to design, develop and evaluate different caching and prefetching strategies. The analysis of the observation paths showed that a complete accordance of two image requests is a very rare event. But more frequently a partial covering of two requested image areas can be found. In total 257 diagnostic paths from 131 WSI have been extracted and analysed. On average a diagnostic path consists of 16 image requests and takes 189 seconds between first and last image request. The mean linearity was 0,41 and the mean 3-point-linearity 0,85. Three different caching algorithms have been compared with respect to hit rate and additional image requests on the WSI server. Tests demonstrated that 95% of the diagnostic paths could be loaded without any deletion of entries in the cache (cache size 12,2 Megapixel). If the image parts are stored after JPEG compression this complies with less than 2 MB. WSI telepathology is a technology which offers the possibility to break the limitations of conventional static telepathology. The complete histological slide may be investigated instead of sets of images of lesions sampled by the presenting pathologist. The benefit is demonstrated by the high diagnostic security of 95% accordance between first and second diagnosis.
Wienert, Stephan; Beil, Michael; Saeger, Kai; Hufnagl, Peter; Schrader, Thomas
2009-01-01
Background The virtual microscopy is widely accepted in Pathology for educational purposes and teleconsultation but is far from the routine use in surgical pathology due to the technical requirements and some limitations. A technical problem is the limited bandwidth of a usual network and the delayed transmission rate and presentation time on the screen. Methods In this study the process of secondary diagnostic was evaluated using the "T.Konsult Pathologie" service of the Professional Association of German Pathologists within the German breast cancer screening program. The characteristics of the access to the WSI (Whole Slide Images) have been analyzed to explore the possibilities of prefetching and caching to reduce the presentation and transfer time with the goal to increase user acceptance. The log files of the web server were analyzed to reconstruct the movements of the pathologist on the WSI and to create the observation path. Using a specialized tool the observation paths were extracted automatically from the log files. The attributes linearity, 3-point-linearity, changes per request, and number of consecutive requests were calculated to design, develop and evaluate different caching and prefetching strategies. Results The analysis of the observation paths showed that a complete accordance of two image requests is a very rare event. But more frequently a partial covering of two requested image areas can be found. In total 257 diagnostic paths from 131 WSI have been extracted and analysed. On average a diagnostic path consists of 16 image requests and takes 189 seconds between first and last image request. The mean linearity was 0,41 and the mean 3-point-linearity 0,85. Three different caching algorithms have been compared with respect to hit rate and additional image requests on the WSI server. Tests demonstrated that 95% of the diagnostic paths could be loaded without any deletion of entries in the cache (cache size 12,2 Megapixel). If the image parts are stored after JPEG compression this complies with less than 2 MB. Discussion WSI telepathology is a technology which offers the possibility to break the limitations of conventional static telepathology. The complete histological slide may be investigated instead of sets of images of lesions sampled by the presenting pathologist. The benefit is demonstrated by the high diagnostic security of 95% accordance between first and second diagnosis. PMID:19134181
Interference Lattice-based Loop Nest Tilings for Stencil Computations
NASA Technical Reports Server (NTRS)
VanderWijngaart, Rob F.; Frumkin, Michael
2000-01-01
A common method for improving performance of stencil operations on structured multi-dimensional discretization grids is loop tiling. Tile shapes and sizes are usually determined heuristically, based on the size of the primary data cache. We provide a lower bound on the numbers of cache misses that must be incurred by any tiling, and a close achievable bound using a particular tiling based on the grid interference lattice. The latter tiling is used to derive highly efficient loop orderings. The total number of cache misses of a code is the sum of (necessary) cold misses and misses caused by elements being dropped from the cache between successive loads (replacement misses). Maximizing temporal locality is equivalent to minimizing replacement misses. Temporal locality of loop nests implementing stencil operations is optimized by tilings that avoid data conflicts. We divide the loop nest iteration space into conflict-free tiles, derived from the cache miss equation. The tiling involves the definition of the grid interference lattice an equivalence class of grid points whose images in main memory map to the same location in the cache-and the construction of a special basis for the lattice. Conflicts only occur on the boundaries of the tiles, unless the tiles are too thin. We show that the surface area of the tiles is bounded for grids of any dimensionality, and for caches of any associativity, provided the eccentricity of the fundamental parallelepiped (the tile spanned by the basis) of the lattice is bounded. Eccentricity is determined by two factors, aspect ratio and skewness. The aspect ratio of the parallelepiped can be bounded by appropriate array padding. The skewness can be bounded by the choice of a proper basis. Combining these two strategies ensures that pathologically thin tiles are avoided. They do not, however, minimize replacement misses per se. The reason is that tile visitation order influences the number of data conflicts on the tile boundaries. If two adjacent tiles are visited successively, there will be no replacement misses on the shared boundary. The iteration space may be covered with pencils larger than the size of the cache while avoiding data conflicts if the pencils are traversed by a scanning-face method. Replacement misses are incurred only on the boundaries of the pencils, and the number of misses is minimized by maximizing the volume of the scanning face, not the volume of the tile. We present an algorithm for constructing the most efficient scanning face for a given grid and stencil operator. In two dimensions it is based on a continued fraction algorithm. In three dimensions it follows Voronoi's successive minima algorithm. We show experimental results of using the scanning face, and compare with canonical loop orderings.
Parallelization Issues and Particle-In Codes.
NASA Astrophysics Data System (ADS)
Elster, Anne Cathrine
1994-01-01
"Everything should be made as simple as possible, but not simpler." Albert Einstein. The field of parallel scientific computing has concentrated on parallelization of individual modules such as matrix solvers and factorizers. However, many applications involve several interacting modules. Our analyses of a particle-in-cell code modeling charged particles in an electric field, show that these accompanying dependencies affect data partitioning and lead to new parallelization strategies concerning processor, memory and cache utilization. Our test-bed, a KSR1, is a distributed memory machine with a globally shared addressing space. However, most of the new methods presented hold generally for hierarchical and/or distributed memory systems. We introduce a novel approach that uses dual pointers on the local particle arrays to keep the particle locations automatically partially sorted. Complexity and performance analyses with accompanying KSR benchmarks, have been included for both this scheme and for the traditional replicated grids approach. The latter approach maintains load-balance with respect to particles. However, our results demonstrate it fails to scale properly for problems with large grids (say, greater than 128-by-128) running on as few as 15 KSR nodes, since the extra storage and computation time associated with adding the grid copies, becomes significant. Our grid partitioning scheme, although harder to implement, does not need to replicate the whole grid. Consequently, it scales well for large problems on highly parallel systems. It may, however, require load balancing schemes for non-uniform particle distributions. Our dual pointer approach may facilitate this through dynamically partitioned grids. We also introduce hierarchical data structures that store neighboring grid-points within the same cache -line by reordering the grid indexing. This alignment produces a 25% savings in cache-hits for a 4-by-4 cache. A consideration of the input data's effect on the simulation may lead to further improvements. For example, in the case of mean particle drift, it is often advantageous to partition the grid primarily along the direction of the drift. The particle-in-cell codes for this study were tested using physical parameters, which lead to predictable phenomena including plasma oscillations and two-stream instabilities. An overview of the most central references related to parallel particle codes is also given.
Study on data acquisition system based on reconfigurable cache technology
NASA Astrophysics Data System (ADS)
Zhang, Qinchuan; Li, Min; Jiang, Jun
2018-03-01
Waveform capture rate is one of the key features of digital acquisition systems, which represents the waveform processing capability of the system in a unit time. The higher the waveform capture rate is, the larger the chance to capture elusive events is and the more reliable the test result is. First, this paper analyzes the impact of several factors on the waveform capture rate of the system, then the novel technology based on reconfigurable cache is further proposed to optimize system architecture, and the simulation results show that the signal-to-noise ratio of signal, capacity, and structure of cache have significant effects on the waveform capture rate. Finally, the technology is demonstrated by the engineering practice, and the results show that the waveform capture rate of the system is improved substantially without significant increase of system's cost, and the technology proposed has a broad application prospect.
Population-based analysis of Alzheimer's disease risk alleles implicates genetic interactions.
Ebbert, Mark T W; Ridge, Perry G; Wilson, Andrew R; Sharp, Aaron R; Bailey, Matthew; Norton, Maria C; Tschanz, JoAnn T; Munger, Ronald G; Corcoran, Christopher D; Kauwe, John S K
2014-05-01
Reported odds ratios and population attributable fractions (PAF) for late-onset Alzheimer's disease (LOAD) risk loci (BIN1, ABCA7, CR1, MS4A4E, CD2AP, PICALM, MS4A6A, CD33, and CLU) come from clinically ascertained samples. Little is known about the combined PAF for these LOAD risk alleles and the utility of these combined markers for case-control prediction. Here we evaluate these loci in a large population-based sample to estimate PAF and explore the effects of additive and nonadditive interactions on LOAD status prediction performance. 2419 samples from the Cache County Memory Study were genotyped for APOE and nine LOAD risk loci from AlzGene.org. We used logistic regression and receiver operator characteristic analysis to assess the LOAD status prediction performance of these loci using additive and nonadditive models and compared odds ratios and PAFs between AlzGene.org and Cache County. Odds ratios were comparable between Cache County and AlzGene.org when identical single nucleotide polymorphisms were genotyped. PAFs from AlzGene.org ranged from 2.25% to 37%; those from Cache County ranged from .05% to 20%. Including non-APOE alleles significantly improved LOAD status prediction performance (area under the curve = .80) over APOE alone (area under the curve = .78) when not constrained to an additive relationship (p < .03). We identified potential allelic interactions (p values uncorrected): CD33-MS4A4E (synergy factor = 5.31; p < .003) and CLU-MS4A4E (synergy factor = 3.81; p < .016). Although nonadditive interactions between loci significantly improve diagnostic ability, the improvement does not reach the desired sensitivity or specificity for clinical use. Nevertheless, these results suggest that understanding gene-gene interactions may be important in resolving Alzheimer's disease etiology. Copyright © 2014 Society of Biological Psychiatry. Published by Elsevier Inc. All rights reserved.
Freye, Chris E; Moore, Nicholas R; Synovec, Robert E
2018-02-16
The complementary information provided by tandem ionization time-of-flight mass spectrometry (TI-TOFMS) is investigated for comparative discovery-based analysis, when coupled with comprehensive two-dimensional gas chromatography (GC × GC). The TI conditions implemented were a hard ionization energy (70 eV) concurrently collected with a soft ionization energy (14 eV). Tile-based Fisher ratio (F-ratio) analysis is used to analyze diesel fuel spiked with twelve analytes at a nominal concentration of 50 ppm. F-ratio analysis is a supervised discovery-based technique that compares two different sample classes, in this case spiked and unspiked diesel, to reduce the complex GC × GC-TI-TOFMS data into a hit list of class distinguishing analyte features. Hit lists of the 70 eV and 14 eV data sets, and the single hit list produced when the two data sets are fused together, are all investigated. For the 70 eV hit list, eleven of the twelve analytes were found in the top thirteen hits. For the 14 eV hit list, nine of the twelve analytes were found in the top nine hits, with the other three analytes either not found or well down the hit list. As expected, the F-ratios per m/z used to calculate each average F-ratio per hit were generally smaller fragment ions for the 70 eV data set, while the larger fragment ions were emphasized in the 14 eV data set, supporting the notion that complementary information was provided. The discovery rate was improved when F-ratio analysis was performed on the fused data sets resulted in eleven of the twelve analytes being at the top of the single hit list. Using PARAFAC, analytes that were "discovered" were deconvoluted in order to obtain their identification via match values (MV). Location of the analytes and the "F-ratio spectra" obtained from F-ratio analysis were used to guide the deconvolution. Eight of the twelve analytes where successfully deconvoluted and identified using the in-house library for the 70 eV data set. PARAFAC deconvolution of the two separate data sets provided increased confidence in identification of "discovered" analytes. Herein, we explore the limit of analyte discovery and limit of analyte identification, and demonstrate a general workflow for the investigation of key chemical features in complex samples. Copyright © 2018 Elsevier B.V. All rights reserved.
The ontogeny of food-caching behaviour in New Zealand robins (Petroica longipes).
Clark, Lisabertha L; Shaw, Rachael C
2018-06-01
Hoarding or caching behaviour is a widely-used paradigm for examining a range of cognitive processes in birds, such as social cognition and spatial memory. However, much is still unknown about how caching develops in young birds, especially in the wild. Studying the ontogeny of caching in the wild will help researchers to identify the mechanisms that shape this advantageous foraging strategy. We examined the ontogeny of food caching behaviour in a wild New Zealand passerine, the North Island robin (Petroica longipes). For 12-weeks following fledging, we observed 34 juveniles to examine the development of caching and cache retrieval. Additionally, we compared the caching behaviour of juveniles at 12 weeks post-fledging to 35 adult robins to determine whether juveniles had developed adult-like caching behaviour by this age. Juveniles began caching mealworms shortly after achieving foraging independency. Multivariate analyses revealed that caching rate increased and handling time decreased with increasing age. Juveniles spontaneously began retrieving caches as soon as they had begun to cache and their retrieval rates then remained constant throughout their ensuing development. Likewise, the number of sites used by juveniles did not change with age. Juvenile sex, caregiver sex and the duration of post-fledging parental care did not influence the development of caching, cache retrieval, the number of cache sites used and the time juveniles spent handling mealworms. At 12 weeks post-fledging, juveniles demonstrated levels of caching, cache retrieval and cache site usage that were comparable to adults. However, juvenile prey handling time was still longer than adults. The spontaneous emergence of cache retrieval and the consistency in the number of cache sites used throughout development suggests that these aspects of caching in North Island robins are likely to be innate, but that age and experience have an important role in the development of adult caching behaviours. Copyright © 2018 Elsevier B.V. All rights reserved.
The Helicity Injected Torus Program
NASA Astrophysics Data System (ADS)
Jarboe, T. R.; Nelson, B. A.; Jewell, P. D.; Liptac, J. E.; McCollam, K. J.; Raman, R.; Redd, A. J.; Rogers, J. A.; Sieck, P. E.; Shumlak, U.; Smith, R. J.; Nagata, M.; Uyama, T.
1999-11-01
The Helicity Injected Torus--II (HIT--II) spherical torus is capable of both Coaxial Helicity Injection (CHI) and transformer action current drive. HIT--II has a major radius R = 0.3, minor radius a = 0.2, aspect ratio A = R/a = 1.5, with an on axis magnetic field of up to Bo = 0.67 T. HIT--II provides equilibrium control, CHI flux boundary conditions, and transformer action using 28 poloidal field coils, using active flux feedback control. HIT--II has driven up to 200 kA of plasma current, using either CHI or transformer drive. An overview and recent results of the HIT--II program will be presented. The development of a locked-electron current drive model for HIT and HIT--II has led to the design of a constant inductive helicity injection method for spherical torii. This method is incorporated in the design of the Helicity Injected Torus -- Steady Inductive (HIT-- SI)(T.R. Jarboe, Fusion Technology, 36) (1), p. 85, 1999 experiment. HIT--SI can form a high-beta spheromak, a low aspect ratio RFP, or a spherical tokamak in a steady-state manner without using electrodes. The HIT--SI design and methodology will be presented.
Clark’s Nutcrackers (Nucifraga columbiana) Flexibly Adapt Caching Behavior to a Cooperative Context
Clary, Dawson; Kelly, Debbie M.
2016-01-01
Corvids recognize when their caches are at risk of being stolen by others and have developed strategies to protect these caches from pilferage. For instance, Clark’s nutcrackers will suppress the number of caches they make if being observed by a potential thief. However, cache protection has most often been studied using competitive contexts, so it is unclear whether corvids can adjust their caching in beneficial ways to accommodate non-competitive situations. Therefore, we examined whether Clark’s nutcrackers, a non-social corvid, would flexibly adapt their caching behaviors to a cooperative context. To do so, birds were given a caching task during which caches made by one individual were reciprocally exchanged for the caches of a partner bird over repeated trials. In this scenario, if caching behaviors can be flexibly deployed, then the birds should recognize the cooperative nature of the task and maintain or increase caching levels over time. However, if cache protection strategies are applied independent of social context and simply in response to cache theft, then cache suppression should occur. In the current experiment, we found that the birds maintained caching throughout the experiment. We report that males increased caching in response to a manipulation in which caches were artificially added, suggesting the birds could adapt to the cooperative nature of the task. Additionally, we show that caching decisions were not solely due to motivational factors, instead showing an additional influence attributed to the behavior of the partner bird. PMID:27826273
The Helicity Injected Torus (HIT) Program
NASA Astrophysics Data System (ADS)
Jarboe, T. R.; Gu, P.; Hamp, W.; Izzo, V.; Jewell, P.; Liptac, J.; McCollam, K. J.; Nelson, B. A.; Raman, R.; Redd, A. J.; Shumlak, U.; Sieck, P. E.; Smith, R. J.; Jain, K. K.; Nagata, M.; Uyama, T.
2000-10-01
The purpose of the Helicity Injected Torus (HIT) program is to develop current drive techniques for low-aspect-ratio toroidal plasmas. The present HIT-II spherical tokamak experiment is capable of both Coaxial Helicity Injection (CHI) and transformer action current drive. The HIT-II device itself is modestly sized (major radius R = 0.3 m, minor radius a = 0.2 m, with an on-axis magnetic field of up to Bo = 0.5 T), but has demonstrated toroidal plasma currents of up to 200 kA, using either CHI or transformer drive. An overview of ongoing research on HIT-II plasmas, including recent results, will be presented. An electron-locking model has been developed for helicity injection current drive; a description of this model will be presented, as well as comparisons to experimental results from the HIT and HIT-II devices. Empirical results from both the HIT program and past spheromak research, buttressed by theoretical developments, have led to the design of the upcoming HIT-SI (Helicity Injected Torus with Steady Inductive helicity injection) device (T.R. Jarboe, Fusion Technology 36, p. 85, 1999). HIT-SI will be able to form a high-beta spheromak, a low aspect ratio RFP or a spherical tokamak using constant inductive helicity injection. The HIT-SI design and construction progress will be presented.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, Lingda; Hayes, Ari; Song, Shuaiwen
Modern GPUs employ cache to improve memory system efficiency. However, large amount of cache space is underutilized due to irregular memory accesses and poor spatial locality which exhibited commonly in GPU applications. Our experiments show that using smaller cache lines could improve cache space utilization, but it also frequently suffers from significant performance loss by introducing large amount of extra cache requests. In this work, we propose a novel cache design named tag-split cache (TSC) that enables fine-grained cache storage to address the problem of cache space underutilization while keeping memory request number unchanged. TSC divides tag into two partsmore » to reduce storage overhead, and it supports multiple cache line replacement in one cycle.« less
NASA Astrophysics Data System (ADS)
Yuasa, H.; Hara, M.; Murakami, S.; Fuji, Y.; Fukuzawa, H.; Zhang, K.; Li, M.; Schreck, E.; Wang, P.; Chen, M.
2010-09-01
We have enhanced magnetoresistance (MR) for current-perpendicular-to-plane giant-magnetoresistive (CPP-GMR) films with a current-confined-path nano-oxide layer (CCP-NOL). In order to realize higher purity in Cu for CCPs, hydrogen ion treatment (HIT) was applied as the CuOx reduction process. By applying the HIT process, an MR ratio was increased to 27.4% even in the case of using conventional FeCo magnetic layer, from 13.0% for a reference without the HIT process. Atom probe tomography data confirmed oxygen reduction by the HIT process in the CCP-NOL. The relationship between oxygen counts and MR ratio indicates that further oxygen reduction would realize an MR ratio greater than 50%.
Effects of experience and social context on prospective caching strategies by scrub jays.
Emery, N J; Clayton, N S
2001-11-22
Social life has costs associated with competition for resources such as food. Food storing may reduce this competition as the food can be collected quickly and hidden elsewhere; however, it is a risky strategy because caches can be pilfered by others. Scrub jays (Aphelocoma coerulescens) remember 'what', 'where' and 'when' they cached. Like other corvids, they remember where conspecifics have cached, pilfering them when given the opportunity, but may also adjust their own caching strategies to minimize potential pilfering. To test this, jays were allowed to cache either in private (when the other bird's view was obscured) or while a conspecific was watching, and then recover their caches in private. Here we show that jays with prior experience of pilfering another bird's caches subsequently re-cached food in new cache sites during recovery trials, but only when they had been observed caching. Jays without pilfering experience did not, even though they had observed other jays caching. Our results suggest that jays relate information about their previous experience as a pilferer to the possibility of future stealing by another bird, and modify their caching strategy accordingly.
Cache-Cache Comparison for Supporting Meaningful Learning
ERIC Educational Resources Information Center
Wang, Jingyun; Fujino, Seiji
2015-01-01
The paper presents a meaningful discovery learning environment called "cache-cache comparison" for a personalized learning support system. The processing of seeking hidden relations or concepts in "cache-cache comparison" is intended to encourage learners to actively locate new knowledge in their knowledge framework and check…
a Cache Design Method for Spatial Information Visualization in 3d Real-Time Rendering Engine
NASA Astrophysics Data System (ADS)
Dai, X.; Xiong, H.; Zheng, X.
2012-07-01
A well-designed cache system has positive impacts on the 3D real-time rendering engine. As the amount of visualization data getting larger, the effects become more obvious. They are the base of the 3D real-time rendering engine to smoothly browsing through the data, which is out of the core memory, or from the internet. In this article, a new kind of caches which are based on multi threads and large file are introduced. The memory cache consists of three parts, the rendering cache, the pre-rendering cache and the elimination cache. The rendering cache stores the data that is rendering in the engine; the data that is dispatched according to the position of the view point in the horizontal and vertical directions is stored in the pre-rendering cache; the data that is eliminated from the previous cache is stored in the eliminate cache and is going to write to the disk cache. Multi large files are used in the disk cache. When a disk cache file size reaches the limit length(128M is the top in the experiment), no item will be eliminated from the file, but a new large cache file will be created. If the large file number is greater than the maximum number that is pre-set, the earliest file will be deleted from the disk. In this way, only one file is opened for writing and reading, and the rest are read-only so the disk cache can be used in a high asynchronous way. The size of the large file is limited in order to map to the core memory to save loading time. Multi-thread is used to update the cache data. The threads are used to load data to the rendering cache as soon as possible for rendering, to load data to the pre-rendering cache for rendering next few frames, and to load data to the elimination cache which is not necessary for the moment. In our experiment, two threads are designed. The first thread is to organize the memory cache according to the view point, and created two threads: the adding list and the deleting list, the adding list index the data that should be loaded to the pre-rendering cache immediately, the deleting list index the data that is no longer visible in the rendering scene and should be moved to the eliminate cache; the other thread is to move the data in the memory and disk cache according to the adding and the deleting list, and create the download requests when the data is indexed in the adding but cannot be found either in memory cache or disk cache, eliminate cache data is moved to the disk cache when the adding list and deleting are empty. The cache designed as described above in our experiment shows reliable and efficient, and the data loading time and files I/O time decreased sharply, especially when the rendering data getting larger.
Optimizing Maintenance of Constraint-Based Database Caches
NASA Astrophysics Data System (ADS)
Klein, Joachim; Braun, Susanne
Caching data reduces user-perceived latency and often enhances availability in case of server crashes or network failures. DB caching aims at local processing of declarative queries in a DBMS-managed cache close to the application. Query evaluation must produce the same results as if done at the remote database backend, which implies that all data records needed to process such a query must be present and controlled by the cache, i. e., to achieve “predicate-specific” loading and unloading of such record sets. Hence, cache maintenance must be based on cache constraints such that “predicate completeness” of the caching units currently present can be guaranteed at any point in time. We explore how cache groups can be maintained to provide the data currently needed. Moreover, we design and optimize loading and unloading algorithms for sets of records keeping the caching units complete, before we empirically identify the costs involved in cache maintenance.
Ostojić, Ljerka; Legg, Edward W; Brecht, Katharina F; Lange, Florian; Deininger, Chantal; Mendl, Michael; Clayton, Nicola S
2017-01-23
Many corvid species accurately remember the locations where they have seen others cache food, allowing them to pilfer these caches efficiently once the cachers have left the scene [1]. To protect their caches, corvids employ a suite of different cache-protection strategies that limit the observers' visual or acoustic access to the cache site [2,3]. In cases where an observer's sensory access cannot be reduced it has been suggested that cachers might be able to minimise the risk of pilfering if they avoid caching food the observer is most motivated to pilfer [4]. In the wild, corvids have been reported to pilfer others' caches as soon as possible after the caching event [5], such that the cacher might benefit from adjusting its caching behaviour according to the observer's current desire. In the current study, observers pilfered according to their current desire: they preferentially pilfered food that they were not sated on. Cachers adjusted their caching behaviour accordingly: they protected their caches by selectively caching food that observers were not motivated to pilfer. The same cache-protection behaviour was found when cachers could not see on which food the observers were sated. Thus, the cachers' ability to respond to the observer's desire might have been driven by the observer's behaviour at the time of caching. Copyright © 2017 The Author(s). Published by Elsevier Ltd.. All rights reserved.
Pilfering Eurasian jays use visual and acoustic information to locate caches.
Shaw, Rachael C; Clayton, Nicola S
2014-11-01
Pilfering corvids use observational spatial memory to accurately locate caches that they have seen another individual make. Accordingly, many corvid cache-protection strategies limit the transfer of visual information to potential thieves. Eurasian jays (Garrulus glandarius) employ strategies that reduce the amount of visual and auditory information that is available to competitors. Here, we test whether or not the jays recall and use both visual and auditory information when pilfering other birds' caches. When jays had no visual or acoustic information about cache locations, the proportion of available caches that they found did not differ from the proportion expected if jays were searching at random. By contrast, after observing and listening to a conspecific caching in gravel or sand, jays located a greater proportion of caches, searched more frequently in the correct substrate type and searched in fewer empty locations to find the first cache than expected. After only listening to caching in gravel and sand, jays also found a larger proportion of caches and searched in the substrate type where they had heard caching take place more frequently than expected. These experiments demonstrate that Eurasian jays possess observational spatial memory and indicate that pilfering jays may gain information about cache location merely by listening to caching. This is the first evidence that a corvid may use recalled acoustic information to locate and pilfer caches.
The Effects of Cache Modification on Food Caching and Retrieval Behavior by Rats
ERIC Educational Resources Information Center
McKenzie, T.L.B.; Bird, L.R.; Roberts, W.A.
2005-01-01
Rats cached pieces of cheese on four different arms of an eight-arm radial maze. On a retrieval test given 45min later, rats learned to return to arms where food was cached before arms where food had not been cached. Tests were then performed in which cache sites on one side of the maze were always modified (pilfered or degraded), but cache sites…
Don’t make cache too complex: A simple probability-based cache management scheme for SSDs
Cho, Sangyeun; Choi, Jongmoo
2017-01-01
Solid-state drives (SSDs) have recently become a common storage component in computer systems, and they are fueled by continued bit cost reductions achieved with smaller feature sizes and multiple-level cell technologies. However, as the flash memory stores more bits per cell, the performance and reliability of the flash memory degrade substantially. To solve this problem, a fast non-volatile memory (NVM-)based cache has been employed within SSDs to reduce the long latency required to write data. Absorbing small writes in a fast NVM cache can also reduce the number of flash memory erase operations. To maximize the benefits of an NVM cache, it is important to increase the NVM cache utilization. In this paper, we propose and study ProCache, a simple NVM cache management scheme, that makes cache-entrance decisions based on random probability testing. Our scheme is motivated by the observation that frequently written hot data will eventually enter the cache with a high probability, and that infrequently accessed cold data will not enter the cache easily. Owing to its simplicity, ProCache is easy to implement at a substantially smaller cost than similar previously studied techniques. We evaluate ProCache and conclude that it achieves comparable performance compared to a more complex reference counter-based cache-management scheme. PMID:28358897
Don't make cache too complex: A simple probability-based cache management scheme for SSDs.
Baek, Seungjae; Cho, Sangyeun; Choi, Jongmoo
2017-01-01
Solid-state drives (SSDs) have recently become a common storage component in computer systems, and they are fueled by continued bit cost reductions achieved with smaller feature sizes and multiple-level cell technologies. However, as the flash memory stores more bits per cell, the performance and reliability of the flash memory degrade substantially. To solve this problem, a fast non-volatile memory (NVM-)based cache has been employed within SSDs to reduce the long latency required to write data. Absorbing small writes in a fast NVM cache can also reduce the number of flash memory erase operations. To maximize the benefits of an NVM cache, it is important to increase the NVM cache utilization. In this paper, we propose and study ProCache, a simple NVM cache management scheme, that makes cache-entrance decisions based on random probability testing. Our scheme is motivated by the observation that frequently written hot data will eventually enter the cache with a high probability, and that infrequently accessed cold data will not enter the cache easily. Owing to its simplicity, ProCache is easy to implement at a substantially smaller cost than similar previously studied techniques. We evaluate ProCache and conclude that it achieves comparable performance compared to a more complex reference counter-based cache-management scheme.
Checkpointing in speculative versioning caches
Eichenberger, Alexandre E; Gara, Alan; Gschwind, Michael K; Ohmacht, Martin
2013-08-27
Mechanisms for generating checkpoints in a speculative versioning cache of a data processing system are provided. The mechanisms execute code within the data processing system, wherein the code accesses cache lines in the speculative versioning cache. The mechanisms further determine whether a first condition occurs indicating a need to generate a checkpoint in the speculative versioning cache. The checkpoint is a speculative cache line which is made non-speculative in response to a second condition occurring that requires a roll-back of changes to a cache line corresponding to the speculative cache line. The mechanisms also generate the checkpoint in the speculative versioning cache in response to a determination that the first condition has occurred.
Rapid effects of corticosterone on cache recovery in mountain chickadees (Parus gambeli).
Saldanha, C J; Schlinger, B A; Clayton, N S
2000-03-01
Environmental perturbations increase adrenal activity in several vertebrates. Increases in corticosterone may serve as a proximate trigger whereby organisms can rapidly adapt their behavior to survive environmental fluctuations. In food-caching songbirds, inclement weather may present the need to alter caching and/or retrieval behaviors to ensure food supplies. We hypothesized that corticosterone may increase the rate of caching and/or retrieval behaviors in the mountain chickadee, a food-storing songbird, and tested if these potential effects were mediated by alterations in appetite, activity, or memory for cache sites. Corticosterone or vehicle was administered to subjects 5 min prior to either caching or recovery in a naturalistic laboratory paradigm during which we recorded the number of caching events, sites visited, and seeds eaten (caching) or caches recovered, total sites visited, cache-related visits, and non-cache-related visits (recovery). Data were analyzed using nested ANOVA for treatment within sequential trial. There was no effect on any caching behaviors following treatment. However, birds treated with corticosterone during retrieval recovered more seeds and tended to visit more cache-related sites than did controls. Since groups did not differ in the number of seeds eaten or the total number of sites visited, it seems unlikely that corticosterone affected appetite or activity. Rapid surges in corticosterone may increase the efficacy of an underlying memory process for cache sites which is reflected in higher cache recovery in corticosterone-treated birds than in controls. Thus, rapid alterations in plasma corticosterone following environmental change may alter memory-reliant behaviors which promote survival in the food-caching mountain chickadee. Copyright 2000 Academic Press.
Combining instruction prefetching with partial cache locking to improve WCET in real-time systems.
Ni, Fan; Long, Xiang; Wan, Han; Gao, Xiaopeng
2013-01-01
Caches play an important role in embedded systems to bridge the performance gap between fast processor and slow memory. And prefetching mechanisms are proposed to further improve the cache performance. While in real-time systems, the application of caches complicates the Worst-Case Execution Time (WCET) analysis due to its unpredictable behavior. Modern embedded processors often equip locking mechanism to improve timing predictability of the instruction cache. However, locking the whole cache may degrade the cache performance and increase the WCET of the real-time application. In this paper, we proposed an instruction-prefetching combined partial cache locking mechanism, which combines an instruction prefetching mechanism (termed as BBIP) with partial cache locking to improve the WCET estimates of real-time applications. BBIP is an instruction prefetching mechanism we have already proposed to improve the worst-case cache performance and in turn the worst-case execution time. The estimations on typical real-time applications show that the partial cache locking mechanism shows remarkable WCET improvement over static analysis and full cache locking.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Zhang, Zhao
With each CMOS technology generation, leakage energy consumption has been dramatically increasing and hence, managing leakage power consumption of large last-level caches (LLCs) has become a critical issue in modern processor design. In this paper, we present EnCache, a novel software-based technique which uses dynamic profiling-based cache reconfiguration for saving cache leakage energy. EnCache uses a simple hardware component called profiling cache, which dynamically predicts energy efficiency of an application for 32 possible cache configurations. Using these estimates, system software reconfigures the cache to the most energy efficient configuration. EnCache uses dynamic cache reconfiguration and hence, it does not requiremore » offline profiling or tuning the parameter for each application. Furthermore, EnCache optimizes directly for the overall memory subsystem (LLC and main memory) energy efficiency instead of the LLC energy efficiency alone. The experiments performed with an x86-64 simulator and workloads from SPEC2006 suite confirm that EnCache provides larger energy saving than a conventional energy saving scheme. For single core and dual-core system configurations, the average savings in memory subsystem energy over a shared baseline configuration are 30.0% and 27.3%, respectively.« less
Combining Instruction Prefetching with Partial Cache Locking to Improve WCET in Real-Time Systems
Ni, Fan; Long, Xiang; Wan, Han; Gao, Xiaopeng
2013-01-01
Caches play an important role in embedded systems to bridge the performance gap between fast processor and slow memory. And prefetching mechanisms are proposed to further improve the cache performance. While in real-time systems, the application of caches complicates the Worst-Case Execution Time (WCET) analysis due to its unpredictable behavior. Modern embedded processors often equip locking mechanism to improve timing predictability of the instruction cache. However, locking the whole cache may degrade the cache performance and increase the WCET of the real-time application. In this paper, we proposed an instruction-prefetching combined partial cache locking mechanism, which combines an instruction prefetching mechanism (termed as BBIP) with partial cache locking to improve the WCET estimates of real-time applications. BBIP is an instruction prefetching mechanism we have already proposed to improve the worst-case cache performance and in turn the worst-case execution time. The estimations on typical real-time applications show that the partial cache locking mechanism shows remarkable WCET improvement over static analysis and full cache locking. PMID:24386133
Way-Scaling to Reduce Power of Cache with Delay Variation
NASA Astrophysics Data System (ADS)
Goudarzi, Maziar; Matsumura, Tadayuki; Ishihara, Tohru
The share of leakage in cache power consumption increases with technology scaling. Choosing a higher threshold voltage (Vth) and/or gate-oxide thickness (Tox) for cache transistors improves leakage, but impacts cell delay. We show that due to uncorrelated random within-die delay variation, only some (not all) of cells actually violate the cache delay after the above change. We propose to add a spare cache way to replace delay-violating cache-lines separately in each cache-set. By SPICE and gate-level simulations in a commercial 90nm process, we show that choosing higher Vth, Tox and adding one spare way to a 4-way 16KB cache reduces leakage power by 42%, which depending on the share of leakage in total cache power, gives up to 22.59% and 41.37% reduction of total energy respectively in L1 instruction- and L2 unified-cache with a negligible delay penalty, but without sacrificing cache capacity or timing-yield.
Seedling Establishment of Coast Live Oak in Relation to Seed Caching by Jays
Joe R. McBride; Ed Norberg; Sheauchi Cheng; Ahmad Mossadegh
1991-01-01
The purpose of this study was to simulate the caching of acorns by jays and rodents to see if less costly procedures could be developed for the establishment of coast live oak (Quercus agrifolia). Four treatments [(1) random - single acorn cache, (2) regular - single acorn cache, (3) regular - 5 acorn cache, (4) regular - 10 acorn cache] were planted...
Kelley, Laura A; Clayton, Nicola S
2017-07-01
Some animals hide food to consume later; however, these caches are susceptible to theft by conspecifics and heterospecifics. Caching animals can use protective strategies to minimize sensory cues available to potential pilferers, such as caching in shaded areas and in quiet substrate. Background matching (where object patterning matches the visual background) is commonly seen in prey animals to reduce conspicuousness, and caching animals may also use this tactic to hide caches, for example, by hiding coloured food in a similar coloured substrate. We tested whether California scrub-jays ( Aphelocoma californica ) camouflage their food in this way by offering them caching substrates that either matched or did not match the colour of food available for caching. We also determined whether this caching behaviour was sensitive to social context by allowing the birds to cache when a conspecific potential pilferer could be both heard and seen (acoustic and visual cues present), or unseen (acoustic cues only). When caching events could be both heard and seen by a potential pilferer, birds cached randomly in matching and non-matching substrates. However, they preferentially hid food in the substrate that matched the food colour when only acoustic cues were present. This is a novel cache protection strategy that also appears to be sensitive to social context. We conclude that studies of cache protection strategies should consider the perceptual capabilities of the cacher and potential pilferers. © 2017 The Author(s).
Consolidation and reconsolidation of memory in black-capped chickadees (Poecile atricapillus).
Barrett, Matthew C; Sherry, David F
2012-12-01
Multiple phases of protein synthesis are necessary for the synaptic modifications that consolidate long-term memory. The reconsolidation hypothesis supposes that information in long-term memory becomes labile and subject to change when retrieved and must be reconsolidated into long-term memory. The current study used the protein synthesis inhibitor anisomycin to examine memory consolidation in birds and to test the reconsolidation hypothesis. Black-capped chickadees store food and usually remember which of their caches they have emptied and which they have left full. In Experiment 1, anisomycin was injected either immediately and 2 hr after food caching, or 4 and 6 hr after food caching. Inhibition of protein synthesis impaired memory for cache sites 24 and 48 hr later. In Experiment 2, it was hypothesized that long-term memory for food caches becomes labile as predicted by the reconsolidation hypothesis when birds search for caches. Anisomycin was administered immediately after chickadees had searched for their caches. Inhibition of protein synthesis should disrupt memory for caches left full if these sites are retrieved from long-term memory and require reconsolidation. Control birds were later more likely to revisit full caches than caches they had emptied. Birds given anisomycin revisited both kinds of caches and did not distinguish between them. This result shows that reconsolidation of full caches into long-term memory is not necessary following search for cache sites, but also shows that protein synthesis-dependent consolidation is required for updating the status of emptied caches.
Memory-Intensive Benchmarks: IRAM vs. Cache-Based Machines
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Gaeke, Brian R.; Husbands, Parry; Li, Xiaoye S.; Oliker, Leonid; Yelick, Katherine A.; Biegel, Bryan (Technical Monitor)
2002-01-01
The increasing gap between processor and memory performance has lead to new architectural models for memory-intensive applications. In this paper, we explore the performance of a set of memory-intensive benchmarks and use them to compare the performance of conventional cache-based microprocessors to a mixed logic and DRAM processor called VIRAM. The benchmarks are based on problem statements, rather than specific implementations, and in each case we explore the fundamental hardware requirements of the problem, as well as alternative algorithms and data structures that can help expose fine-grained parallelism or simplify memory access patterns. The benchmarks are characterized by their memory access patterns, their basic control structures, and the ratio of computation to memory operation.
Mobility-Aware Caching and Computation Offloading in 5G Ultra-Dense Cellular Networks
Chen, Min; Hao, Yixue; Qiu, Meikang; Song, Jeungeun; Wu, Di; Humar, Iztok
2016-01-01
Recent trends show that Internet traffic is increasingly dominated by content, which is accompanied by the exponential growth of traffic. To cope with this phenomena, network caching is introduced to utilize the storage capacity of diverse network devices. In this paper, we first summarize four basic caching placement strategies, i.e., local caching, Device-to-Device (D2D) caching, Small cell Base Station (SBS) caching and Macrocell Base Station (MBS) caching. However, studies show that so far, much of the research has ignored the impact of user mobility. Therefore, taking the effect of the user mobility into consideration, we proposes a joint mobility-aware caching and SBS density placement scheme (MS caching). In addition, differences and relationships between caching and computation offloading are discussed. We present a design of a hybrid computation offloading and support it with experimental results, which demonstrate improved performance in terms of energy cost. Finally, we discuss the design of an incentive mechanism by considering network dynamics, differentiated user’s quality of experience (QoE) and the heterogeneity of mobile terminals in terms of caching and computing capabilities. PMID:27347975
Mobility-Aware Caching and Computation Offloading in 5G Ultra-Dense Cellular Networks.
Chen, Min; Hao, Yixue; Qiu, Meikang; Song, Jeungeun; Wu, Di; Humar, Iztok
2016-06-25
Recent trends show that Internet traffic is increasingly dominated by content, which is accompanied by the exponential growth of traffic. To cope with this phenomena, network caching is introduced to utilize the storage capacity of diverse network devices. In this paper, we first summarize four basic caching placement strategies, i.e., local caching, Device-to-Device (D2D) caching, Small cell Base Station (SBS) caching and Macrocell Base Station (MBS) caching. However, studies show that so far, much of the research has ignored the impact of user mobility. Therefore, taking the effect of the user mobility into consideration, we proposes a joint mobility-aware caching and SBS density placement scheme (MS caching). In addition, differences and relationships between caching and computation offloading are discussed. We present a design of a hybrid computation offloading and support it with experimental results, which demonstrate improved performance in terms of energy cost. Finally, we discuss the design of an incentive mechanism by considering network dynamics, differentiated user's quality of experience (QoE) and the heterogeneity of mobile terminals in terms of caching and computing capabilities.
Behavior-aware cache hierarchy optimization for low-power multi-core embedded systems
NASA Astrophysics Data System (ADS)
Zhao, Huatao; Luo, Xiao; Zhu, Chen; Watanabe, Takahiro; Zhu, Tianbo
2017-07-01
In modern embedded systems, the increasing number of cores requires efficient cache hierarchies to ensure data throughput, but such cache hierarchies are restricted by their tumid size and interference accesses which leads to both performance degradation and wasted energy. In this paper, we firstly propose a behavior-aware cache hierarchy (BACH) which can optimally allocate the multi-level cache resources to many cores and highly improved the efficiency of cache hierarchy, resulting in low energy consumption. The BACH takes full advantage of the explored application behaviors and runtime cache resource demands as the cache allocation bases, so that we can optimally configure the cache hierarchy to meet the runtime demand. The BACH was implemented on the GEM5 simulator. The experimental results show that energy consumption of a three-level cache hierarchy can be saved from 5.29% up to 27.94% compared with other key approaches while the performance of the multi-core system even has a slight improvement counting in hardware overhead.
A cache-aided multiprocessor rollback recovery scheme
NASA Technical Reports Server (NTRS)
Wu, Kun-Lung; Fuchs, W. Kent
1989-01-01
This paper demonstrates how previous uniprocessor cache-aided recovery schemes can be applied to multiprocessor architectures, for recovering from transient processor failures, utilizing private caches and a global shared memory. As with cache-aided uniprocessor recovery, the multiprocessor cache-aided recovery scheme of this paper can be easily integrated into standard bus-based snoopy cache coherence protocols. A consistent shared memory state is maintained without the necessity of global check-pointing.
NASA Astrophysics Data System (ADS)
Gardner, R. W.; Hanushevsky, A.; Vukotic, I.; Yang, W.
2017-10-01
As many LHC Tier-3 and some Tier-2 centers look toward streamlining operations, they are considering autonomously managed storage elements as part of the solution. These storage elements are essentially file caching servers. They can operate as whole file or data block level caches. Several implementations exist. In this paper we explore using XRootD caching servers that can operate in either mode. They can also operate autonomously (i.e. demand driven), be centrally managed (i.e. a Rucio managed cache), or operate in both modes. We explore the pros and cons of various configurations as well as practical requirements for caching to be effective. While we focus on XRootD caches, the analysis should apply to other kinds of caches as well.
Gara, Alan; Ohmacht, Martin
2014-09-16
In a multiprocessor system with at least two levels of cache, a speculative thread may run on a core processor in parallel with other threads. When the thread seeks to do a write to main memory, this access is to be written through the first level cache to the second level cache. After the write though, the corresponding line is deleted from the first level cache and/or prefetch unit, so that any further accesses to the same location in main memory have to be retrieved from the second level cache. The second level cache keeps track of multiple versions of data, where more than one speculative thread is running in parallel, while the first level cache does not have any of the versions during speculation. A switch allows choosing between modes of operation of a speculation blind first level cache.
Re-caching by Western scrub-jays (Aphelocoma californica) cannot be attributed to stress.
Thom, James M; Clayton, Nicola S
2013-01-01
Western scrub-jays (Aphelocoma californica) live double lives, storing food for the future while raiding the stores of other birds. One tactic scrub-jays employ to protect stores is "re-caching"-relocating caches out of sight of would-be thieves. Recent computational modelling work suggests that re-caching might be mediated not by complex cognition, but by a combination of memory failure and stress. The "Stress Model" asserts that re-caching is a manifestation of a general drive to cache, rather than a desire to protect existing stores. Here, we present evidence strongly contradicting the central assumption of these models: that stress drives caching, irrespective of social context. In Experiment (i), we replicate the finding that scrub-jays preferentially relocate food they were watched hiding. In Experiment (ii) we find no evidence that stress increases caching. In light of our results, we argue that the Stress Model cannot account for scrub-jay re-caching.
Jakopak, Rhiannon P.; Hall, L. Embere; Chalfoun, Anna D.
2017-01-01
Many mammals create food stores to enhance overwinter survival in seasonal environments. Strategic arrangement of food within caches may facilitate the physical integrity of the cache or improve access to high-quality food to ensure that cached resources meet future nutritional demands. We used the American pika (Ochotona princeps), a food-caching lagomorph, to evaluate variation in haypile (cache) structure (i.e., horizontal layering by plant functional group) in Wyoming, United States. Fifty-five percent of 62 haypiles contained at least 2 discrete layers of vegetation. Adults and juveniles layered haypiles in similar proportions. The probability of layering increased with haypile volume, but not haypile number per individual or nearby forage diversity. Vegetation cached in layered haypiles was also higher in nitrogen compared to vegetation in unlayered piles. We found that American pikas frequently structured their food caches, structured caches were larger, and the cached vegetation in structured piles was of higher nutritional quality. Improving access to stable, high-quality vegetation in haypiles, a critical overwinter food resource, may allow individuals to better persist amidst harsh conditions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Zhang, Zhao; Vetter, Jeffrey S
Recent trends of CMOS scaling and use of large last level caches (LLCs) have led to significant increase in the leakage energy consumption of LLCs and hence, managing their energy consumption has become extremely important in modern processor design. The conventional cache energy saving techniques require offline profiling or provide only coarse granularity of cache allocation. We present FlexiWay, a cache energy saving technique which uses dynamic cache reconfiguration. FlexiWay logically divides the cache sets into multiple (e.g. 16) modules and dynamically turns off suitable and possibly different number of cache ways in each module. FlexiWay has very small implementationmore » overhead and it provides fine-grain cache allocation even with caches of typical associativity, e.g. an 8-way cache. Microarchitectural simulations have been performed using an x86-64 simulator and workloads from SPEC2006 suite. Also, FlexiWay has been compared with two conventional energy saving techniques. The results show that FlexiWay provides largest energy saving and incurs only small loss in performance. For single, dual and quad core systems, the average energy saving using FlexiWay are 26.2%, 25.7% and 22.4%, respectively.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Millar, A. P.; Baranova, T.; Behrmann, G.
For over a decade, dCache has been synonymous with large-capacity, fault-tolerant storage using commodity hardware that supports seamless data migration to and from tape. In this paper we provide some recent news of changes within dCache and the community surrounding it. We describe the flexible nature of dCache that allows both externally developed enhancements to dCache facilities and the adoption of new technologies. Finally, we present information about avenues the dCache team is exploring for possible future improvements in dCache.
Smart caching based on mobile agent of power WebGIS platform.
Wang, Xiaohui; Wu, Kehe; Chen, Fei
2013-01-01
Power information construction is developing towards intensive, platform, distributed direction with the expansion of power grid and improvement of information technology. In order to meet the trend, power WebGIS was designed and developed. In this paper, we first discuss the architecture and functionality of power WebGIS, and then we study caching technology in detail, which contains dynamic display cache model, caching structure based on mobile agent, and cache data model. We have designed experiments of different data capacity to contrast performance between WebGIS with the proposed caching model and traditional WebGIS. The experimental results showed that, with the same hardware environment, the response time of WebGIS with and without caching model increased as data capacity growing, while the larger the data was, the higher the performance of WebGIS with proposed caching model improved.
Elements of episodic-like memory in animals.
Clayton, N S; Griffiths, D P; Emery, N J; Dickinson, A
2001-09-29
A number of psychologists have suggested that episodic memory is a uniquely human phenomenon and, until recently, there was little evidence that animals could recall a unique past experience and respond appropriately. Experiments on food-caching memory in scrub jays question this assumption. On the basis of a single caching episode, scrub jays can remember when and where they cached a variety of foods that differ in the rate at which they degrade, in a way that is inexplicable by relative familiarity. They can update their memory of the contents of a cache depending on whether or not they have emptied the cache site, and can also remember where another bird has hidden caches, suggesting that they encode rich representations of the caching event. They make temporal generalizations about when perishable items should degrade and also remember the relative time since caching when the same food is cached in distinct sites at different times. These results show that jays form integrated memories for the location, content and time of caching. This memory capability fulfils Tulving's behavioural criteria for episodic memory and is thus termed 'episodic-like'. We suggest that several features of episodic memory may not be unique to humans.
Pravosudov, V V; Clayton, N S
2001-02-22
Birds rely, at least in part, on spatial memory for recovering previously hidden caches but accurate cache recovery may be more critical for birds that forage in harsh conditions where the food supply is limited and unpredictable. Failure to find caches in these conditions may potentially result in death from starvation. In order to test this hypothesis we compared the cache recovery behaviour of 24 wild-caught mountain chickadees (Poecile gambeli), half of which were maintained on a limited and unpredictable food supply while the rest were maintained on an ad libitum food supply for 60 days. We then tested their cache retrieval accuracy by allowing birds from both groups to cache seeds in the experimental room and recover them 5 hours later. Our results showed that birds maintained on a limited and unpredictable food supply made significantly fewer visits to non-cache sites when recovering their caches compared to birds maintained on ad libitum food. We found the same difference in performance in two versions of a one-trial associative learning task in which the birds had to rely on memory to find previously encountered hidden food. In a non-spatial memory version of the task, in which the baited feeder was clearly marked, there were no significant differences between the two groups. We therefore concluded that the two groups differed in their efficiency at cache retrieval. We suggest that this difference is more likely to be attributable to a difference in memory (encoding or recall) than to a difference in their motivation to search for hidden food, although the possibility of some motivational differences still exists. Overall, our results suggest that demanding foraging conditions favour more accurate cache retrieval in food-caching birds.
Determinants of seed removal distance by scatter-hoarding rodents in deciduous forests.
Moore, Jeffrey E; McEuen, Amy B; Swihart, Robert K; Contreras, Thomas A; Steele, Michael A
2007-10-01
Scatter-hoarding rodents should space food caches to maximize cache recovery rate (to minimize loss to pilferers) relative to the energetic cost of carrying food items greater distances. Optimization models of cache spacing make two predictions. First, spacing of caches should be greater for food items with greater energy content. Second, the mean distance between caches should increase with food abundance. However, the latter prediction fails to account for the effect of food abundance on the behavior of potential pilferers or on the ability of caching individuals to acquire food by means other than recovering their own caches. When considering these factors, shorter cache distances may be predicted in conditions of higher food abundance. We predicted that seed caching distances would be greater for food items of higher energy content and during lower ambient food abundance and that the effect of seed type on cache distance variation would be lower during higher food abundance. We recorded distances moved for 8636 seeds of five seed types at 15 locations in three forested sites in Pennsylvania, USA, and 29 forest fragments in Indiana, U.S.A., across five different years. Seed production was poor in three years and high in two years. Consistent with previous studies, seeds with greater energy content were moved farther than less profitable food items. Seeds were dispersed less far in seed-rich years than in seed-poor years, contrary to predictions of conventional models. Interactions were important, with seed type effects more evident in seed-poor years. These results suggest that, when food is superabundant, optimal cache distances are more strongly determined by minimizing energy cost of caching than by minimizing pilfering rates and that cache loss rates may be more strongly density-dependent in times of low seed abundance.
Addressing Inter-set Write-Variation for Improving Lifetime of Non-Volatile Caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Vetter, Jeffrey S
We propose a technique which minimizes inter-set write variation in NVM caches for improving its lifetime. Our technique uses cache coloring scheme to add a software-controlled mapping layer between groups of physical pages (called memory regions) and cache sets. Periodically, the number of writes to different colors of the cache is computed and based on this result, the mapping of a few colors is changed to channel the write traffic to least utilized cache colors. This change helps to achieve wear-leveling.
Use of diuretics is associated with reduced risk of Alzheimer's disease: the Cache County Study.
Chuang, Yi-Fang; Breitner, John C S; Chiu, Yen-Ling; Khachaturian, Ara; Hayden, Kathleen; Corcoran, Chris; Tschanz, JoAnn; Norton, Maria; Munger, Ron; Welsh-Bohmer, Kathleen; Zandi, Peter P
2014-11-01
Although the use of antihypertensive medications has been associated with reduced risk of Alzheimer's disease (AD), it remains unclear which class provides the most benefit. The Cache County Study of Memory Health and Aging is a prospective longitudinal cohort study of dementing illnesses among the elderly population of Cache County, Utah. Using waves I to IV data of the Cache County Study, 3417 participants had a mean of 7.1 years of follow-up. Time-varying use of antihypertensive medications including different class of diuretics, angiotensin converting enzyme inhibitors, β-blockers, and calcium channel blockers was used to predict the incidence of AD using Cox proportional hazards analyses. During follow-up, 325 AD cases were ascertained with a total of 23,590 person-years. Use of any antihypertensive medication was associated with lower incidence of AD (adjusted hazard ratio [aHR], 0.77; 95% confidence interval [CI], 0.61-0.97). Among different classes of antihypertensive medications, thiazide (aHR, 0.7; 95% CI, 0.53-0.93), and potassium-sparing diuretics (aHR, 0.69; 95% CI, 0.48-0.99) were associated with the greatest reduction of AD risk. Thiazide and potassium-sparing diuretics were associated with decreased risk of AD. The inverse association of potassium-sparing diuretics confirms an earlier finding in this cohort, now with longer follow-up, and merits further investigation. Copyright © 2014 Elsevier Inc. All rights reserved.
Use of diuretics is associated with reduced risk of Alzheimer’s disease: the Cache County Study
Chuang, Yi-Fang; Breitner, John C.S.; Chiu, Yen-Ling; Khachaturian, Ara; Hayden, Kathleen; Corcoran, Chris; Tschanz, JoAnn; Norton, Maria; Munger, Ron; Welsh-Bohmer, Kathleen; Zandi, Peter P.
2015-01-01
Although the use of antihypertensive medications has been associated with reduced risk of Alzheimer’s disease (AD), it remains unclear which class provides the most benefit. The Cache County Study of Memory Health and Aging is a prospective longitudinal cohort study of dementing illnesses among the elderly population of Cache County, Utah. Using waves I to IV data of the Cache County Study, 3417 participants had a mean of 7.1 years of follow-up. Time-varying use of antihypertensive medications including different class of diuretics, angiotensin converting enzyme inhibitors, β-blockers, and calcium channel blockers was used to predict the incidence of AD using Cox proportional hazards analyses. During follow-up, 325 AD cases were ascertained with a total of 23,590 person-years. Use of any anti-hypertensive medication was associated with lower incidence of AD (adjusted hazard ratio [aHR], 0.77; 95% confidence interval [CI], 0.61–0.97). Among different classes of antihypertensive medications, thiazide (aHR, 0.7; 95% CI, 0.53–0.93), and potassium-sparing diuretics (aHR, 0.69; 95% CI, 0.48–0.99) were associated with the greatest reduction of AD risk. Thiazide and potassium-sparing diuretics were associated with decreased risk of AD. The inverse association of potassium-sparing diuretics confirms an earlier finding in this cohort, now with longer follow-up, and merits further investigation. PMID:24910391
NASA Astrophysics Data System (ADS)
Steele, Michael A.; Bugdal, Melissa; Yuan, Amy; Bartlow, Andrew; Buzalewski, Jarrod; Lichti, Nathan; Swihart, Robert
2011-11-01
Scatter-hoarding mammals are thought to rely on spatial memory to relocate food caches. Yet, we know little about how long these granivores (primarily rodents) recall specific cache locations or whether individual hoarders have an advantage when recovering their own caches. Indeed, a few recent studies suggest that high rates of pilferage are common and that individual hoarders may not have a retriever's advantage. We tested this hypothesis in a high-density (>7 animals/ha) population of eastern gray squirrels ( Sciurus carolinensis) by presenting individually marked animals (>20) with tagged acorns, mapping cache sites, and following the fate of seed caches. PIT tags allowed us to monitor individual seeds without disturbing cache sites. Acorns only remained in the caches for 12-119 h (0.5-5 d). However, when we live-trapped and removed some animals from the site immediately after they stored seeds (thus simulating predation), their seed caches remained intact for significantly longer periods (16-27 d). Cache duration corresponded roughly to the time at which squirrels were returned to the study area. These results suggest that squirrels have a retriever's advantage and may remember specific cache sites longer than previously thought. We further suggest that predation of scatter hoarders who store seeds for long periods and also possess a recovery advantage may be one important mechanism by which seed establishment is achieved.
A Survey of Architectural Techniques For Improving Cache Power Efficiency
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh
Modern processors are using increasingly larger sized on-chip caches. Also, with each CMOS technology generation, there has been a significant increase in their leakage energy consumption. For this reason, cache power management has become a crucial research issue in modern processor design. To address this challenge and also meet the goals of sustainable computing, researchers have proposed several techniques for improving energy efficiency of cache architectures. This paper surveys recent architectural techniques for improving cache power efficiency and also presents a classification of these techniques based on their characteristics. For providing an application perspective, this paper also reviews several real-worldmore » processor chips that employ cache energy saving techniques. The aim of this survey is to enable engineers and researchers to get insights into the techniques for improving cache power efficiency and motivate them to invent novel solutions for enabling low-power operation of caches.« less
Cache-Aware Asymptotically-Optimal Sampling-Based Motion Planning
Ichnowski, Jeffrey; Prins, Jan F.; Alterovitz, Ron
2014-01-01
We present CARRT* (Cache-Aware Rapidly Exploring Random Tree*), an asymptotically optimal sampling-based motion planner that significantly reduces motion planning computation time by effectively utilizing the cache memory hierarchy of modern central processing units (CPUs). CARRT* can account for the CPU’s cache size in a manner that keeps its working dataset in the cache. The motion planner progressively subdivides the robot’s configuration space into smaller regions as the number of configuration samples rises. By focusing configuration exploration in a region for periods of time, nearest neighbor searching is accelerated since the working dataset is small enough to fit in the cache. CARRT* also rewires the motion planning graph in a manner that complements the cache-aware subdivision strategy to more quickly refine the motion planning graph toward optimality. We demonstrate the performance benefit of our cache-aware motion planning approach for scenarios involving a point robot as well as the Rethink Robotics Baxter robot. PMID:25419474
Smart Caching Based on Mobile Agent of Power WebGIS Platform
Wang, Xiaohui; Wu, Kehe; Chen, Fei
2013-01-01
Power information construction is developing towards intensive, platform, distributed direction with the expansion of power grid and improvement of information technology. In order to meet the trend, power WebGIS was designed and developed. In this paper, we first discuss the architecture and functionality of power WebGIS, and then we study caching technology in detail, which contains dynamic display cache model, caching structure based on mobile agent, and cache data model. We have designed experiments of different data capacity to contrast performance between WebGIS with the proposed caching model and traditional WebGIS. The experimental results showed that, with the same hardware environment, the response time of WebGIS with and without caching model increased as data capacity growing, while the larger the data was, the higher the performance of WebGIS with proposed caching model improved. PMID:24288504
Cache-Aware Asymptotically-Optimal Sampling-Based Motion Planning.
Ichnowski, Jeffrey; Prins, Jan F; Alterovitz, Ron
2014-05-01
We present CARRT* (Cache-Aware Rapidly Exploring Random Tree*), an asymptotically optimal sampling-based motion planner that significantly reduces motion planning computation time by effectively utilizing the cache memory hierarchy of modern central processing units (CPUs). CARRT* can account for the CPU's cache size in a manner that keeps its working dataset in the cache. The motion planner progressively subdivides the robot's configuration space into smaller regions as the number of configuration samples rises. By focusing configuration exploration in a region for periods of time, nearest neighbor searching is accelerated since the working dataset is small enough to fit in the cache. CARRT* also rewires the motion planning graph in a manner that complements the cache-aware subdivision strategy to more quickly refine the motion planning graph toward optimality. We demonstrate the performance benefit of our cache-aware motion planning approach for scenarios involving a point robot as well as the Rethink Robotics Baxter robot.
An Effective Cache Algorithm for Heterogeneous Storage Systems
Li, Yong; Feng, Dan
2013-01-01
Modern storage environment is commonly composed of heterogeneous storage devices. However, traditional cache algorithms exhibit performance degradation in heterogeneous storage systems because they were not designed to work with the diverse performance characteristics. In this paper, we present a new cache algorithm called HCM for heterogeneous storage systems. The HCM algorithm partitions the cache among the disks and adopts an effective scheme to balance the work across the disks. Furthermore, it applies benefit-cost analysis to choose the best allocation of cache block to improve the performance. Conducting simulations with a variety of traces and a wide range of cache size, our experiments show that HCM significantly outperforms the existing state-of-the-art storage-aware cache algorithms. PMID:24453890
A Measurement and Simulation Based Methodology for Cache Performance Modeling and Tuning
NASA Technical Reports Server (NTRS)
Waheed, Abdul; Yan, Jerry; Saini, Subhash (Technical Monitor)
1998-01-01
We present a cache performance modeling methodology that facilitates the tuning of uniprocessor cache performance for applications executing on shared memory multiprocessors by accurately predicting the effects of source code level modifications. Measurements on a single processor are initially used for identifying parts of code where cache utilization improvements may significantly impact the overall performance. Cache simulation based on trace-driven techniques can be carried out without gathering detailed address traces. Minimal runtime information for modeling cache performance of a selected code block includes: base virtual addresses of arrays, virtual addresses of variables, and loop bounds for that code block. Rest of the information is obtained from the source code. We show that the cache performance predictions are as reliable as those obtained through trace-driven simulations. This technique is particularly helpful to the exploration of various "what-if' scenarios regarding the cache performance impact for alternative code structures. We explain and validate this methodology using a simple matrix-matrix multiplication program. We then apply this methodology to predict and tune the cache performance of two realistic scientific applications taken from the Computational Fluid Dynamics (CFD) domain.
Pravosudov, Vladimir V
2003-12-22
It is widely assumed that chronic stress and corresponding chronic elevations of glucocorticoid levels have deleterious effects on animals' brain functions such as learning and memory. Some animals, however, appear to maintain moderately elevated levels of glucocorticoids over long periods of time under natural energetically demanding conditions, and it is not clear whether such chronic but moderate elevations may be adaptive. I implanted wild-caught food-caching mountain chickadees (Poecile gambeli), which rely at least in part on spatial memory to find their caches, with 90-day continuous time-release corticosterone pellets designed to approximately double the baseline corticosterone levels. Corticosterone-implanted birds cached and consumed significantly more food and showed more efficient cache recovery and superior spatial memory performance compared with placebo-implanted birds. Thus, contrary to prevailing assumptions, long-term moderate elevations of corticosterone appear to enhance spatial memory in food-caching mountain chickadees. These results suggest that moderate chronic elevation of corticosterone may serve as an adaptation to unpredictable environments by facilitating feeding and food-caching behaviour and by improving cache-retrieval efficiency in food-caching birds.
Cache Scheme Based on Pre-Fetch Operation in ICN
Duan, Jie; Wang, Xiong; Xu, Shizhong; Liu, Yuanni; Xu, Chuan; Zhao, Guofeng
2016-01-01
Many recent researches focus on ICN (Information-Centric Network), in which named content becomes the first citizen instead of end-host. In ICN, Named content can be further divided into many small sized chunks, and chunk-based communication has merits over content-based communication. The universal in-network cache is one of the fundamental infrastructures for ICN. In this work, a chunk-level cache mechanism based on pre-fetch operation is proposed. The main idea is that, routers with cache store should pre-fetch and cache the next chunks which may be accessed in the near future according to received requests and cache policy for reducing the users’ perceived latency. Two pre-fetch driven modes are present to answer when and how to pre-fetch. The LRU (Least Recently Used) is employed for the cache replacement. Simulation results show that the average user perceived latency and hops can be decreased by employed this cache mechanism based on pre-fetch operation. Furthermore, we also demonstrate that the results are influenced by many factors, such as the cache capacity, Zipf parameters and pre-fetch window size. PMID:27362478
Optoelectronic-cache memory system architecture.
Chiarulli, D M; Levitan, S P
1996-05-10
We present an investigation of the architecture of an optoelectronic cache that can integrate terabit optical memories with the electronic caches associated with high-performance uniprocessors and multiprocessors. The use of optoelectronic-cache memories enables these terabit technologies to provide transparently low-latency secondary memory with frame sizes comparable with disk pages but with latencies that approach those of electronic secondary-cache memories. This enables the implementation of terabit memories with effective access times comparable with the cycle times of current microprocessors. The cache design is based on the use of a smart-pixel array and combines parallel free-space optical input-output to-and-from optical memory with conventional electronic communication to the processor caches. This cache and the optical memory system to which it will interface provide a large random-access memory space that has a lower overall latency than that of magnetic disks and disk arrays. In addition, as a consequence of the high-bandwidth parallel input-output capabilities of optical memories, fault service times for the optoelectronic cache are substantially less than those currently achievable with any rotational media.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Boyle, Peter; Christ, Norman; Gara, Alan
A list prefetch engine improves a performance of a parallel computing system. The list prefetch engine receives a current cache miss address. The list prefetch engine evaluates whether the current cache miss address is valid. If the current cache miss address is valid, the list prefetch engine compares the current cache miss address and a list address. A list address represents an address in a list. A list describes an arbitrary sequence of prior cache miss addresses. The prefetch engine prefetches data according to the list, if there is a match between the current cache miss address and the listmore » address.« less
Boyle, Peter [Edinburgh, GB; Christ, Norman [Irvington, NY; Gara, Alan [Yorktown Heights, NY; Kim,; Changhoan, [San Jose, CA; Mawhinney, Robert [New York, NY; Ohmacht, Martin [Yorktown Heights, NY; Sugavanam, Krishnan [Yorktown Heights, NY
2012-08-28
A list prefetch engine improves a performance of a parallel computing system. The list prefetch engine receives a current cache miss address. The list prefetch engine evaluates whether the current cache miss address is valid. If the current cache miss address is valid, the list prefetch engine compares the current cache miss address and a list address. A list address represents an address in a list. A list describes an arbitrary sequence of prior cache miss addresses. The prefetch engine prefetches data according to the list, if there is a match between the current cache miss address and the list address.
Analysis of DNS Cache Effects on Query Distribution
2013-01-01
This paper studies the DNS cache effects that occur on query distribution at the CN top-level domain (TLD) server. We first filter out the malformed DNS queries to purify the log data pollution according to six categories. A model for DNS resolution, more specifically DNS caching, is presented. We demonstrate the presence and magnitude of DNS cache effects and the cache sharing effects on the request distribution through analytic model and simulation. CN TLD log data results are provided and analyzed based on the cache model. The approximate TTL distribution for domain name is inferred quantificationally. PMID:24396313
Analysis of DNS cache effects on query distribution.
Wang, Zheng
2013-01-01
This paper studies the DNS cache effects that occur on query distribution at the CN top-level domain (TLD) server. We first filter out the malformed DNS queries to purify the log data pollution according to six categories. A model for DNS resolution, more specifically DNS caching, is presented. We demonstrate the presence and magnitude of DNS cache effects and the cache sharing effects on the request distribution through analytic model and simulation. CN TLD log data results are provided and analyzed based on the cache model. The approximate TTL distribution for domain name is inferred quantificationally.
An Adaptive Insertion and Promotion Policy for Partitioned Shared Caches
NASA Astrophysics Data System (ADS)
Mahrom, Norfadila; Liebelt, Michael; Raof, Rafikha Aliana A.; Daud, Shuhaizar; Hafizah Ghazali, Nur
2018-03-01
Cache replacement policies in chip multiprocessors (CMP) have been investigated extensively and proven able to enhance shared cache management. However, competition among multiple processors executing different threads that require simultaneous access to a shared memory may cause cache contention and memory coherence problems on the chip. These issues also exist due to some drawbacks of the commonly used Least Recently Used (LRU) policy employed in multiprocessor systems, which are because of the cache lines residing in the cache longer than required. In image processing analysis of for example extra pulmonary tuberculosis (TB), an accurate diagnosis for tissue specimen is required. Therefore, a fast and reliable shared memory management system to execute algorithms for processing vast amount of specimen image is needed. In this paper, the effects of the cache replacement policy in a partitioned shared cache are investigated. The goal is to quantify whether better performance can be achieved by using less complex replacement strategies. This paper proposes a Middle Insertion 2 Positions Promotion (MI2PP) policy to eliminate cache misses that could adversely affect the access patterns and the throughput of the processors in the system. The policy employs a static predefined insertion point, near distance promotion, and the concept of ownership in the eviction policy to effectively improve cache thrashing and to avoid resource stealing among the processors.
Cache Energy Optimization Techniques For Modern Processors
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh
2013-01-01
Modern multicore processors are employing large last-level caches, for example Intel's E7-8800 processor uses 24MB L3 cache. Further, with each CMOS technology generation, leakage energy has been dramatically increasing and hence, leakage energy is expected to become a major source of energy dissipation, especially in last-level caches (LLCs). The conventional schemes of cache energy saving either aim at saving dynamic energy or are based on properties specific to first-level caches, and thus these schemes have limited utility for last-level caches. Further, several other techniques require offline profiling or per-application tuning and hence are not suitable for product systems. In thismore » book, we present novel cache leakage energy saving schemes for single-core and multicore systems; desktop, QoS, real-time and server systems. Also, we present cache energy saving techniques for caches designed with both conventional SRAM devices and emerging non-volatile devices such as STT-RAM (spin-torque transfer RAM). We present software-controlled, hardware-assisted techniques which use dynamic cache reconfiguration to configure the cache to the most energy efficient configuration while keeping the performance loss bounded. To profile and test a large number of potential configurations, we utilize low-overhead, micro-architecture components, which can be easily integrated into modern processor chips. We adopt a system-wide approach to save energy to ensure that cache reconfiguration does not increase energy consumption of other components of the processor. We have compared our techniques with state-of-the-art techniques and have found that our techniques outperform them in terms of energy efficiency and other relevant metrics. The techniques presented in this book have important applications in improving energy-efficiency of higher-end embedded, desktop, QoS, real-time, server processors and multitasking systems. This book is intended to be a valuable guide for both newcomers and veterans in the field of cache power management. It will help graduate students, CAD tool developers and designers in understanding the need of energy efficiency in modern computing systems. Further, it will be useful for researchers in gaining insights into algorithms and techniques for micro-architectural and system-level energy optimization using dynamic cache reconfiguration. We sincerely believe that the ``food for thought'' presented in this book will inspire the readers to develop even better ideas for designing ``green'' processors of tomorrow.« less
Donovan, Patrick M; Blum, Joel D; Singer, Michael Bliss; Marvin-DiPasquale, Mark; Tsui, Martin T K
2016-10-15
Monomethyl mercury (MMHg) and total mercury (THg) concentrations and Hg stable isotope ratios (δ(202)Hg and Δ(199)Hg) were measured in sediment and aquatic organisms from Cache Creek (California Coast Range) and Yolo Bypass (Sacramento Valley). Cache Creek sediment had a large range in THg (87 to 3870ng/g) and δ(202)Hg (-1.69 to -0.20‰) reflecting the heterogeneity of Hg mining sources in sediment. The δ(202)Hg of Yolo Bypass wetland sediment suggests a mixture of high and low THg sediment sources. Relationships between %MMHg (the percent ratio of MMHg to THg) and Hg isotope values (δ(202)Hg and Δ(199)Hg) in fish and macroinvertebrates were used to identify and estimate the isotopic composition of MMHg. Deviation from linear relationships was found between %MMHg and Hg isotope values, which is indicative of the bioaccumulation of isotopically distinct pools of MMHg. The isotopic composition of pre-photodegraded MMHg (i.e., subtracting fractionation from photochemical reactions) was estimated and contrasting relationships were observed between the estimated δ(202)Hg of pre-photodegraded MMHg and sediment IHg. Cache Creek had mass dependent fractionation (MDF; δ(202)Hg) of at least -0.4‰ whereas Yolo Bypass had MDF of +0.2 to +0.5‰. This result supports the hypothesis that Hg isotope fractionation between IHg and MMHg observed in rivers (-MDF) is unique compared to +MDF observed in non-flowing water environments such as wetlands, lakes, and the coastal ocean. Copyright © 2016 Elsevier B.V. All rights reserved.
Donovan, Patrick M.; Blum, Joel D.; Singer, Michael B.; Marvin-DiPasquale, Mark C.; Tsui, Martin T.K.
2016-01-01
Monomethyl mercury (MMHg) and total mercury (THg) concentrations and Hg stable isotope ratios (δ202Hg and Δ199Hg) were measured in sediment and aquatic organisms from Cache Creek (California Coast Range) and Yolo Bypass (Sacramento Valley). Cache Creek sediment had a large range in THg (87 to 3870 ng/g) and δ202Hg (−1.69 to −0.20‰) reflecting the heterogeneity of Hg mining sources in sediment. The δ202Hg of Yolo Bypass wetland sediment suggests a mixture of high and low THg sediment sources. Relationships between %MMHg (the percent ratio of MMHg to THg) and Hg isotope values (δ202Hg and Δ199Hg) in fish and macroinvertebrates were used to identify and estimate the isotopic composition of MMHg. Deviation from linear relationships was found between %MMHg and Hg isotope values, which is indicative of the bioaccumulation of isotopically distinct pools of MMHg. The isotopic composition of pre-photodegraded MMHg (i.e., subtracting fractionation from photochemical reactions) was estimated and contrasting relationships were observed between the estimated δ202Hg of pre-photodegraded MMHg and sediment IHg. Cache Creek had mass dependent fractionation (MDF; δ202Hg) of at least −0.4‰ whereas Yolo Bypass had MDF of +0.2 to +0.5‰. This result supports the hypothesis that Hg isotope fractionation between IHg and MMHg observed in rivers (−MDF) is unique compared to +MDF observed in non-flowing water environments such as wetlands, lakes, and the coastal ocean.
Cache as point of coherence in multiprocessor system
Blumrich, Matthias A.; Ceze, Luis H.; Chen, Dong; Gara, Alan; Heidelberger, Phlip; Ohmacht, Martin; Steinmacher-Burow, Burkhard; Zhuang, Xiaotong
2016-11-29
In a multiprocessor system, a conflict checking mechanism is implemented in the L2 cache memory. Different versions of speculative writes are maintained in different ways of the cache. A record of speculative writes is maintained in the cache directory. Conflict checking occurs as part of directory lookup. Speculative versions that do not conflict are aggregated into an aggregated version in a different way of the cache. Speculative memory access requests do not go to main memory.
Pravosudov, V V; Lavenex, P; Clayton, N S
2002-05-01
Earlier reports suggested that seasonal variation in food-caching behavior (caching intensity and cache retrieval accuracy) might correlate with morphological changes in the hippocampal formation, a brain structure thought to play a role in remembering cache locations. We demonstrated that changes in cache retrieval accuracy can also be triggered by experimental variation in food supply: captive mountain chickadees (Poecile gambeli) maintained on limited and unpredictable food supply were more accurate at recovering their caches and performed better on spatial memory tests than birds maintained on ad libitum food. In this study, we investigated whether these two treatment groups also differed in the volume and neuron number of the hippocampal formation. If variation in memory for food caches correlates with hippocampal size, then our birds with enhanced cache recovery and spatial memory performance should have larger hippocampal volumes and total neuron numbers. Contrary to this prediction we found no significant differences in volume or total neuron number of the hippocampal formation between the two treatment groups. Our results therefore indicate that changes in food-caching behavior and spatial memory performance, as mediated by experimental variations in food supply, are not necessarily accompanied by morphological changes in volume or neuron number of the hippocampal formation in fully developed, experienced food-caching birds. Copyright 2002 Wiley Periodicals, Inc.
A Refreshable, On-line Cache for HST Data Retrieval
NASA Astrophysics Data System (ADS)
Fraquelli, Dorothy A.; Ellis, Tracy A.; Ridgaway, Michael; DPAS Team
2016-01-01
We discuss upgrades to the HST Data Processing System, with an emphasis on the changes Hubble Space Telescope (HST) Archive users will experience. In particular, data are now held on-line (in a cache) removing the need to reprocess the data every time they are requested from the Archive. OTFR (on the fly reprocessing) has been replaced by a reprocessing system, which runs in the background. Data in the cache are automatically placed in the reprocessing queue when updated calibration reference files are received or when an improved calibration algorithm is installed. Data in the on-line cache are expected to be the most up to date version. These changes were phased in throughout 2015 for all active instruments.The on-line cache was populated instrument by instrument over the course of 2015. As data were placed in the cache, the flag that triggers OTFR was reset so that OTFR no longer runs on these data. "Hybrid" requests to the Archive are handled transparently, with data not yet in the cache provided via OTFR and the remaining data provided from the cache. Users do not need to make separate requests.Users of the MAST Portal will be able to download data from the cache immediately. For data not in the cache, the Portal will send the user to the standard "Retrieval Options Page," allowing the user to direct the Archive to process and deliver the data.The classic MAST Search and Retrieval interface has the same look and feel as previously. Minor changes, unrelated to the cache, have been made to the format of the Retrieval Options Page.
An Analysis of Instruction-Cached SIMD Computer Architecture
1993-12-01
ASSEBLE SIMULATE SCHEDULE VERIFY :t og ... . .. ... V~JSRUCTONSFOR PECIIEDCOMPARE ASSEMBLEI SIMULATE Ift*U1II ~ ~ SCHEDULEIinw ;. & VERIFY...Cache to Place Blocks ................. 70 4.5.4 Step 4: Schedule Cache Blocks ............................. 70 4.5.5 Step 5: Store Cache Blocks...167 B.4 Scheduler .............................................. 167 B.4.1 Basic Block Definition
Norton, Maria C; Singh, Archana; Skoog, Ingmar; Corcoran, Christopher; Tschanz, Joann T; Zandi, Peter P; Breitner, John C S; Welsh-Bohmer, Kathleen A; Steffens, David C
2008-05-01
We examined the relation between church attendance, membership in the Church of Jesus Christ of Latter-Day Saints (LDS), and major depressive episode, in a population-based study of aging and dementia in Cache County, Utah. Participants included 2,989 nondemented individuals aged between 65 and 100 years who were interviewed initially in 1995 to 1996 and again in 1998 to 1999. LDS church members reported twice the rate of major depression that non-LDS members did (odds ratio = 2.56, 95% confidence interval = 1.07-6.08). Individuals attending church weekly or more often had a significantly lower risk for major depression. After controlling for demographic and health variables and the strongest predictor of future episodes of depression, a prior depression history, we found that church attendance more often than weekly remained a significant protectant (odds ratio = 0.51, 95% confidence interval = 0.28-0.92). Results suggest that there may be a threshold of church attendance that is necessary for a person to garner long-term protection from depression. We discuss sociological factors relevant to LDS culture.
Leroux, Dorothée; Hezard, Nathalie; Lebreton, Aurélien; Bauters, Anne; Suchon, Pierre; de Maistre, Emmanuel; Biron, Christine; Huisse, Marie-Genevieve; Ternisien, Catherine; Voisin, Sophie; Gruel, Yves; Pouplard, Claire
2014-09-01
A rapid lateral flow immunoassay (LFIA) (STic Expert(®) HIT), recently developed for the diagnosis of heparin-induced thrombocytopenia (HIT), was evaluated in a prospective multicentre cohort of 334 consecutive patients. The risk of HIT was estimated by the 4Ts score as low, intermediate and high in 28·7%, 61·7% and 9·6% of patients, respectively. Definite HIT was diagnosed in 40 patients (12·0%) with positive results on both enzyme-linked immunosorbent assay (Asserachrom(®) HPIA IgG) and serotonin release assay. The inter-reader reproducibility of results obtained was excellent (kappa ratio > 0·9). The negative predictive value of LFIA with plasma samples was 99·6% with a negative likelihood ratio (LR) of 0·03, and was comparable to those of the particle gel immunoassay (H/PF4-PaGIA(®) ) performed in 124 cases. Positive predictive value and positive LR were 44·4% and 5·87, respectively, and the results were similar for serum samples. The probability of HIT in intermediate risk patients decreased from 11·2% to 0·4% when the LFIA result was negative and increased to 42·5% when it was positive. In conclusion, the STic Expert(®) HIT combined with the 4Ts score is a reliable tool to rule out the diagnosis of HIT. © 2014 John Wiley & Sons Ltd.
Effects of simulated mountain lion caching on decomposition of ungulate carcasses
Bischoff-Mattson, Z.; Mattson, D.
2009-01-01
Caching of animal remains is common among carnivorous species of all sizes, yet the effects of caching on larger prey are unstudied. We conducted a summer field experiment designed to test the effects of simulated mountain lion (Puma concolor) caching on mass loss, relative temperature, and odor dissemination of 9 prey-like carcasses. We deployed all but one of the carcasses in pairs, with one of each pair exposed and the other shaded and shallowly buried (cached). Caching substantially reduced wastage during dry and hot (drought) but not wet and cool (monsoon) periods, and it also reduced temperature and discernable odor to some degree during both seasons. These results are consistent with the hypotheses that caching serves to both reduce competition from arthropods and microbes and reduce odds of detection by larger vertebrates such as bears (Ursus spp.), wolves (Canis lupus), or other lions.
NASA Astrophysics Data System (ADS)
Tu, H.-Yu.; Tasneem, Sarah
Most of modern microprocessors employ on—chip cache memories to meet the memory bandwidth demand. These caches are now occupying a greater real es tate of chip area. Also, continuous down scaling of transistors increases the possi bility of defects in the cache area which already starts to occupies more than 50% of chip area. For this reason, various techniques have been proposed to tolerate defects in cache blocks. These techniques can be classified into three different cat egories, namely, cache line disabling, replacement with spare block, and decoder reconfiguration without spare blocks. This chapter examines each of those fault tol erant techniques with a fixed typical size and organization of L1 cache, through extended simulation using SPEC2000 benchmark on individual techniques. The de sign and characteristics of each technique are summarized with a view to evaluate the scheme. We then present our simulation results and comparative study of the three different methods.
Conditional load and store in a shared memory
Blumrich, Matthias A; Ohmacht, Martin
2015-02-03
A method, system and computer program product for implementing load-reserve and store-conditional instructions in a multi-processor computing system. The computing system includes a multitude of processor units and a shared memory cache, and each of the processor units has access to the memory cache. In one embodiment, the method comprises providing the memory cache with a series of reservation registers, and storing in these registers addresses reserved in the memory cache for the processor units as a result of issuing load-reserve requests. In this embodiment, when one of the processor units makes a request to store data in the memory cache using a store-conditional request, the reservation registers are checked to determine if an address in the memory cache is reserved for that processor unit. If an address in the memory cache is reserved for that processor, the data are stored at this address.
A Survey Of Techniques for Managing and Leveraging Caches in GPUs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh
2014-09-01
Initially introduced as special-purpose accelerators for graphics applications, graphics processing units (GPUs) have now emerged as general purpose computing platforms for a wide range of applications. To address the requirements of these applications, modern GPUs include sizable hardware-managed caches. However, several factors, such as unique architecture of GPU, rise of CPU–GPU heterogeneous computing, etc., demand effective management of caches to achieve high performance and energy efficiency. Recently, several techniques have been proposed for this purpose. In this paper, we survey several architectural and system-level techniques proposed for managing and leveraging GPU caches. We also discuss the importance and challenges ofmore » cache management in GPUs. The aim of this paper is to provide the readers insights into cache management techniques for GPUs and motivate them to propose even better techniques for leveraging the full potential of caches in the GPUs of tomorrow.« less
Cache coherency without line exclusivity in MP systems having store-in caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pomerene, J.H.; Puzak, T.R.; Rechtschaffen, R.N.
1983-11-01
By modifying the function of the storage control unit, a multiprocessor (MP) system having store-in caches is enabled to operate with the same versatility as an MP system having store-through caches, thereby eliminating the requirement for line exclusivity and greatly reducing the occurrence of cross-interrogates.
Nature as a treasure map! Teaching geoscience with the help of earth caches?!
NASA Astrophysics Data System (ADS)
Zecha, Stefanie; Schiller, Thomas
2015-04-01
This presentation looks at how earth caches are influence the learning process in the field of geo science in non-formal education. The development of mobile technologies using Global Positioning System (GPS) data to point geographical location together with the evolving Web 2.0 supporting the creation and consumption of content, suggest a potential for collaborative informal learning linked to location. With the help of the GIS in smartphones you can go directly in nature, search for information by your smartphone, and learn something about nature. Earth caches are a very good opportunity, which are organized and supervised geocaches with special information about physical geography high lights. Interested people can inform themselves about aspects in geoscience area by earth caches. The main question of this presentation is how these caches are created in relation to learning processes. As is not possible, to analyze all existing earth caches, there was focus on Bavaria and a certain feature of earth caches. At the end the authors show limits and potentials for the use of earth caches and give some remark for the future.
Explicit Content Caching at Mobile Edge Networks with Cross-Layer Sensing
Chen, Lingyu; Su, Youxing; Luo, Wenbin; Hong, Xuemin; Shi, Jianghong
2018-01-01
The deployment density and computational power of small base stations (BSs) are expected to increase significantly in the next generation mobile communication networks. These BSs form the mobile edge network, which is a pervasive and distributed infrastructure that can empower a variety of edge/fog computing applications. This paper proposes a novel edge-computing application called explicit caching, which stores selective contents at BSs and exposes such contents to local users for interactive browsing and download. We formulate the explicit caching problem as a joint content recommendation, caching, and delivery problem, which aims to maximize the expected user quality-of-experience (QoE) with varying degrees of cross-layer sensing capability. Optimal and effective heuristic algorithms are presented to solve the problem. The theoretical performance bounds of the explicit caching system are derived in simplified scenarios. The impacts of cache storage space, BS backhaul capacity, cross-layer information, and user mobility on the system performance are simulated and discussed in realistic scenarios. Results suggest that, compared with conventional implicit caching schemes, explicit caching can better exploit the mobile edge network infrastructure for personalized content dissemination. PMID:29565313
Explicit Content Caching at Mobile Edge Networks with Cross-Layer Sensing.
Chen, Lingyu; Su, Youxing; Luo, Wenbin; Hong, Xuemin; Shi, Jianghong
2018-03-22
The deployment density and computational power of small base stations (BSs) are expected to increase significantly in the next generation mobile communication networks. These BSs form the mobile edge network, which is a pervasive and distributed infrastructure that can empower a variety of edge/fog computing applications. This paper proposes a novel edge-computing application called explicit caching, which stores selective contents at BSs and exposes such contents to local users for interactive browsing and download. We formulate the explicit caching problem as a joint content recommendation, caching, and delivery problem, which aims to maximize the expected user quality-of-experience (QoE) with varying degrees of cross-layer sensing capability. Optimal and effective heuristic algorithms are presented to solve the problem. The theoretical performance bounds of the explicit caching system are derived in simplified scenarios. The impacts of cache storage space, BS backhaul capacity, cross-layer information, and user mobility on the system performance are simulated and discussed in realistic scenarios. Results suggest that, compared with conventional implicit caching schemes, explicit caching can better exploit the mobile edge network infrastructure for personalized content dissemination.
Value-Based Caching in Information-Centric Wireless Body Area Networks
Al-Turjman, Fadi M.; Imran, Muhammad; Vasilakos, Athanasios V.
2017-01-01
We propose a resilient cache replacement approach based on a Value of sensed Information (VoI) policy. To resolve and fetch content when the origin is not available due to isolated in-network nodes (fragmentation) and harsh operational conditions, we exploit a content caching approach. Our approach depends on four functional parameters in sensory Wireless Body Area Networks (WBANs). These four parameters are: age of data based on periodic request, popularity of on-demand requests, communication interference cost, and the duration for which the sensor node is required to operate in active mode to capture the sensed readings. These parameters are considered together to assign a value to the cached data to retain the most valuable information in the cache for prolonged time periods. The higher the value, the longer the duration for which the data will be retained in the cache. This caching strategy provides significant availability for most valuable and difficult to retrieve data in the WBANs. Extensive simulations are performed to compare the proposed scheme against other significant caching schemes in the literature while varying critical aspects in WBANs (e.g., data popularity, cache size, publisher load, connectivity-degree, and severe probabilities of node failures). These simulation results indicate that the proposed VoI-based approach is a valid tool for the retrieval of cached content in disruptive and challenging scenarios, such as the one experienced in WBANs, since it allows the retrieval of content for a long period even while experiencing severe in-network node failures. PMID:28106817
NASA Astrophysics Data System (ADS)
Chien, Chi-Sheng; Hung, Yu-Chien; Hong, Ting-Fu; Wu, Chung-Chun; Kuo, Tsung-Yuan; Lee, Tzer-Min; Liao, Tze-Yuan; Lin, Huan-Chang; Chuang, Cheng-Hsin
2017-03-01
Fluorapatite (FA) has better chemical and thermal stability than hydroxyapatite (HA), and has thus attracted significant interest for biomaterial applications in recent years. In this study, porous bioceramic layers were prepared on pure titanium surfaces using a micro-arc oxidation (MAO) technique with an applied voltage of 450 V and an oxidation time of 5 min. The MAO process was performed using three different electrolyte solutions containing calcium fluoride (CaF2), calcium acetate monohydrate (Ca(CH3COO)2·H2O), and sodium phosphate monobasic dihydrate (NaH2PO4·2H2O) mixed in ratios of 0:2:1, 1:1:1, and 2:0:1, respectively. The surface morphology, composition, micro-hardness, porosity, and biological properties of the various MAO coatings were examined and compared. The results showed that as the CaF2/Ca(CH3COO)2·H2O ratio increased, the elemental composition of the MAO coating transformed from HA, A-TiO2 (Anatase) and R-TiO2 (Rutile); to A-TiO2, R-TiO2, and a small amount of HA; and finally A-TiO2, R-TiO2, CaF2, TiP2O5, and FA. The change in elemental composition was accompanied by a higher micro-hardness and a lower porosity. The coatings exhibited a similar in vitro bioactivity performance during immersion in simulated body fluid for 7-28 days. Furthermore, for in initial in vitro biocompatibility tests performed for 24 h using Dulbecco's Modified Eagle Medium (DMEM) supplement containing 10%Fetal bovine serum, the attachment and spreading of osteoblast-like osteosarcoma MG63 cells were found to increase slightly with an increasing CaF2/Ca(CH3COO)2·H2O ratio. In general, the results presented in this study show that all three MAO coatings possess a certain degree of in vitro bioactivity and biocompatibility.
Corvid re-caching without 'theory of mind': a model.
van der Vaart, Elske; Verbrugge, Rineke; Hemelrijk, Charlotte K
2012-01-01
Scrub jays are thought to use many tactics to protect their caches. For instance, they predominantly bury food far away from conspecifics, and if they must cache while being watched, they often re-cache their worms later, once they are in private. Two explanations have been offered for such observations, and they are intensely debated. First, the birds may reason about their competitors' mental states, with a 'theory of mind'; alternatively, they may apply behavioral rules learned in daily life. Although this second hypothesis is cognitively simpler, it does seem to require a different, ad-hoc behavioral rule for every caching and re-caching pattern exhibited by the birds. Our new theory avoids this drawback by explaining a large variety of patterns as side-effects of stress and the resulting memory errors. Inspired by experimental data, we assume that re-caching is not motivated by a deliberate effort to safeguard specific caches from theft, but by a general desire to cache more. This desire is brought on by stress, which is determined by the presence and dominance of onlookers, and by unsuccessful recovery attempts. We study this theory in two experiments similar to those done with real birds with a kind of 'virtual bird', whose behavior depends on a set of basic assumptions about corvid cognition, and a well-established model of human memory. Our results show that the 'virtual bird' acts as the real birds did; its re-caching reflects whether it has been watched, how dominant its onlooker was, and how close to that onlooker it has cached. This happens even though it cannot attribute mental states, and it has only a single behavioral rule assumed to be previously learned. Thus, our simulations indicate that corvid re-caching can be explained without sophisticated social cognition. Given our specific predictions, our theory can easily be tested empirically.
Corvid Re-Caching without ‘Theory of Mind’: A Model
van der Vaart, Elske; Verbrugge, Rineke; Hemelrijk, Charlotte K.
2012-01-01
Scrub jays are thought to use many tactics to protect their caches. For instance, they predominantly bury food far away from conspecifics, and if they must cache while being watched, they often re-cache their worms later, once they are in private. Two explanations have been offered for such observations, and they are intensely debated. First, the birds may reason about their competitors' mental states, with a ‘theory of mind’; alternatively, they may apply behavioral rules learned in daily life. Although this second hypothesis is cognitively simpler, it does seem to require a different, ad-hoc behavioral rule for every caching and re-caching pattern exhibited by the birds. Our new theory avoids this drawback by explaining a large variety of patterns as side-effects of stress and the resulting memory errors. Inspired by experimental data, we assume that re-caching is not motivated by a deliberate effort to safeguard specific caches from theft, but by a general desire to cache more. This desire is brought on by stress, which is determined by the presence and dominance of onlookers, and by unsuccessful recovery attempts. We study this theory in two experiments similar to those done with real birds with a kind of ‘virtual bird’, whose behavior depends on a set of basic assumptions about corvid cognition, and a well-established model of human memory. Our results show that the ‘virtual bird’ acts as the real birds did; its re-caching reflects whether it has been watched, how dominant its onlooker was, and how close to that onlooker it has cached. This happens even though it cannot attribute mental states, and it has only a single behavioral rule assumed to be previously learned. Thus, our simulations indicate that corvid re-caching can be explained without sophisticated social cognition. Given our specific predictions, our theory can easily be tested empirically. PMID:22396799
Efficacy of Code Optimization on Cache-based Processors
NASA Technical Reports Server (NTRS)
VanderWijngaart, Rob F.; Chancellor, Marisa K. (Technical Monitor)
1997-01-01
The current common wisdom in the U.S. is that the powerful, cost-effective supercomputers of tomorrow will be based on commodity (RISC) micro-processors with cache memories. Already, most distributed systems in the world use such hardware as building blocks. This shift away from vector supercomputers and towards cache-based systems has brought about a change in programming paradigm, even when ignoring issues of parallelism. Vector machines require inner-loop independence and regular, non-pathological memory strides (usually this means: non-power-of-two strides) to allow efficient vectorization of array operations. Cache-based systems require spatial and temporal locality of data, so that data once read from main memory and stored in high-speed cache memory is used optimally before being written back to main memory. This means that the most cache-friendly array operations are those that feature zero or unit stride, so that each unit of data read from main memory (a cache line) contains information for the next iteration in the loop. Moreover, loops ought to be 'fat', meaning that as many operations as possible are performed on cache data-provided instruction caches do not overflow and enough registers are available. If unit stride is not possible, for example because of some data dependency, then care must be taken to avoid pathological strides, just ads on vector computers. For cache-based systems the issues are more complex, due to the effects of associativity and of non-unit block (cache line) size. But there is more to the story. Most modern micro-processors are superscalar, which means that they can issue several (arithmetic) instructions per clock cycle, provided that there are enough independent instructions in the loop body. This is another argument for providing fat loop bodies. With these restrictions, it appears fairly straightforward to produce code that will run efficiently on any cache-based system. It can be argued that although some of the important computational algorithms employed at NASA Ames require different programming styles on vector machines and cache-based machines, respectively, neither architecture class appeared to be favored by particular algorithms in principle. Practice tells us that the situation is more complicated. This report presents observations and some analysis of performance tuning for cache-based systems. We point out several counterintuitive results that serve as a cautionary reminder that memory accesses are not the only factors that determine performance, and that within the class of cache-based systems, significant differences exist.
Warkentin, Theodore E; Sheppard, Jo-Ann I; Linkins, Lori-Ann; Arnold, Donald M; Nazy, Ishac
2017-05-01
Heparin-induced thrombocytopenia (HIT) is a prothrombotic drug reaction caused by platelet-activating anti-PF4/heparin antibodies. Given time-sensitive treatment considerations, a rapid and accurate laboratory test for HIT antibodies is needed. To determine operating characteristics for the HemosIL ® HIT-Ab (PF4/H) , a rapid, on-demand, fully-automated, latex immunoturbidimetric assay (LIA), for diagnosis of HIT. We evaluated LIA sensitivity, specificity, negative (NPV) and positive predictive value (PPV), negative (LR-) and positive likelihood ratio (LR+), using citrated-plasma from 429 patients (prospective cohort study of 4Ts scoring; HIT, n=31), and from consecutive HIT patients (n=125), using reference standard serotonin-release assay (SRA). Comparators included two PF4-dependent enzyme-immunoassays (EIAs). We used stratum-specific likelihood ratios (SSLRs) to determine how differing magnitudes of LIA-positivity influenced post-test probability of HIT. LIA operating characteristics were: sensitivity=97.4% (152/156); specificity=94.0% (374/398); PPV=55.6% (30/54); and NPV=99.7% (374/375). At manufacturers' cutoffs, LIA specificity and PPV were superior to the EIAs. Although a negative LIA pointed strongly against HIT (LR-, 0.034), the post-test probability was ~2% with high 4Ts score. The LIA's LR+ was high (16.0), with SSLRs rising substantially with greater LIA-positivity: 5.7 (1.0-4.9U/mL), 31 (5.0-15.9U/mL), and 128 (≥16U/mL). A LIA-positive result (at 1.0 cutoff) indicated at least 24% HIT probability (low 4Ts score), rising to 90% with high 4Ts score. Although approximately 1 in 40 SRA-positive patients tested LIA-negative, the LIA's high NPV and PPV indicate that this rapid assay is useful for the diagnostic evaluation of HIT, including in low pre-test situations. Copyright © 2017 The Authors. Published by Elsevier Ltd.. All rights reserved.
44 CFR 208.24 - Purchase and maintenance of items not listed on Equipment Cache List.
Code of Federal Regulations, 2011 CFR
2011-10-01
... items not listed on Equipment Cache List. 208.24 Section 208.24 Emergency Management and Assistance... of items not listed on Equipment Cache List. (a) Requests for purchase or maintenance of equipment and supplies not appearing on the Equipment Cache List, or that exceed the number specified in the...
Effects of cacheing on multitasking efficiency and programming strategy on an ELXSI 6400
DOE Office of Scientific and Technical Information (OSTI.GOV)
Montry, G.R.; Benner, R.E.
1985-12-01
The impact of a cache/shared memory architecture, and, in particular, the cache coherency problem, upon concurrent algorithm and program development is discussed. In this context, a simple set of programming strategies are proposed which streamline code development and improve code performance when multitasking in a cache/shared memory or distributed memory environment.
Cache-based error recovery for shared memory multiprocessor systems
NASA Technical Reports Server (NTRS)
Wu, Kun-Lung; Fuchs, W. Kent; Patel, Janak H.
1989-01-01
A multiprocessor cache-based checkpointing and recovery scheme for of recovering from transient processor errors in a shared-memory multiprocessor with private caches is presented. New implementation techniques that use checkpoint identifiers and recovery stacks to reduce performance degradation in processor utilization during normal execution are examined. This cache-based checkpointing technique prevents rollback propagation, provides for rapid recovery, and can be integrated into standard cache coherence protocols. An analytical model is used to estimate the relative performance of the scheme during normal execution. Extensions that take error latency into account are presented.
Vincent, Grace; Lamon, Séverine; Gant, Nicholas; Vincent, Peter J.; MacDonald, Julia R.; Markworth, James F.; Edge, Johann A.; Hickey, Anthony J. R.
2015-01-01
Purpose: High-intensity short-duration interval training (HIT) stimulates functional and metabolic adaptation in skeletal muscle, but the influence of HIT on mitochondrial function remains poorly studied in humans. Mitochondrial metabolism as well as mitochondrial-associated protein expression were tested in untrained participants performing HIT over a 2-week period. Methods: Eight males performed a single-leg cycling protocol (12 × 1 min intervals at 120% peak power output, 90 s recovery, 4 days/week). Muscle biopsies (vastus lateralis) were taken pre- and post-HIT. Mitochondrial respiration in permeabilized fibers, citrate synthase (CS) activity and protein expression of peroxisome proliferator-activated receptor gamma coactivator (PGC-1α) and respiratory complex components were measured. Results: HIT training improved peak power and time to fatigue. Increases in absolute oxidative phosphorylation (OXPHOS) capacities and CS activity were observed, but not in the ratio of CCO to the electron transport system (CCO/ETS), the respiratory control ratios (RCR-1 and RCR-2) or mitochondrial-associated protein expression. Specific increases in OXPHOS flux were not apparent after normalization to CS, indicating that gross changes mainly resulted from increased mitochondrial mass. Conclusion: Over only 2 weeks HIT significantly increased mitochondrial function in skeletal muscle independently of detectable changes in mitochondrial-associated and mitogenic protein expression. PMID:25759671
Forest resources of the Wasatch-Cache National Forest
Renee A. O' Brien; Jesse Pope
1997-01-01
The 1,215,219 acres in the Wasatch-Cache National Forest encompass 863,906 acres of forest land, made up of 90 percent (776,239 acres) "timberland" and 10 percent (87,667 acres) "woodland." The other 351,313 acres of the Wasatch-Cache are nonforest or water (fig. 1). This report discusses forest land only. In the Wasatch-Cache, 26 percent...
Problems faced by food-caching corvids and the evolution of cognitive solutions
Grodzinski, Uri; Clayton, Nicola S.
2010-01-01
The scatter hoarding of food, or caching, is a widespread and well-studied behaviour. Recent experiments with caching corvids have provided evidence for episodic-like memory, future planning and possibly mental attribution, all cognitive abilities that were thought to be unique to humans. In addition to the complexity of making flexible, informed decisions about caching and recovering, this behaviour is underpinned by a motivationally controlled compulsion to cache. In this review, we shall first discuss the compulsive side of caching both during ontogeny and in the caching behaviour of adult corvids. We then consider some of the problems that these birds face and review the evidence for the cognitive abilities they use to solve them. Thus, the emergence of episodic-like memory is viewed as a solution for coping with food perishability, while the various cache-protection and pilfering strategies may be sophisticated tools to deprive competitors of information, either by reducing the quality of information they can gather, or invalidating the information they already have. Finally, we shall examine whether such future-oriented behaviour involves future planning and ask why this and other cognitive abilities might have evolved in corvids. PMID:20156820
A Morphometric Assessment of the Intended Function of Cached Clovis Points
Buchanan, Briggs; Kilby, J. David; Huckell, Bruce B.; O'Brien, Michael J.; Collard, Mark
2012-01-01
A number of functions have been proposed for cached Clovis points. The least complicated hypothesis is that they were intended to arm hunting weapons. It has also been argued that they were produced for use in rituals or in connection with costly signaling displays. Lastly, it has been suggested that some cached Clovis points may have been used as saws. Here we report a study in which we morphometrically compared Clovis points from caches with Clovis points recovered from kill and camp sites to test two predictions of the hypothesis that cached Clovis points were intended to arm hunting weapons: 1) cached points should be the same shape as, but generally larger than, points from kill/camp sites, and 2) cached points and points from kill/camp sites should follow the same allometric trajectory. The results of the analyses are consistent with both predictions and therefore support the hypothesis. A follow-up review of the fit between the results of the analyses and the predictions of the other hypotheses indicates that the analyses support only the hunting equipment hypothesis. We conclude from this that cached Clovis points were likely produced with the intention of using them to arm hunting weapons. PMID:22348012
WATCHMAN: A Data Warehouse Intelligent Cache Manager
NASA Technical Reports Server (NTRS)
Scheuermann, Peter; Shim, Junho; Vingralek, Radek
1996-01-01
Data warehouses store large volumes of data which are used frequently by decision support applications. Such applications involve complex queries. Query performance in such an environment is critical because decision support applications often require interactive query response time. Because data warehouses are updated infrequently, it becomes possible to improve query performance by caching sets retrieved by queries in addition to query execution plans. In this paper we report on the design of an intelligent cache manager for sets retrieved by queries called WATCHMAN, which is particularly well suited for data warehousing environment. Our cache manager employs two novel, complementary algorithms for cache replacement and for cache admission. WATCHMAN aims at minimizing query response time and its cache replacement policy swaps out entire retrieved sets of queries instead of individual pages. The cache replacement and admission algorithms make use of a profit metric, which considers for each retrieved set its average rate of reference, its size, and execution cost of the associated query. We report on a performance evaluation based on the TPC-D and Set Query benchmarks. These experiments show that WATCHMAN achieves a substantial performance improvement in a decision support environment when compared to a traditional LRU replacement algorithm.
Novel dynamic caching for hierarchically distributed video-on-demand systems
NASA Astrophysics Data System (ADS)
Ogo, Kenta; Matsuda, Chikashi; Nishimura, Kazutoshi
1998-02-01
It is difficult to simultaneously serve the millions of video streams that will be needed in the age of 'Mega-Media' networks by using only one high-performance server. To distribute the service load, caching servers should be location near users. However, in previously proposed caching mechanisms, the grade of service depends on whether the data is already cached at a caching server. To make the caching servers transparent to the users, the ability to randomly access the large volume of data stored in the central server should be supported, and the operational functions of the provided service should not be narrowly restricted. We propose a mechanism for constructing a video-stream-caching server that is transparent to the users and that will always support all special playback functions for all available programs to all the contents with a latency of only 1 or 2 seconds. This mechanism uses Variable-sized-quantum-segment- caching technique derived from an analysis of the historical usage log data generated by a line-on-demand-type service experiment and based on the basic techniques used by a time- slot-based multiple-stream video-on-demand server.
Do Clark's nutcrackers demonstrate what-where-when memory on a cache-recovery task?
Gould, Kristy L; Ort, Amy J; Kamil, Alan C
2012-01-01
What-where-when (WWW) memory during cache recovery was investigated in six Clark's nutcrackers. During caching, both red- and blue-colored pine seeds were cached by the birds in holes filled with sand. Either a short (3 day) retention interval (RI) or a long (9 day) RI was followed by a recovery session during which caches were replaced with either a single seed or wooden bead depending upon the color of the cache and length of the retention interval. Knowledge of what was in the cache (seed or bead), where it was located, and when the cache had been made (3 or 9 days ago) were the three WWW memory components under investigation. Birds recovered items (bead or seed) at above chance levels, demonstrating accurate spatial memory. They also recovered seeds more than beads after the long RI, but not after the short RI, when they recovered seeds and beads equally often. The differential recovery after the long RI demonstrates that nutcrackers may have the capacity for WWW memory during this task, but it is not clear why it was influenced by RI duration.
Problems faced by food-caching corvids and the evolution of cognitive solutions.
Grodzinski, Uri; Clayton, Nicola S
2010-03-27
The scatter hoarding of food, or caching, is a widespread and well-studied behaviour. Recent experiments with caching corvids have provided evidence for episodic-like memory, future planning and possibly mental attribution, all cognitive abilities that were thought to be unique to humans. In addition to the complexity of making flexible, informed decisions about caching and recovering, this behaviour is underpinned by a motivationally controlled compulsion to cache. In this review, we shall first discuss the compulsive side of caching both during ontogeny and in the caching behaviour of adult corvids. We then consider some of the problems that these birds face and review the evidence for the cognitive abilities they use to solve them. Thus, the emergence of episodic-like memory is viewed as a solution for coping with food perishability, while the various cache-protection and pilfering strategies may be sophisticated tools to deprive competitors of information, either by reducing the quality of information they can gather, or invalidating the information they already have. Finally, we shall examine whether such future-oriented behaviour involves future planning and ask why this and other cognitive abilities might have evolved in corvids.
Short-term observational spatial memory in Jackdaws (Corvus monedula) and Ravens (Corvus corax).
Scheid, Christelle; Bugnyar, Thomas
2008-10-01
Observational spatial memory (OSM) refers to the ability of remembering food caches made by other individuals, enabling observers to find and pilfer the others' caches. Within birds, OSM has only been demonstrated in corvids, with more social species such as Mexican jays (Aphelocoma ultramarine) showing a higher accuracy of finding conspecific' caches than less social species such as Clark's nutcrackers (Nucifraga columbiana). However, socially dynamic corvids such as ravens (Corvus corax) are capable of sophisticated pilfering manoeuvres based on OSM. We here compared the performance of ravens and jackdaws (Corvus monedula) in a short-term OSM task. In contrast to ravens, jackdaws are socially cohesive but hardly cache and compete over food caches. Birds had to recover food pieces after watching a human experimenter hiding them in 2, 4 or 6 out of 10 possible locations. Results showed that for tests with two, four and six caches, ravens performed more accurately than expected by chance whereas jackdaws did not. Moreover, ravens made fewer re-visits to already inspected cache sites than jackdaws. These findings suggest that the development of observational spatial memory skills is linked with the species' reliance on food caches rather than with a social life style per se.
EqualChance: Addressing Intra-set Write Variation to Increase Lifetime of Non-volatile Caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Vetter, Jeffrey S
To address the limitations of SRAM such as high-leakage and low-density, researchers have explored use of non-volatile memory (NVM) devices, such as ReRAM (resistive RAM) and STT-RAM (spin transfer torque RAM) for designing on-chip caches. A crucial limitation of NVMs, however, is that their write endurance is low and the large intra-set write variation introduced by existing cache management policies may further exacerbate this problem, thereby reducing the cache lifetime significantly. We present EqualChance, a technique to increase cache lifetime by reducing intra-set write variation. EqualChance works by periodically changing the physical cache-block location of a write-intensive data item withinmore » a set to achieve wear-leveling. Simulations using workloads from SPEC CPU2006 suite and HPC (high-performance computing) field show that EqualChance improves the cache lifetime by 4.29X. Also, its implementation overhead is small, and it incurs very small performance and energy loss.« less
Store-operate-coherence-on-value
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Dong; Heidelberger, Philip; Kumar, Sameer
A system, method and computer program product for performing various store-operate instructions in a parallel computing environment that includes a plurality of processors and at least one cache memory device. A queue in the system receives, from a processor, a store-operate instruction that specifies under which condition a cache coherence operation is to be invoked. A hardware unit in the system runs the received store-operate instruction. The hardware unit evaluates whether a result of the running the received store-operate instruction satisfies the condition. The hardware unit invokes a cache coherence operation on a cache memory address associated with the receivedmore » store-operate instruction if the result satisfies the condition. Otherwise, the hardware unit does not invoke the cache coherence operation on the cache memory device.« less
Multicast for savings in cache-based video distribution
NASA Astrophysics Data System (ADS)
Griwodz, Carsten; Zink, Michael; Liepert, Michael; On, Giwon; Steinmetz, Ralf
1999-12-01
Internet video-on-demand (VoD) today streams videos directly from server to clients, because re-distribution is not established yet. Intranet solutions exist but are typically managed centrally. Caching may overcome these management needs, however existing web caching strategies are not applicable because they work in different conditions. We propose movie distribution by means of caching, and study the feasibility from the service providers' point of view. We introduce the combination of our reliable multicast protocol LCRTP for caching hierarchies combined with our enhancement to the patching technique for bandwidth friendly True VoD, not depending on network resource guarantees.
Store operations to maintain cache coherence
Evangelinos, Constantinos; Nair, Ravi; Ohmacht, Martin
2017-08-01
In one embodiment, a computer-implemented method includes encountering a store operation during a compile-time of a program, where the store operation is applicable to a memory line. It is determined, by a computer processor, that no cache coherence action is necessary for the store operation. A store-without-coherence-action instruction is generated for the store operation, responsive to determining that no cache coherence action is necessary. The store-without-coherence-action instruction specifies that the store operation is to be performed without a cache coherence action, and cache coherence is maintained upon execution of the store-without-coherence-action instruction.
Urhan, A Utku; Emilsson, Ellen; Brodin, Anders
2017-01-01
Many species in the family Paridae, such as marsh tits Poecile palustris , are large-scale scatter hoarders of food that make cryptic caches and disperse these in large year-round territories. The perhaps most well-known species in the family, the great tit Parus major , does not store food itself but is skilled in stealing caches from the other species. We have previously demonstrated that great tits are able to memorise positions of caches they have observed marsh tits make and later return and steal the food. As great tits are explorative in nature and unusually good learners, it is possible that such "memorisation of caches from a distance" is a unique ability of theirs. The other possibility is that this ability is general in the parid family. Here, we tested marsh tits in the same experimental set-up as where we previously have tested great tits. We allowed caged marsh tits to observe a caching conspecific in a specially designed indoor arena. After a retention interval of 1 or 24 h, we allowed the observer to enter the arena and search for the caches. The marsh tits showed no evidence of such observational memorization ability, and we believe that such ability is more useful for a non-hoarding species. Why should a marsh tit that memorises hundreds of their own caches in the field bother with the difficult task of memorising other individuals' caches? We argue that the close-up memorisation procedure that marsh tits use at their own caches may be a different type of observational learning than memorisation of caches made by others. For example, the latter must be done from a distance and hence may require the ability to adopt an allocentric perspective, i.e. the ability to visualise the cache from the hoarder's perspective. Members of the Paridae family are known to possess foraging techniques that are cognitively advanced. Previously, we have demonstrated that a non-hoarding parid species, the great tit P. major , is able to memorise positions of caches that they have observed marsh tits P. palustris make. However, it is unknown whether this cognitively advanced foraging strategy is unique to great tits or if it occurs also in other parids. Here, we demonstrated that "pilfering by observational memorization strategy" is not a general strategy in parids. We believe that such ability is important for a non-hoarding species such as the great tit and, most likely, birds owning many caches do not need this foraging strategy.
Cache directory look-up re-use as conflict check mechanism for speculative memory requests
Ohmacht, Martin
2013-09-10
In a cache memory, energy and other efficiencies can be realized by saving a result of a cache directory lookup for sequential accesses to a same memory address. Where the cache is a point of coherence for speculative execution in a multiprocessor system, with directory lookups serving as the point of conflict detection, such saving becomes particularly advantageous.
Nelson, C; Wengreen, H J; Munger, R G; Corcoran, C D
2009-12-01
To examine associations between dietary and supplemental folate, vitamin B-12 and vitamin B-6 and incident Alzheimer's disease (AD) among elderly men and women. Data collected were from participants of the Cache County Memory, Health and Aging Study, a longitudinal study of 5092 men and women 65 years and older who were residents of Cache County, Utah in 1995. Multistage clinical assessment procedures were used to identify incident cases of AD. Dietary data were collected using a 142-item food frequency questionnaire. Cox Proportional Hazards (CPH) modeling was used to determine hazard ratios across quintiles of micronutrient intake. 202 participants were diagnosed with incident AD during follow-up (1995-2004). In multivariable CPH models that controlled for the effects of gender, age, education, and other covariates there were no observed differences in risk of AD or dementia by increasing quintiles of total intake of folate, vitamin B-12, or vitamin B-6. Similarly, there were no observed differences in risk of AD by regular use of either folate or B6 supplements. Dietary intake of B-vitamins from food and supplemental sources appears unrelated to incidence of dementia and AD. Further studies examining associations between dietary intakes of B-vitamins, biomarkers of B-vitamin status and cognitive endpoints are warranted.
Effective Padding of Multi-Dimensional Arrays to Avoid Cache Conflict Misses
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hong, Changwan; Bao, Wenlei; Cohen, Albert
Caches are used to significantly improve performance. Even with high degrees of set-associativity, the number of accessed data elements mapping to the same set in a cache can easily exceed the degree of associativity, causing conflict misses and lowered performance, even if the working set is much smaller than cache capacity. Array padding (increasing the size of array dimensions) is a well known optimization technique that can reduce conflict misses. In this paper, we develop the first algorithms for optimal padding of arrays for a set associative cache for arbitrary tile sizes, In addition, we develop the first solution tomore » padding for nested tiles and multi-level caches. The techniques are in implemented in PAdvisor tool. Experimental results with multiple benchmarks demonstrate significant performance improvement from use of PAdvisor for padding.« less
Zwolak, Rafał; Bogdziewicz, Michał; Wróbel, Aleksandra; Crone, Elizabeth E
2016-03-01
The predator satiation and predator dispersal hypotheses provide alternative explanations for masting. Both assume satiation of seed-eating vertebrates. They differ in whether satiation occurs before or after seed removal and caching by granivores (predator satiation and predator dispersal, respectively). This difference is largely unrecognized, but it is demographically important because cached seeds are dispersed and often have a microsite advantage over nondispersed seeds. We conducted rodent exclosure experiments in two mast and two nonmast years to test predictions of the predator dispersal hypothesis in our study system of yellow-necked mice (Apodemus flavicollis) and European beech (Fagus sylvatica). Specifically, we tested whether the fraction of seeds removed from the forest floor is similar during mast and nonmast years (i.e., lack of satiation before seed caching), whether masting decreases the removal of cached seeds (i.e., satiation after seed storage), and whether seed caching increases the probability of seedling emergence. We found that masting did not result in satiation at the seed removal stage. However, masting decreased the removal of cached seeds, and seed caching dramatically increased the probability of seedling emergence relative to noncached seeds. European beech thus benefits from masting through the satiation of scatterhoarders that occurs only after seeds are removed and cached. Although these findings do not exclude other evolutionary advantages of beech masting, they indicate that fitness benefits of masting extend beyond the most commonly considered advantages of predator satiation and increased pollination efficiency.
Saccadic entropy of head impulses in acute unilateral vestibular loss.
Hsieh, Li-Chun; Lin, Hung-Ching; Lee, Guo-She
2017-10-01
To evaluate the complexity of vestibular-ocular reflex (VOR) in patients with acute unilateral vestibular loss (AUVL) via entropy analysis of head impulses. Horizontal head impulse test (HIT) with high-velocity alternating directions was used to evaluate 12 participants with AUVL and 16 healthy volunteers. Wireless electro-oculography and electronic gyrometry were used to acquire eye positional signals and head velocity signals. The eye velocity signals were then obtained through differentiation, band-pass filtering. The approximate entropy of eye velocity to head velocity (R ApEn ) was used to evaluate chaos property. VOR gain, gain asymmetry ratio, and R ApEn asymmetry ratio were also used to compare the groups. For the lesion-side HIT of the patient group, the mean VOR gain was significantly lower and the mean R ApEn was significantly greater compared with both nonlesion-side HIT and healthy controls (p < 0.01, one-way analysis of variance). Both the R ApEn asymmetry ratio and gain asymmetry ratio of the AUVL group were significantly greater compared with those of the control group (p < 0.05, independent sample t test). Entropy and gain analysis of HIT using wireless electro-oculography system could be used to detect the VOR dysfunctions of AUVL and may become effective methods for evaluating vestibular disorders. Copyright © 2017. Published by Elsevier B.V.
Corvid caching: Insights from a cognitive model.
van der Vaart, Elske; Verbrugge, Rineke; Hemelrijk, Charlotte K
2011-07-01
Caching and recovery of food by corvids is well-studied, but some ambiguous results remain. To help clarify these, we built a computational cognitive model. It is inspired by similar models built for humans, and it assumes that memory strength depends on frequency and recency of use. We compared our model's behavior to that of real birds in previously published experiments. Our model successfully replicated the outcomes of two experiments on recovery behavior and two experiments on cache site choice. Our "virtual birds" reproduced declines in recovery accuracy across sessions, revisits to previously emptied cache sites, a lack of correlation between caching and recovery order, and a preference for caching in safe locations. The model also produced two new explanations. First, that Clark's nutcrackers may become less accurate as recovery progresses not because of differential memory for different cache sites, as was once assumed, but because of chance effects. And second, that Western scrub jays may choose their cache sites not on the basis of negative recovery experiences only, as was previously thought, but on the basis of positive recovery experiences instead. Alternatively, both "punishment" and "reward" may be playing a role. We conclude with a set of new insights, a testable prediction, and directions for future work. PsycINFO Database Record (c) 2011 APA, all rights reserved
OS friendly microprocessor architecture: Hardware level computer security
NASA Astrophysics Data System (ADS)
Jungwirth, Patrick; La Fratta, Patrick
2016-05-01
We present an introduction to the patented OS Friendly Microprocessor Architecture (OSFA) and hardware level computer security. Conventional microprocessors have not tried to balance hardware performance and OS performance at the same time. Conventional microprocessors have depended on the Operating System for computer security and information assurance. The goal of the OS Friendly Architecture is to provide a high performance and secure microprocessor and OS system. We are interested in cyber security, information technology (IT), and SCADA control professionals reviewing the hardware level security features. The OS Friendly Architecture is a switched set of cache memory banks in a pipeline configuration. For light-weight threads, the memory pipeline configuration provides near instantaneous context switching times. The pipelining and parallelism provided by the cache memory pipeline provides for background cache read and write operations while the microprocessor's execution pipeline is running instructions. The cache bank selection controllers provide arbitration to prevent the memory pipeline and microprocessor's execution pipeline from accessing the same cache bank at the same time. This separation allows the cache memory pages to transfer to and from level 1 (L1) caching while the microprocessor pipeline is executing instructions. Computer security operations are implemented in hardware. By extending Unix file permissions bits to each cache memory bank and memory address, the OSFA provides hardware level computer security.
Experimental evaluation of multiprocessor cache-based error recovery
NASA Technical Reports Server (NTRS)
Janssens, Bob; Fuchs, W. K.
1991-01-01
Several variations of cache-based checkpointing for rollback error recovery in shared-memory multiprocessors have been recently developed. By modifying the cache replacement policy, these techniques use the inherent redundancy in the memory hierarchy to periodically checkpoint the computation state. Three schemes, different in the manner in which they avoid rollback propagation, are evaluated. By simulation with address traces from parallel applications running on an Encore Multimax shared-memory multiprocessor, the performance effect of integrating the recovery schemes in the cache coherence protocol are evaluated. The results indicate that the cache-based schemes can provide checkpointing capability with low performance overhead but uncontrollable high variability in the checkpoint interval.
Version pressure feedback mechanisms for speculative versioning caches
Eichenberger, Alexandre E.; Gara, Alan; O& #x27; Brien, Kathryn M.; Ohmacht, Martin; Zhuang, Xiaotong
2013-03-12
Mechanisms are provided for controlling version pressure on a speculative versioning cache. Raw version pressure data is collected based on one or more threads accessing cache lines of the speculative versioning cache. One or more statistical measures of version pressure are generated based on the collected raw version pressure data. A determination is made as to whether one or more modifications to an operation of a data processing system are to be performed based on the one or more statistical measures of version pressure, the one or more modifications affecting version pressure exerted on the speculative versioning cache. An operation of the data processing system is modified based on the one or more determined modifications, in response to a determination that one or more modifications to the operation of the data processing system are to be performed, to affect the version pressure exerted on the speculative versioning cache.
Norton, Maria C.; Singh, Archana; Skoog, Ingmar; Corcoran, Christopher; Tschanz, JoAnn T.; Zandi, Peter P.; Breitner, John C. S.; Welsh-Bohmer, Kathleen A.; Steffens, David C.
2009-01-01
We examined the relation between church attendance, membership in the Church of Jesus Christ of Latter-Day Saints (LDS), and major depressive episode, in a population-based study of aging and dementia in Cache County, Utah. Participants included 2,989 nondemented individuals aged between 65 and 100 years who were interviewed initially in 1995 to 1996 and again in 1998 to 1999. LDS church members reported twice the rate of major depression that non-LDS members did (odds ratio = 2.56, 95% confidence interval = 1.07-6.08). Individuals attending church weekly or more often had a significantly lower risk for major depression. After controlling for demographic and health variables and the strongest predictor of future episodes of depression, a prior depression history, we found that church attendance more often than weekly remained a significant protectant (odds ratio = 0.51, 95% confidence interval = 0.28-0.92). Results suggest that there may be a threshold of church attendance that is necessary for a person to garner long-term protection from depression. We discuss sociological factors relevant to LDS culture. PMID:18559677
Hoarding patterns in the southern flying squirrel (Glaucomys volans)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Clemmer, S.M.
1984-05-01
Southern flying squirrels, Glaucomys volans, were individually offered two size classes of pecans in a 1:1 ratio to establish preference. All but one squirrel preferred small pecans. As relative abundance of preferred food diminished to 0.10, squirrels switched preference. Absolute abundance of either food did not affect caching levels. A difficulty-of-retrieval experiment did not result in switching of preference. No effect of sex on hoarding was exhibited. There was an inverse correlation between individual storing effort and caching levels of the same squirrels tested as pairs, with individual non-storers showing increases in numbers of pecans stored. Animals that were activemore » storers as individuals showed decreases when paired. Total number stored did not decrease significantly when squirrels were offered previously marked pecans. When offered own-marked and other-marked pecans, squirrels did not discriminate. 43 references, 3 figures, 6 tables.« less
Enabling MPEG-2 video playback in embedded systems through improved data cache efficiency
NASA Astrophysics Data System (ADS)
Soderquist, Peter; Leeser, Miriam E.
1999-01-01
Digital video decoding, enabled by the MPEG-2 Video standard, is an important future application for embedded systems, particularly PDAs and other information appliances. Many such system require portability and wireless communication capabilities, and thus face severe limitations in size and power consumption. This places a premium on integration and efficiency, and favors software solutions for video functionality over specialized hardware. The processors in most embedded system currently lack the computational power needed to perform video decoding, but a related and equally important problem is the required data bandwidth, and the need to cost-effectively insure adequate data supply. MPEG data sets are very large, and generate significant amounts of excess memory traffic for standard data caches, up to 100 times the amount required for decoding. Meanwhile, cost and power limitations restrict cache sizes in embedded systems. Some systems, including many media processors, eliminate caches in favor of memories under direct, painstaking software control in the manner of digital signal processors. Yet MPEG data has locality which caches can exploit if properly optimized, providing fast, flexible, and automatic data supply. We propose a set of enhancements which target the specific needs of the heterogeneous types within the MPEG decoder working set. These optimizations significantly improve the efficiency of small caches, reducing cache-memory traffic by almost 70 percent, and can make an enhanced 4 KB cache perform better than a standard 1 MB cache. This performance improvement can enable high-resolution, full frame rate video playback in cheaper, smaller system than woudl otherwise be possible.
MELOC - Memory and Location Optimized Caching for Mobile Ad Hoc Networks
2011-01-01
required for such environments. Moreover, nodes located at centre have to be chosen as cache location, since it reduces the chance of being attacked...Figure 1.1. MANET Formed by Armed Forces 47 Example 3: Sharing of music and videos are famous among mobile users. Instead of downloading...The two tier caching scheme discussed in this paper is acoustic . The characteristics of two-tier caching are as follows, the content of data to be
Visual landmark-directed scatter-hoarding of Siberian chipmunks Tamias sibiricus.
Zhang, Dongyuan; Li, Jia; Wang, Zhenyu; Yi, Xianfeng
2016-05-01
Spatial memory of cached food items plays an important role in cache recovery by scatter-hoarding animals. However, whether scatter-hoarding animals intentionally select cache sites with respect to visual landmarks in the environment and then rely on them to recover their cached seeds for later use has not been extensively explored. Furthermore, there is a lack of evidence on whether there are sex differences in visual landmark-based food-hoarding behaviors in small rodents even though male and female animals exhibit different spatial abilities. In the present study, we used a scatter-hoarding animal, the Siberian chipmunk, Tamias sibiricus to explore these questions in semi-natural enclosures. Our results showed that T. sibiricus preferred to establish caches in the shallow pits labeled with visual landmarks (branches of Pinus sylvestris, leaves of Athyrium brevifrons and PVC tubes). In addition, visual landmarks of P. sylvestris facilitated cache recovery by T. sibiricus. We also found significant sex differences in visual landmark-based food-hoarding strategies in Siberian chipmunks. Males, rather than females, chipmunks tended to establish their caches with respect to the visual landmarks. Our studies show that T. sibiricus rely on visual landmarks to establish and recover their caches, and that sex differences exist in visual landmark-based food hoarding in Siberian chipmunks. © 2015 International Society of Zoological Sciences, Institute of Zoology/Chinese Academy of Sciences and John Wiley & Sons Australia, Ltd.
Cost aware cache replacement policy in shared last-level cache for hybrid memory based fog computing
NASA Astrophysics Data System (ADS)
Jia, Gangyong; Han, Guangjie; Wang, Hao; Wang, Feng
2018-04-01
Fog computing requires a large main memory capacity to decrease latency and increase the Quality of Service (QoS). However, dynamic random access memory (DRAM), the commonly used random access memory, cannot be included into a fog computing system due to its high consumption of power. In recent years, non-volatile memories (NVM) such as Phase-Change Memory (PCM) and Spin-transfer torque RAM (STT-RAM) with their low power consumption have emerged to replace DRAM. Moreover, the currently proposed hybrid main memory, consisting of both DRAM and NVM, have shown promising advantages in terms of scalability and power consumption. However, the drawbacks of NVM, such as long read/write latency give rise to potential problems leading to asymmetric cache misses in the hybrid main memory. Current last level cache (LLC) policies are based on the unified miss cost, and result in poor performance in LLC and add to the cost of using NVM. In order to minimize the cache miss cost in the hybrid main memory, we propose a cost aware cache replacement policy (CACRP) that reduces the number of cache misses from NVM and improves the cache performance for a hybrid memory system. Experimental results show that our CACRP behaves better in LLC performance, improving performance up to 43.6% (15.5% on average) compared to LRU.
Optimal and Scalable Caching for 5G Using Reinforcement Learning of Space-Time Popularities
NASA Astrophysics Data System (ADS)
Sadeghi, Alireza; Sheikholeslami, Fatemeh; Giannakis, Georgios B.
2018-02-01
Small basestations (SBs) equipped with caching units have potential to handle the unprecedented demand growth in heterogeneous networks. Through low-rate, backhaul connections with the backbone, SBs can prefetch popular files during off-peak traffic hours, and service them to the edge at peak periods. To intelligently prefetch, each SB must learn what and when to cache, while taking into account SB memory limitations, the massive number of available contents, the unknown popularity profiles, as well as the space-time popularity dynamics of user file requests. In this work, local and global Markov processes model user requests, and a reinforcement learning (RL) framework is put forth for finding the optimal caching policy when the transition probabilities involved are unknown. Joint consideration of global and local popularity demands along with cache-refreshing costs allow for a simple, yet practical asynchronous caching approach. The novel RL-based caching relies on a Q-learning algorithm to implement the optimal policy in an online fashion, thus enabling the cache control unit at the SB to learn, track, and possibly adapt to the underlying dynamics. To endow the algorithm with scalability, a linear function approximation of the proposed Q-learning scheme is introduced, offering faster convergence as well as reduced complexity and memory requirements. Numerical tests corroborate the merits of the proposed approach in various realistic settings.
Salwiczek, Lucie H.; Schlinger, Barney; Emery, Nathan J.; Clayton, Nicola S.
2010-01-01
Recent studies on the food-caching behavior of corvids have revealed complex physical and social skills, yet little is known about the ontogeny of food caching in relation to the development of cognitive capacities. Piagetian object permanence is the understanding that objects continue to exist even when they are no longer visible. Here, the authors focus on Piagetian Stages 3 and 4, because they are hallmarks in the cognitive development of both young children and animals. Our aim is to determine in a food-caching corvid, the Western scrub-jay, whether (1) Piagetian Stage 4 competence and tentative caching (i.e., hiding an item invisibly and retrieving it without delay), emerge concomitantly or consecutively; (2) whether experiencing the reappearance of hidden objects enhances the timing of the appearance of object permanence; and (3) discuss how the development of object permanence is related to behavioral development and sensorimotor intelligence. Our findings suggest that object permanence Stage 4 emerges before tentative caching, and independent of environmental influences, but that once the birds have developed simple object-permanence, then social learning might advance the interval after which tentative caching commences. PMID:19685971
Salwiczek, Lucie H; Emery, Nathan J; Schlinger, Barney; Clayton, Nicola S
2009-08-01
Recent studies on the food-caching behavior of corvids have revealed complex physical and social skills, yet little is known about the ontogeny of food caching in relation to the development of cognitive capacities. Piagetian object permanence is the understanding that objects continue to exist even when they are no longer visible. Here, the authors focus on Piagetian Stages 3 and 4, because they are hallmarks in the cognitive development of both young children and animals. Our aim is to determine in a food-caching corvid, the Western scrub-jay, whether (1) Piagetian Stage 4 competence and tentative caching (i.e., hiding an item invisibly and retrieving it without delay), emerge concomitantly or consecutively; (2) whether experiencing the reappearance of hidden objects enhances the timing of the appearance of object permanence; and (3) discuss how the development of object permanence is related to behavioral development and sensorimotor intelligence. Our findings suggest that object permanence Stage 4 emerges before tentative caching, and independent of environmental influences, but that once the birds have developed simple object-permanence, then social learning might advance the interval after which tentative caching commences. Copyright 2009 APA, all rights reserved.
Binary mesh partitioning for cache-efficient visualization.
Tchiboukdjian, Marc; Danjean, Vincent; Raffin, Bruno
2010-01-01
One important bottleneck when visualizing large data sets is the data transfer between processor and memory. Cache-aware (CA) and cache-oblivious (CO) algorithms take into consideration the memory hierarchy to design cache efficient algorithms. CO approaches have the advantage to adapt to unknown and varying memory hierarchies. Recent CA and CO algorithms developed for 3D mesh layouts significantly improve performance of previous approaches, but they lack of theoretical performance guarantees. We present in this paper a {\\schmi O}(N\\log N) algorithm to compute a CO layout for unstructured but well shaped meshes. We prove that a coherent traversal of a N-size mesh in dimension d induces less than N/B+{\\schmi O}(N/M;{1/d}) cache-misses where B and M are the block size and the cache size, respectively. Experiments show that our layout computation is faster and significantly less memory consuming than the best known CO algorithm. Performance is comparable to this algorithm for classical visualization algorithm access patterns, or better when the BSP tree produced while computing the layout is used as an acceleration data structure adjusted to the layout. We also show that cache oblivious approaches lead to significant performance increases on recent GPU architectures.
76 FR 26981 - Proposed Flood Elevation Determinations
Federal Register 2010, 2011, 2012, 2013, 2014
2011-05-10
... table provided here represents the flooding sources, location of referenced elevations, effective and.... Specifically, it addresses the following flooding sources: Cache Creek, Cache Creek Left Bank Overflow, and... ``Unincorporated Areas of Yolo County, California'' addressed the flooding source Cache Creek Settling Basin. That...
Xrootd in dCache - design and experiences
NASA Astrophysics Data System (ADS)
Behrmann, Gerd; Ozerov, Dmitry; Zangerl, Thomas
2011-12-01
dCache is a well established distributed storage solution used in both high energy physics computing and other disciplines. An overview of the implementation of the xrootd data access protocol within dCache is presented. The performance of various access mechanisms is studied and compared and it is concluded that our implementation is as perfomant as other protocols. This makes dCache a compelling alternative to the Scalla software suite implementation of xrootd, with added value from broad protocol support, including the IETF approved NFS 4.1 protocol.
Performance of defect-tolerant set-associative cache memories
NASA Technical Reports Server (NTRS)
Frenzel, J. F.
1991-01-01
The increased use of on-chip cache memories has led researchers to investigate their performance in the presence of manufacturing defects. Several techniques for yield improvement are discussed and results are presented which indicate that set-associativity may be used to provide defect tolerance as well as improve the cache performance. Tradeoffs between several cache organizations and replacement strategies are investigated and it is shown that token-based replacement may be a suitable alternative to the widely-used LRU strategy.
The Effects of Block Size on the Performance of Coherent Caches in Shared-Memory Multiprocessors
1993-05-01
increase with the bandwidth and latency. For those applications with poor spatial locality, the best choice of cache line size is determined by the...observation was used in the design of two schemes: LimitLESS di- rectories and Tag caches. LimitLESS directories [15] were designed for the ALEWIFE...small packets may be used to avoid network congestion. The most important factor influencing the choice of cache line size for a multipro- cessor is the
Pravosudov, Vladimir V; Mendoza, Sally P; Clayton, Nicola S
2003-08-01
It has been hypothesized that in avian social groups subordinate individuals should maintain more energy reserves than dominants, as an insurance against increased perceived risk of starvation. Subordinates might also have elevated baseline corticosterone levels because corticosterone is known to facilitate fattening in birds. Recent experiments showed that moderately elevated corticosterone levels resulting from unpredictable food supply are correlated with enhanced cache retrieval efficiency and more accurate performance on a spatial memory task. Given the correlation between corticosterone and memory, a further prediction is that subordinates might be more efficient at cache retrieval and show more accurate performance on spatial memory tasks. We tested these predictions in dominant-subordinate pairs of mountain chickadees (Poecile gambeli). Each pair was housed in the same cage but caching behavior was tested individually in an adjacent aviary to avoid the confounding effects of small spaces in which birds could unnaturally and directly influence each other's behavior. In sharp contrast to our hypothesis, we found that subordinate chickadees cached less food, showed less efficient cache retrieval, and performed significantly worse on the spatial memory task than dominants. Although the behavioral differences could have resulted from social stress of subordination, and dominant birds reached significantly higher levels of corticosterone during their response to acute stress compared to subordinates, there were no significant differences between dominants and subordinates in baseline levels or in the pattern of adrenocortical stress response. We find no evidence, therefore, to support the hypothesis that subordinate mountain chickadees maintain elevated baseline corticosterone levels whereas lower caching rates and inferior cache retrieval efficiency might contribute to reduced survival of subordinates commonly found in food-caching parids.
AYUSH: A Technique for Extending Lifetime of SRAM-NVM Hybrid Caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Vetter, Jeffrey S
2014-01-01
Recently, researchers have explored way-based hybrid SRAM-NVM (non-volatile memory) last level caches (LLCs) to bring the best of SRAM and NVM together. However, the limited write endurance of NVMs restricts the lifetime of these hybrid caches. We present AYUSH, a technique to enhance the lifetime of hybrid caches, which works by using data-migration to preferentially use SRAM for storing frequently-reused data. Microarchitectural simulations confirm that AYUSH achieves larger improvement in lifetime than a previous technique and also maintains performance and energy efficiency. For single, dual and quad-core workloads, the average increase in cache lifetime with AYUSH is 6.90X, 24.06X andmore » 47.62X, respectively.« less
Agulleiro, Jose-Ignacio; Fernandez, Jose-Jesus
2015-01-01
Cache blocking is a technique widely used in scientific computing to minimize the exchange of information with main memory by reusing the data kept in cache memory. In tomographic reconstruction on standard computers using vector instructions, cache blocking turns out to be central to optimize performance. To this end, sinograms of the tilt-series and slices of the volumes to be reconstructed have to be divided into small blocks that fit into the different levels of cache memory. The code is then reorganized so as to operate with a block as much as possible before proceeding with another one. This data article is related to the research article titled Tomo3D 2.0 – Exploitation of Advanced Vector eXtensions (AVX) for 3D reconstruction (Agulleiro and Fernandez, 2015) [1]. Here we present data of a thorough study of the performance of tomographic reconstruction by varying cache block sizes, which allows derivation of expressions for their automatic quasi-optimal tuning. PMID:26217710
Agulleiro, Jose-Ignacio; Fernandez, Jose-Jesus
2015-06-01
Cache blocking is a technique widely used in scientific computing to minimize the exchange of information with main memory by reusing the data kept in cache memory. In tomographic reconstruction on standard computers using vector instructions, cache blocking turns out to be central to optimize performance. To this end, sinograms of the tilt-series and slices of the volumes to be reconstructed have to be divided into small blocks that fit into the different levels of cache memory. The code is then reorganized so as to operate with a block as much as possible before proceeding with another one. This data article is related to the research article titled Tomo3D 2.0 - Exploitation of Advanced Vector eXtensions (AVX) for 3D reconstruction (Agulleiro and Fernandez, 2015) [1]. Here we present data of a thorough study of the performance of tomographic reconstruction by varying cache block sizes, which allows derivation of expressions for their automatic quasi-optimal tuning.
Temperature and leakage aware techniques to improve cache reliability
NASA Astrophysics Data System (ADS)
Akaaboune, Adil
Decreasing power consumption in small devices such as handhelds, cell phones and high-performance processors is now one of the most critical design concerns. On-chip cache memories dominate the chip area in microprocessors and thus arises the need for power efficient cache memories. Cache is the simplest cost effective method to attain high speed memory hierarchy and, its performance is extremely critical for high speed computers. Cache is used by the microprocessor for channeling the performance gap between processor and main memory (RAM) hence the memory bandwidth is frequently a bottleneck which can affect the peak throughput significantly. In the design of any cache system, the tradeoffs of area/cost, performance, power consumption, and thermal management must be taken into consideration. Previous work has mainly concentrated on performance and area/cost constraints. More recent works have focused on low power design especially for portable devices and media-processing systems, however fewer research has been done on the relationship between heat management, Leakage power and cost per die. Lately, the focus of power dissipation in the new generations of microprocessors has shifted from dynamic power to idle power, a previously underestimated form of power loss that causes battery charge to drain and shutdown too early due the waste of energy. The problem has been aggravated by the aggressive scaling of process; device level method used originally by designers to enhance performance, conserve dissipation and reduces the sizes of digital circuits that are increasingly condensed. This dissertation studies the impact of hotspots, in the cache memory, on leakage consumption and microprocessor reliability and durability. The work will first prove that by eliminating hotspots in the cache memory, leakage power will be reduced and therefore, the reliability will be improved. The second technique studied is data quality management that improves the quality of the data stored in the cache to reduce power consumption. The initial work done on this subject focuses on the type of data that increases leakage consumption and ways to manage without impacting the performance of the microprocessor. The second phase of the project focuses on managing the data storage in different blocks of the cache to smooth the leakage power as well as dynamic power consumption. The last technique is a voltage controlled cache to reduce the leakage consumption of the cache while in execution and even in idle state. Two blocks of the 4-way set associative cache go through a voltage regulator before getting to the voltage well, and the other two are directly connected to the voltage well. The idea behind this technique is to use the replacement algorithm information to increase or decrease voltage of the two blocks depending on the need of the information stored on them.
Karnes, Jason H; Shaffer, Christian M; Cronin, Robert; Bastarache, Lisa; Gaudieri, Silvana; James, Ian; Pavlos, Rebecca; Steiner, Heidi E; Mosley, Jonathan D; Mallal, Simon; Denny, Joshua C; Phillips, Elizabeth J; Roden, Dan M
2017-09-01
Heparin-induced thrombocytopenia (HIT) is an unpredictable, life-threatening, immune-mediated reaction to heparin. Variation in human leukocyte antigen (HLA) genes is now used to prevent immune-mediated adverse drug reactions. Combinations of HLA alleles and killer cell immunoglobulin-like receptors (KIR) are associated with multiple autoimmune diseases and infections. The objective of this study is to evaluate the association of HLA alleles and KIR types, alone or in the presence of different HLA ligands, with HIT. HIT cases and heparin-exposed controls were identified in BioVU, an electronic health record coupled to a DNA biobank. HLA sequencing and KIR type imputation using Illumina OMNI-Quad data were performed. Odds ratios for HLA alleles and KIR types and HLA*KIR interactions using conditional logistic regressions were determined in the overall population and by race/ethnicity. Analysis was restricted to KIR types and HLA alleles with a frequency greater than 0.01. The p values for HLA and KIR association were corrected by using a false discovery rate q<0.05 and HLA*KIR interactions were considered significant at p<0.05. Sixty-five HIT cases and 350 matched controls were identified. No statistical differences in baseline characteristics were observed between cases and controls. The HLA-DRB3*01:01 allele was significantly associated with HIT in the overall population (odds ratio 2.81 [1.57-5.02], p=2.1×10 -4 , q=0.02) and in individuals with European ancestry, independent of other alleles. No KIR types were associated with HIT, although a significant interaction was observed between KIR2DS5 and the HLA-C1 KIR binding group (p=0.03). The HLA-DRB3*01:01 allele was identified as a potential risk factor for HIT. This class II HLA gene and allele represent biologically plausible candidates for influencing HIT pathogenesis. We found limited evidence of the role of KIR types in HIT pathogenesis. Replication and further study of the HLA-DRB3*01:01 association is necessary. © 2017 Pharmacotherapy Publications, Inc.
NELSON, C.; WENGREEN, H.J.; MUNGER, R.G.; CORCORAN, C.D.
2013-01-01
Objective To examine associations between dietary and supplemental folate, vitamin B-12 and vitamin B-6 and incident Alzheimer’s disease (AD) among elderly men and women. Design, Setting and Participants Data collected were from participants of the Cache County Memory, Health and Aging Study, a longitudinal study of 5092 men and women 65 years and older who were residents of Cache County, Utah in 1995. Measurements Multistage clinical assessment procedures were used to identify incident cases of AD. Dietary data were collected using a 142-item food frequency questionnaire. Cox Proportional Hazards (CPH) modeling was used to determine hazard ratios across quintiles of micronutrient intake. Results 202 participants were diagnosed with incident AD during follow-up (1995–2004). In multivariable CPH models that controlled for the effects of gender, age, education, and other covariates there were no observed differences in risk of AD or dementia by increasing quintiles of total intake of folate, vitamin B-12, or vitamin B-6. Similarly, there were no observed differences in risk of AD by regular use of either folate or B6 supplements. Conclusion Dietary intake of B-vitamins from food and supplemental sources appears unrelated to incidence of dementia and AD. Further studies examining associations between dietary intakes of B-vitamins, biomarkers of B-vitamin status and cognitive endpoints are warranted. PMID:19924351
Video head impulse test: a review of the literature.
Alhabib, Salman F; Saliba, Issam
2017-03-01
Video head impulse test (vHIT) is a new testing which able to identify the overt and covert saccades and study the gain of vestibulo-ocular reflex (VOR) of each semicircular canal. The aim of this study is to review the clinical use of vHIT in patients with vestibular disorders in different diseases. PubMed and Cochrane databases were searched for all articles that defined vHIT, compared vHIT with another clinical test, and studied the efficacy of vHIT as diagnostic tools with vestibular disease. 37 articles about vHIT were reviewed. All articles studied the vHIT in English and French languages up to May 2015 were included in the review. Editorial articles or short comments, conference abstracts, animal studies, and language restriction were excluded from the review. Four systems were used in the literature to do the vHIT. vHIT is physiological quick test, which studied the VOR at high frequency of each semicircular canal by calculating the duration ratio between the head impulse and gaze deviation. vHIT is more sensitive than clinical head impulse test (cHIT), especially in patient with isolated covert saccades. vHIT test is diagnostic of vestibular weakness by gain reduction and the appearance of overt and covert saccades. If the vHIT is normal, then caloric test is mandatory to rule out a peripheral origin of vertigo. It is recommended to test each semicircular canal, as isolated vertical canal weakness was identified in the literature. More investigation would be required to determine the evolution of the VOR gain with the progression of the vestibular disease.
Locality in Search Engine Queries and Its Implications for Caching
2001-05-01
in the question of whether caching might be effective for search engines as well. They study two real search engine traces by examining query...locality and its implications for caching. The two search engines studied are Vivisimo and Excite. Their trace analysis results show that queries have
Predictive Caching Using the TDAG Algorithm
NASA Technical Reports Server (NTRS)
Laird, Philip; Saul, Ronald
1992-01-01
We describe how the TDAG algorithm for learning to predict symbol sequences can be used to design a predictive cache store. A model of a two-level mass storage system is developed and used to calculate the performance of the cache under various conditions. Experimental simulations provide good confirmation of the model.
Mammal caching of oak acorns in a red pine and a mixed oak stand
E.R. Thorn; W.M. Tzilkowski
1991-01-01
Small mammal caching of oak (Quercus spp.) acorns in adjacent red pine (Pinus resinosa) and mixed-oak stands was investigated at The Penn State Experimental Forest, Huntingdon Co., Pennsylvania. Gray squirrels (Sciurus carolinensis) and mice (Peromyscus spp.) were the most common acorn-caching...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Evangelinos, Constantinos; Nair, Ravi; Ohmacht, Martin
In one embodiment, a computer-implemented method includes encountering a store operation during a compile-time of a program, where the store operation is applicable to a memory line. It is determined, by a computer processor, that no cache coherence action is necessary for the store operation. A store-without-coherence-action instruction is generated for the store operation, responsive to determining that no cache coherence action is necessary. The store-without-coherence-action instruction specifies that the store operation is to be performed without a cache coherence action, and cache coherence is maintained upon execution of the store-without-coherence-action instruction.
Population substructure in Cache County, Utah: the Cache County study
2014-01-01
Background Population stratification is a key concern for genetic association analyses. In addition, extreme homogeneity of ethnic origins of a population can make it difficult to interpret how genetic associations in that population may translate into other populations. Here we have evaluated the genetic substructure of samples from the Cache County study relative to the HapMap Reference populations and data from the Alzheimer's Disease Neuroimaging Initiative (ADNI). Results Our findings show that the Cache County study is similar in ethnic diversity to the self-reported "Whites" in the ADNI sample and less homogenous than the HapMap CEU population. Conclusions We conclude that the Cache County study is genetically representative of the general European American population in the USA and is an appropriate population for conducting broadly applicable genetic studies. PMID:25078123
EqualWrites: Reducing Intra-set Write Variations for Enhancing Lifetime of Non-volatile Caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Vetter, Jeffrey S.
Driven by the trends of increasing core-count and bandwidth-wall problem, the size of last level caches (LLCs) has greatly increased and hence, the researchers have explored non-volatile memories (NVMs) which provide high density and consume low-leakage power. Since NVMs have low write-endurance and the existing cache management policies are write variation-unaware, effective wear-leveling techniques are required for achieving reasonable cache lifetimes using NVMs. We present EqualWrites, a technique for mitigating intra-set write variation. In this paper, our technique works by recording the number of writes on a block and changing the cache-block location of a hot data-item to redirect themore » future writes to a cold block to achieve wear-leveling. Simulation experiments have been performed using an x86-64 simulator and benchmarks from SPEC06 and HPC (high-performance computing) field. The results show that for single, dual and quad-core system configurations, EqualWrites improves cache lifetime by 6.31X, 8.74X and 10.54X, respectively. In addition, its implementation overhead is very small and it provides larger improvement in lifetime than three other intra-set wear-leveling techniques and a cache replacement policy.« less
Improving energy efficiency of Embedded DRAM Caches for High-end Computing Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Vetter, Jeffrey S; Li, Dong
2014-01-01
With increasing system core-count, the size of last level cache (LLC) has increased and since SRAM consumes high leakage power, power consumption of LLCs is becoming a significant fraction of processor power consumption. To address this, researchers have used embedded DRAM (eDRAM) LLCs which consume low-leakage power. However, eDRAM caches consume a significant amount of energy in the form of refresh energy. In this paper, we propose ESTEEM, an energy saving technique for embedded DRAM caches. ESTEEM uses dynamic cache reconfiguration to turn-off a portion of the cache to save both leakage and refresh energy. It logically divides the cachemore » sets into multiple modules and turns-off possibly different number of ways in each module. Microarchitectural simulations confirm that ESTEEM is effective in improving performance and energy efficiency and provides better results compared to a recently-proposed eDRAM cache energy saving technique, namely Refrint. For single and dual-core simulations, the average saving in memory subsystem (LLC+main memory) on using ESTEEM is 25.8% and 32.6%, respectively and average weighted speedup are 1.09X and 1.22X, respectively. Additional experiments confirm that ESTEEM works well for a wide-range of system parameters.« less
EqualWrites: Reducing Intra-set Write Variations for Enhancing Lifetime of Non-volatile Caches
Mittal, Sparsh; Vetter, Jeffrey S.
2015-01-29
Driven by the trends of increasing core-count and bandwidth-wall problem, the size of last level caches (LLCs) has greatly increased and hence, the researchers have explored non-volatile memories (NVMs) which provide high density and consume low-leakage power. Since NVMs have low write-endurance and the existing cache management policies are write variation-unaware, effective wear-leveling techniques are required for achieving reasonable cache lifetimes using NVMs. We present EqualWrites, a technique for mitigating intra-set write variation. In this paper, our technique works by recording the number of writes on a block and changing the cache-block location of a hot data-item to redirect themore » future writes to a cold block to achieve wear-leveling. Simulation experiments have been performed using an x86-64 simulator and benchmarks from SPEC06 and HPC (high-performance computing) field. The results show that for single, dual and quad-core system configurations, EqualWrites improves cache lifetime by 6.31X, 8.74X and 10.54X, respectively. In addition, its implementation overhead is very small and it provides larger improvement in lifetime than three other intra-set wear-leveling techniques and a cache replacement policy.« less
Choi, Ji Eun; Kim, Yi-Kyung; Cho, Young Sang; Lee, Kieun; Park, Hyun Woo; Yoon, Sung Hoon; Kim, Hyung-Jin; Chung, Won-Ho
2017-01-01
The purpose of this study was to prove the hypothesis that caloric response in Ménière's disease (MD) is reduced by hydropic expansion of the vestibular labyrinth, not by vestibular hypofunction, by evaluating the correlation morphologically using an intravenous Gadolinium (IV-Gd) inner ear MRI. In study I, the prevalence of abnormal video Head Impulse Test (vHIT) results among the patients with definite unilateral MD (n = 24) and vestibular neuritis (VN) (n = 22) were investigated. All patients showed abnormal canal paresis (CP) (> 26%) on caloric tests. The prevalence of abnormal vHIT in patients with abnormal CP was significantly lower in MD patients (12.5%) than that in VN patients (81.8%) (p < 0.001). In study II, morphological correlation between caloric tests and vestibular hydrops level was evaluated in unilateral MD patients (n = 16) who had normal vHIT results. Eleven patients (61%) had abnormal CP. After taking the images of IV-Gd inner ear MRI, the vestibular hydrops ratio (endolymph volume/total lymph volume = %VH) was measured. In addition, the relative vestibular hydrops ratio (%RVH = (%VHaffected ear-%VHunaffected ear) / (%VHaffected ear + %VHunaffected ear)) was calculated. Each ratio (%VH and %RVH) was compared with average peak slow phase velocity (PSPV) and CP, respectively. In the MD patients, %VH of the affected ear correlated significantly with mean PSPV on the same side (rs = -0.569, p = 0.024), while %RVH correlated significantly with CP (rs = 0.602, p = 0.014). In most MD patients (87.5%) compared to VN patients, vHIT results were normal even though the caloric function was reduced. In addition, the reduced caloric function with normal vHIT was related to the severity of the vestibular hydrops measured by the IV-Gd inner ear MRI. These findings concluded that the abnormal caloric tests with normal vHIT in MD indicated severe endolymphatic hydrops rather than vestibular hypofunction.
Winter prey caching by northern hawk owls in Minnesota
Richard R. Schaefer; D. Craig Rudolph; Jesse F. Fagan
2007-01-01
Northern Hawk Owls (Surnia ulula) have been reported to cache prey during the breeding season for later consumption, but detailed reports of prey caching during the non-breeding season are comparatively rare. We provided prey to four individual Northern Hawk Owls in wintering areas in northeastern Minnesota during 2001 and 2005 and observed their...
A measurement-based study of concurrency in a multiprocessor
NASA Technical Reports Server (NTRS)
Mcguire, Patrick John
1987-01-01
A systematic measurement-based methodology for characterizing the amount of concurrency present in a workload, and the effect of concurrency on system performance indices such as cache miss rate and bus activity are developed. Hardware and software instrumentation of an Alliant FX/8 was used to obtain data from a real workload environment. Results show that 35% of the workload is concurrent, with the concurrent periods typically using all available processors. Measurements of periods of change in concurrency show uneven usage of processors during these times. Other system measures, including cache miss rate and processor bus activity, are analyzed with respect to the concurrency measures. Probability of a cache miss is seen to increase with concurrency. The change in cache miss rate is much more sensitive to the fraction of concurrent code in the worklaod than the number of processors active during concurrency. Regression models are developed to quantify the relationships between cache miss rate, bus activity, and the concurrency measures. The model for cache miss rate predicts an increase in the median miss rate value as much as 300% for a 100% increase in concurrency in the workload.
NASA Astrophysics Data System (ADS)
Fang, Juan; Hao, Xiaoting; Fan, Qingwen; Chang, Zeqing; Song, Shuying
2017-05-01
In the Heterogeneous multi-core architecture, CPU and GPU processor are integrated on the same chip, which poses a new challenge to the last-level cache management. In this architecture, the CPU application and the GPU application execute concurrently, accessing the last-level cache. CPU and GPU have different memory access characteristics, so that they have differences in the sensitivity of last-level cache (LLC) capacity. For many CPU applications, a reduced share of the LLC could lead to significant performance degradation. On the contrary, GPU applications can tolerate increase in memory access latency when there is sufficient thread-level parallelism. Taking into account the GPU program memory latency tolerance characteristics, this paper presents a method that let GPU applications can access to memory directly, leaving lots of LLC space for CPU applications, in improving the performance of CPU applications and does not affect the performance of GPU applications. When the CPU application is cache sensitive, and the GPU application is insensitive to the cache, the overall performance of the system is improved significantly.
Episodic-like memory during cache recovery by scrub jays.
Clayton, N S; Dickinson, A
1998-09-17
The recollection of past experiences allows us to recall what a particular event was, and where and when it occurred, a form of memory that is thought to be unique to humans. It is known, however, that food-storing birds remember the spatial location and contents of their caches. Furthermore, food-storing animals adapt their caching and recovery strategies to the perishability of food stores, which suggests that they are sensitive to temporal factors. Here we show that scrub jays (Aphelocoma coerulescens) remember 'when' food items are stored by allowing them to recover perishable 'wax worms' (wax-moth larvae) and non-perishable peanuts which they had previously cached in visuospatially distinct sites. Jays searched preferentially for fresh wax worms, their favoured food, when allowed to recover them shortly after caching. However, they rapidly learned to avoid searching for worms after a longer interval during which the worms had decayed. The recovery preference of jays demonstrates memory of where and when particular food items were cached, thereby fulfilling the behavioural criteria for episodic-like memory in non-human animals.
Ordering of guarded and unguarded stores for no-sync I/O
Gara, Alan; Ohmacht, Martin
2013-06-25
A parallel computing system processes at least one store instruction. A first processor core issues a store instruction. A first queue, associated with the first processor core, stores the store instruction. A second queue, associated with a first local cache memory device of the first processor core, stores the store instruction. The first processor core updates first data in the first local cache memory device according to the store instruction. The third queue, associated with at least one shared cache memory device, stores the store instruction. The first processor core invalidates second data, associated with the store instruction, in the at least one shared cache memory. The first processor core invalidates third data, associated with the store instruction, in other local cache memory devices of other processor cores. The first processor core flushing only the first queue.
Pravosudov, Vladimir V; Clayton, Nicola S
2002-08-01
To test the hypothesis that accurate cache recovery is more critical for birds that live in harsh conditions where the food supply is limited and unpredictable, the authors compared food caching, memory, and the hippocampus of black-capped chickadees (Poecile atricapilla) from Alaska and Colorado. Under identical laboratory conditions, Alaska chickadees (a) cached significantly more food; (b) were more efficient at cache recovery: (c) performed more accurately on one-trial associative learning tasks in which birds had to rely on spatial memory, but did not differ when tested on a nonspatial version of this task; and (d) had significantly larger hippocampal volumes containing more neurons compared with Colorado chickadees. The results support the hypothesis that these population differences may reflect adaptations to a harsh environment.
Accurate low-cost methods for performance evaluation of cache memory systems
NASA Technical Reports Server (NTRS)
Laha, Subhasis; Patel, Janak H.; Iyer, Ravishankar K.
1988-01-01
Methods of simulation based on statistical techniques are proposed to decrease the need for large trace measurements and for predicting true program behavior. Sampling techniques are applied while the address trace is collected from a workload. This drastically reduces the space and time needed to collect the trace. Simulation techniques are developed to use the sampled data not only to predict the mean miss rate of the cache, but also to provide an empirical estimate of its actual distribution. Finally, a concept of primed cache is introduced to simulate large caches by the sampling-based method.
Ramírez-Vélez, Robinson; Tordecilla-Sanders, Alejandra; Téllez-T, Luis Andrés; Camelo-Prieto, Diana; Hernández-Quiñonez, Paula Andrea; Correa-Bautista, Jorge Enrique; Garcia-Hermoso, Antonio; Ramírez-Campillo, Rodrigo; Izquierdo, Mikel
2017-02-01
We investigated the effect of moderate versus high-intensity interval exercise training on the HRV indices in physically inactive adults. Twenty inactive adults were randomly allocated to receive either moderate intensity training (MCT group) or high-intensity interval training (HIT group). The MCT group performed aerobic training at an intensity of 55-75%, which consisted of walking on a treadmill at 60-80% of the maximum heart rate (HRmax) until the expenditure of 300 kcal. The HIT group ran on a treadmill for 4 minutes at 85-95% peak HRmax and had a recovery of 4 minutes at 65% peak HRmax until the expenditure of 300 kcal. Supine resting HRV indices (time domain: SDNN, standard deviation of normal-to-normal intervals; rMSSD, Root mean square successive difference of RR intervals and frequency domain: HFLn, high-frequency spectral power; LF, low-frequency spectral power and HF/LF ratio) were measured at baseline and 12 weeks thereafter. The SDNN changes were 3.4 (8.9) ms in the MCT group and 29.1 (7.6) ms in the HIT group (difference between groups 32.6 [95% CI, 24.9 to 40.4 (P = 0.01)]. The LF/HFLn ratio change 0.19 (0.03) ms in the MCT group and 0.13 (0.01) ms in the HIT group (P between groups = 0.016). No significant group differences were observed for the rMSSD, HF and LF parameters. In inactive adults, this study showed that a 12-week HIT training program could increase short-term HRV, mostly in vagally mediated indices such as SDNN and HF/LFLn ratio power.
dCache on Steroids - Delegated Storage Solutions
Mkrtchyan, Tigran; Adeyemi, F.; Ashish, A.; ...
2017-11-23
For over a decade, dCache.org has delivered a robust software used at more than 80 Universities and research institutes around the world, allowing these sites to provide reliable storage services for the WLCG experiments as well as many other scientific communities. The flexible architecture of dCache allows running it in a wide variety of configurations and platforms - from a SoC based all-in-one Raspberry-Pi up to hundreds of nodes in a multipetabyte installation. Due to lack of managed storage at the time, dCache implemented data placement, replication and data integrity directly. Today, many alternatives are available: S3, GlusterFS, CEPH andmore » others. While such solutions position themselves as scalable storage systems, they cannot be used by many scientific communities out of the box. The absence of community-accepted authentication and authorization mechanisms, the use of product specific protocols and the lack of namespace are some of the reasons that prevent wide-scale adoption of these alternatives. Most of these limitations are already solved by dCache. By delegating low-level storage management functionality to the above-mentioned new systems and providing the missing layer through dCache, we provide a solution which combines the benefits of both worlds - industry standard storage building blocks with the access protocols and authentication required by scientific communities. In this paper, we focus on CEPH, a popular software for clustered storage that supports file, block and object interfaces. CEPH is often used in modern computing centers, for example as a backend to OpenStack services. We will show prototypes of dCache running with a CEPH backend and discuss the benefits and limitations of such an approach. As a result, we will also outline the roadmap for supporting ‘delegated storage’ within the dCache releases.« less
dCache on Steroids - Delegated Storage Solutions
NASA Astrophysics Data System (ADS)
Mkrtchyan, T.; Adeyemi, F.; Ashish, A.; Behrmann, G.; Fuhrmann, P.; Litvintsev, D.; Millar, P.; Rossi, A.; Sahakyan, M.; Starek, J.
2017-10-01
For over a decade, dCache.org has delivered a robust software used at more than 80 Universities and research institutes around the world, allowing these sites to provide reliable storage services for the WLCG experiments as well as many other scientific communities. The flexible architecture of dCache allows running it in a wide variety of configurations and platforms - from a SoC based all-in-one Raspberry-Pi up to hundreds of nodes in a multipetabyte installation. Due to lack of managed storage at the time, dCache implemented data placement, replication and data integrity directly. Today, many alternatives are available: S3, GlusterFS, CEPH and others. While such solutions position themselves as scalable storage systems, they cannot be used by many scientific communities out of the box. The absence of community-accepted authentication and authorization mechanisms, the use of product specific protocols and the lack of namespace are some of the reasons that prevent wide-scale adoption of these alternatives. Most of these limitations are already solved by dCache. By delegating low-level storage management functionality to the above-mentioned new systems and providing the missing layer through dCache, we provide a solution which combines the benefits of both worlds - industry standard storage building blocks with the access protocols and authentication required by scientific communities. In this paper, we focus on CEPH, a popular software for clustered storage that supports file, block and object interfaces. CEPH is often used in modern computing centers, for example as a backend to OpenStack services. We will show prototypes of dCache running with a CEPH backend and discuss the benefits and limitations of such an approach. We will also outline the roadmap for supporting ‘delegated storage’ within the dCache releases.
dCache on Steroids - Delegated Storage Solutions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mkrtchyan, Tigran; Adeyemi, F.; Ashish, A.
For over a decade, dCache.org has delivered a robust software used at more than 80 Universities and research institutes around the world, allowing these sites to provide reliable storage services for the WLCG experiments as well as many other scientific communities. The flexible architecture of dCache allows running it in a wide variety of configurations and platforms - from a SoC based all-in-one Raspberry-Pi up to hundreds of nodes in a multipetabyte installation. Due to lack of managed storage at the time, dCache implemented data placement, replication and data integrity directly. Today, many alternatives are available: S3, GlusterFS, CEPH andmore » others. While such solutions position themselves as scalable storage systems, they cannot be used by many scientific communities out of the box. The absence of community-accepted authentication and authorization mechanisms, the use of product specific protocols and the lack of namespace are some of the reasons that prevent wide-scale adoption of these alternatives. Most of these limitations are already solved by dCache. By delegating low-level storage management functionality to the above-mentioned new systems and providing the missing layer through dCache, we provide a solution which combines the benefits of both worlds - industry standard storage building blocks with the access protocols and authentication required by scientific communities. In this paper, we focus on CEPH, a popular software for clustered storage that supports file, block and object interfaces. CEPH is often used in modern computing centers, for example as a backend to OpenStack services. We will show prototypes of dCache running with a CEPH backend and discuss the benefits and limitations of such an approach. As a result, we will also outline the roadmap for supporting ‘delegated storage’ within the dCache releases.« less
The Optimization of In-Memory Space Partitioning Trees for Cache Utilization
NASA Astrophysics Data System (ADS)
Yeo, Myung Ho; Min, Young Soo; Bok, Kyoung Soo; Yoo, Jae Soo
In this paper, a novel cache conscious indexing technique based on space partitioning trees is proposed. Many researchers investigated efficient cache conscious indexing techniques which improve retrieval performance of in-memory database management system recently. However, most studies considered data partitioning and targeted fast information retrieval. Existing data partitioning-based index structures significantly degrade performance due to the redundant accesses of overlapped spaces. Specially, R-tree-based index structures suffer from the propagation of MBR (Minimum Bounding Rectangle) information by updating data frequently. In this paper, we propose an in-memory space partitioning index structure for optimal cache utilization. The proposed index structure is compared with the existing index structures in terms of update performance, insertion performance and cache-utilization rate in a variety of environments. The results demonstrate that the proposed index structure offers better performance than existing index structures.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gala, Alan; Ohmacht, Martin
A multiprocessor system includes nodes. Each node includes a data path that includes a core, a TLB, and a first level cache implementing disambiguation. The system also includes at least one second level cache and a main memory. For thread memory access requests, the core uses an address associated with an instruction format of the core. The first level cache uses an address format related to the size of the main memory plus an offset corresponding to hardware thread meta data. The second level cache uses a physical main memory address plus software thread meta data to store the memorymore » access request. The second level cache accesses the main memory using the physical address with neither the offset nor the thread meta data after resolving speculation. In short, this system includes mapping of a virtual address to a different physical addresses for value disambiguation for different threads.« less
Kowalska, M Anna; Krishnaswamy, Sriram; Rauova, Lubica; Zhai, Li; Hayes, Vincent; Amirikian, Karine; Esko, Jeffrey D; Bougie, Daniel W; Aster, Richard H; Cines, Douglas B; Poncz, Mortimer
2011-09-08
Heparin-induced thrombocytopenia (HIT) is caused by antibodies that recognize complexes between platelet factor 4 (PF4) and heparin or glycosaminoglycan side chains. These antibodies can lead to a limb- and life-threatening prothrombotic state. We now show that HIT antibodies are able to inhibit generation of activated protein C (aPC) by thrombin/thrombomodulin (IIa/TM) in the presence of PF4. Tetrameric PF4 potentiates aPC generation by formation of complexes with chondroitin sulfate (CS) on TM. Formation of these complexes occurs at a specific molar ratio of PF4 to glycosaminoglycan. This observation and the finding that the effect of heparin on aPC generation depends on the concentration of PF4 suggest similarity between PF4/CS complexes and those that bind HIT antibodies. HIT antibodies reduced the ability of PF4 to augment aPC formation. Cationic protamine sulfate, which forms similar complexes with heparin, also enhanced aPC generation, but its activity was not blocked by HIT antibodies. Our studies provide evidence that complexes formed between PF4 and TM's CS may play a physiologic role in potentiating aPC generation. Recognition of these complexes by HIT antibodies reverses the PF4-dependent enhancement in aPC generation and may contribute to the prothrombotic nature of HIT.
Balásházy, Imre; Farkas, Arpád; Madas, Balázs Gergely; Hofmann, Werner
2009-06-01
Cellular hit probabilities of alpha particles emitted by inhaled radon progenies in sensitive bronchial epithelial cell nuclei were simulated at low exposure levels to obtain useful data for the rejection or support of the linear-non-threshold (LNT) hypothesis. In this study, local distributions of deposited inhaled radon progenies in airway bifurcation models were computed at exposure conditions characteristic of homes and uranium mines. Then, maximum local deposition enhancement factors at bronchial airway bifurcations, expressed as the ratio of local to average deposition densities, were determined to characterise the inhomogeneity of deposition and to elucidate their effect on resulting hit probabilities. The results obtained suggest that in the vicinity of the carinal regions of the central airways the probability of multiple hits can be quite high, even at low average doses. Assuming a uniform distribution of activity there are practically no multiple hits and the hit probability as a function of dose exhibits a linear shape in the low dose range. The results are quite the opposite in the case of hot spots revealed by realistic deposition calculations, where practically all cells receive multiple hits and the hit probability as a function of dose is non-linear in the average dose range of 10-100 mGy.
Cache directory lookup reader set encoding for partial cache line speculation support
Gara, Alan; Ohmacht, Martin
2014-10-21
In a multiprocessor system, with conflict checking implemented in a directory lookup of a shared cache memory, a reader set encoding permits dynamic recordation of read accesses. The reader set encoding includes an indication of a portion of a line read, for instance by indicating boundaries of read accesses. Different encodings may apply to different types of speculative execution.
Nelson, Michael E.
2011-01-01
A single Gray Wolf (Canis lupus) killed an adult male White-tailed Deer (Odocoileus virginianus) and cached the intact carcass in 76 cm of snow. The carcass was revisited and entirely consumed between four and seven days later. This is the first recorded observation of a Gray Wolf caching an entire adult deer.
Formal verification of an MMU and MMU cache
NASA Technical Reports Server (NTRS)
Schubert, E. T.
1991-01-01
We describe the formal verification of a hardware subsystem consisting of a memory management unit and a cache. These devices are verified independently and then shown to interact correctly when composed. The MMU authorizes memory requests and translates virtual addresses to real addresses. The cache improves performance by maintaining a LRU (least recently used) list from the memory resident segment table.
NASA Astrophysics Data System (ADS)
Pleros, Nikos; Maniotis, Pavlos; Alexoudi, Theonitsa; Fitsios, Dimitris; Vagionas, Christos; Papaioannou, Sotiris; Vyrsokinos, K.; Kanellos, George T.
2014-03-01
The processor-memory performance gap, commonly referred to as "Memory Wall" problem, owes to the speed mismatch between processor and electronic RAM clock frequencies, forcing current Chip Multiprocessor (CMP) configurations to consume more than 50% of the chip real-estate for caching purposes. In this article, we present our recent work spanning from Si-based integrated optical RAM cell architectures up to complete optical cache memory architectures for Chip Multiprocessor configurations. Moreover, we discuss on e/o router subsystems with up to Tb/s routing capacity for cache interconnection purposes within CMP configurations, currently pursued within the FP7 PhoxTrot project.
A search game model of the scatter hoarder's problem
Alpern, Steve; Fokkink, Robbert; Lidbetter, Thomas; Clayton, Nicola S.
2012-01-01
Scatter hoarders are animals (e.g. squirrels) who cache food (nuts) over a number of sites for later collection. A certain minimum amount of food must be recovered, possibly after pilfering by another animal, in order to survive the winter. An optimal caching strategy is one that maximizes the survival probability, given worst case behaviour of the pilferer. We modify certain ‘accumulation games’ studied by Kikuta & Ruckle (2000 J. Optim. Theory Appl.) and Kikuta & Ruckle (2001 Naval Res. Logist.), which modelled the problem of optimal diversification of resources against catastrophic loss, to include the depth at which the food is hidden at each caching site. Optimal caching strategies can then be determined as equilibria in a new ‘caching game’. We show how the distribution of food over sites and the site-depths of the optimal caching varies with the animal's survival requirements and the amount of pilfering. We show that in some cases, ‘decoy nuts’ are required to be placed above other nuts that are buried further down at the same site. Methods from the field of search games are used. Some empirically observed behaviour can be shown to be optimal in our model. PMID:22012971
Image matrix processor for fast multi-dimensional computations
Roberson, George P.; Skeate, Michael F.
1996-01-01
An apparatus for multi-dimensional computation which comprises a computation engine, including a plurality of processing modules. The processing modules are configured in parallel and compute respective contributions to a computed multi-dimensional image of respective two dimensional data sets. A high-speed, parallel access storage system is provided which stores the multi-dimensional data sets, and a switching circuit routes the data among the processing modules in the computation engine and the storage system. A data acquisition port receives the two dimensional data sets representing projections through an image, for reconstruction algorithms such as encountered in computerized tomography. The processing modules include a programmable local host, by which they may be configured to execute a plurality of different types of multi-dimensional algorithms. The processing modules thus include an image manipulation processor, which includes a source cache, a target cache, a coefficient table, and control software for executing image transformation routines using data in the source cache and the coefficient table and loading resulting data in the target cache. The local host processor operates to load the source cache with a two dimensional data set, loads the coefficient table, and transfers resulting data out of the target cache to the storage system, or to another destination.
Joshua tree (Yucca brevifolia) seeds are dispersed by seed-caching rodents
Vander Wall, S.B.; Esque, T.; Haines, D.; Garnett, M.; Waitman, B.A.
2006-01-01
Joshua tree (Yucca brevifolia) is a distinctive and charismatic plant of the Mojave Desert. Although floral biology and seed production of Joshua tree and other yuccas are well understood, the fate of Joshua tree seeds has never been studied. We tested the hypothesis that Joshua tree seeds are dispersed by seed-caching rodents. We radioactively labelled Joshua tree seeds and followed their fates at five source plants in Potosi Wash, Clark County, Nevada, USA. Rodents made a mean of 30.6 caches, usually within 30 m of the base of source plants. Caches contained a mean of 5.2 seeds buried 3-30 nun deep. A variety of rodent species appears to have prepared the caches. Three of the 836 Joshua tree seeds (0.4%) cached germinated the following spring. Seed germination using rodent exclosures was nearly 15%. More than 82% of seeds in open plots were removed by granivores, and neither microsite nor supplemental water significantly affected germination. Joshua tree produces seeds in indehiscent pods or capsules, which rodents dismantle to harvest seeds. Because there is no other known means of seed dispersal, it is possible that the Joshua tree-rodent seed dispersal interaction is an obligate mutualism for the plant.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shoopman, J. D.
This report documents Livermore Computing (LC) activities in support of ASC L2 milestone 5589: Modernization and Expansion of LLNL Archive Disk Cache, due March 31, 2016. The full text of the milestone is included in Attachment 1. The description of the milestone is: Description: Configuration of archival disk cache systems will be modernized to reduce fragmentation, and new, higher capacity disk subsystems will be deployed. This will enhance archival disk cache capability for ASC archive users, enabling files written to the archives to remain resident on disk for many (6–12) months, regardless of file size. The milestone was completed inmore » three phases. On August 26, 2015 subsystems with 6PB of disk cache were deployed for production use in LLNL’s unclassified HPSS environment. Following that, on September 23, 2015 subsystems with 9 PB of disk cache were deployed for production use in LLNL’s classified HPSS environment. On January 31, 2016, the milestone was fully satisfied when the legacy Data Direct Networks (DDN) archive disk cache subsystems were fully retired from production use in both LLNL’s unclassified and classified HPSS environments, and only the newly deployed systems were in use.« less
NASA Astrophysics Data System (ADS)
Dykstra, D.; Bockelman, B.; Blomer, J.; Herner, K.; Levshina, T.; Slyz, M.
2015-12-01
A common use pattern in the computing models of particle physics experiments is running many distributed applications that read from a shared set of data files. We refer to this data is auxiliary data, to distinguish it from (a) event data from the detector (which tends to be different for every job), and (b) conditions data about the detector (which tends to be the same for each job in a batch of jobs). Relatively speaking, conditions data also tends to be relatively small per job where both event data and auxiliary data are larger per job. Unlike event data, auxiliary data comes from a limited working set of shared files. Since there is spatial locality of the auxiliary data access, the use case appears to be identical to that of the CernVM- Filesystem (CVMFS). However, we show that distributing auxiliary data through CVMFS causes the existing CVMFS infrastructure to perform poorly. We utilize a CVMFS client feature called "alien cache" to cache data on existing local high-bandwidth data servers that were engineered for storing event data. This cache is shared between the worker nodes at a site and replaces caching CVMFS files on both the worker node local disks and on the site's local squids. We have tested this alien cache with the dCache NFSv4.1 interface, Lustre, and the Hadoop Distributed File System (HDFS) FUSE interface, and measured performance. In addition, we use high-bandwidth data servers at central sites to perform the CVMFS Stratum 1 function instead of the low-bandwidth web servers deployed for the CVMFS software distribution function. We have tested this using the dCache HTTP interface. As a result, we have a design for an end-to-end high-bandwidth distributed caching read-only filesystem, using existing client software already widely deployed to grid worker nodes and existing file servers already widely installed at grid sites. Files are published in a central place and are soon available on demand throughout the grid and cached locally on the site with a convenient POSIX interface. This paper discusses the details of the architecture and reports performance measurements.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dykstra, D.; Bockelman, B.; Blomer, J.
A common use pattern in the computing models of particle physics experiments is running many distributed applications that read from a shared set of data files. We refer to this data is auxiliary data, to distinguish it from (a) event data from the detector (which tends to be different for every job), and (b) conditions data about the detector (which tends to be the same for each job in a batch of jobs). Relatively speaking, conditions data also tends to be relatively small per job where both event data and auxiliary data are larger per job. Unlike event data, auxiliarymore » data comes from a limited working set of shared files. Since there is spatial locality of the auxiliary data access, the use case appears to be identical to that of the CernVM- Filesystem (CVMFS). However, we show that distributing auxiliary data through CVMFS causes the existing CVMFS infrastructure to perform poorly. We utilize a CVMFS client feature called 'alien cache' to cache data on existing local high-bandwidth data servers that were engineered for storing event data. This cache is shared between the worker nodes at a site and replaces caching CVMFS files on both the worker node local disks and on the site's local squids. We have tested this alien cache with the dCache NFSv4.1 interface, Lustre, and the Hadoop Distributed File System (HDFS) FUSE interface, and measured performance. In addition, we use high-bandwidth data servers at central sites to perform the CVMFS Stratum 1 function instead of the low-bandwidth web servers deployed for the CVMFS software distribution function. We have tested this using the dCache HTTP interface. As a result, we have a design for an end-to-end high-bandwidth distributed caching read-only filesystem, using existing client software already widely deployed to grid worker nodes and existing file servers already widely installed at grid sites. Files are published in a central place and are soon available on demand throughout the grid and cached locally on the site with a convenient POSIX interface. This paper discusses the details of the architecture and reports performance measurements.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Smith, Tyler Barratt; Urrea, Jorge Mario
2012-06-01
The aim of the Authenticating Cache architecture is to ensure that machine instructions in a Read Only Memory (ROM) are legitimate from the time the ROM image is signed (immediately after compilation) to the time they are placed in the cache for the processor to consume. The proposed architecture allows the detection of ROM image modifications during distribution or when it is loaded into memory. It also ensures that modified instructions will not execute in the processor-as the cache will not be loaded with a page that fails an integrity check. The authenticity of the instruction stream can also bemore » verified in this architecture. The combination of integrity and authenticity assurance greatly improves the security profile of a system.« less
Replication Strategy for Spatiotemporal Data Based on Distributed Caching System
Xiong, Lian; Tao, Yang; Xu, Juan; Zhao, Lun
2018-01-01
The replica strategy in distributed cache can effectively reduce user access delay and improve system performance. However, developing a replica strategy suitable for varied application scenarios is still quite challenging, owing to differences in user access behavior and preferences. In this paper, a replication strategy for spatiotemporal data (RSSD) based on a distributed caching system is proposed. By taking advantage of the spatiotemporal locality and correlation of user access, RSSD mines high popularity and associated files from historical user access information, and then generates replicas and selects appropriate cache node for placement. Experimental results show that the RSSD algorithm is simple and efficient, and succeeds in significantly reducing user access delay. PMID:29342897
Error recovery in shared memory multiprocessors using private caches
NASA Technical Reports Server (NTRS)
Wu, Kun-Lung; Fuchs, W. Kent; Patel, Janak H.
1990-01-01
The problem of recovering from processor transient faults in shared memory multiprocesses systems is examined. A user-transparent checkpointing and recovery scheme using private caches is presented. Processes can recover from errors due to faulty processors by restarting from the checkpointed computation state. Implementation techniques using checkpoint identifiers and recovery stacks are examined as a means of reducing performance degradation in processor utilization during normal execution. This cache-based checkpointing technique prevents rollback propagation, provides rapid recovery, and can be integrated into standard cache coherence protocols. An analytical model is used to estimate the relative performance of the scheme during normal execution. Extensions to take error latency into account are presented.
Weng, Tzu-Pin; Huang, Shu-Chun; Chuang, Yu-Fen; Wang, Jong-Shyan
2013-01-01
Exercise is linked with the type/intensity-dependent adaptive immune responses, whereas hypoxic stress facilitates the programmed death of CD4 lymphocytes. This study investigated how high intensity-interval (HIT) and moderate intensity-continuous (MCT) exercise training influence hypoxia-induced apoptosis and autophagy of CD4 lymphocytes in sedentary men. Thirty healthy sedentary males were randomized to engage either HIT (3-minute intervals at 40% and 80%VO2max, n=10) or MCT (sustained 60%VO2max, n=10) for 30 minutes/day, 5 days/week for 5 weeks, or to a control group that did not received exercise intervention (CTL, n=10). CD4 lymphocyte apoptotic and autophagic responses to hypoxic exercise (HE, 100 W under 12%O2 for 30 minutes) were determined before and after various regimens. The results demonstrated that HIT exhibited higher enhancements of pulmonary ventilation, cardiac output, and VO2 at ventilatory threshold and peak performance than MCT did. Before the intervention, HE significantly down-regulated autophagy by decreased beclin-1, Atg-1, LC3-II, Atg-12, and LAMP-2 expressions and acridine orange staining, and simultaneously enhanced apoptosis by increased phospho-Bcl-2 and active caspase-9/-3 levels and phosphotidylserine exposure in CD4 lymphocytes. However, five weeks of HIT and MCT, but not CTL, reduced the extents of declined autophagy and potentiated apoptosis in CD4 lymphocytes caused by HE. Furthermore, both HIT and MCT regimens manifestly lowered plasma myeloperoxidase and interleukin-4 levels and elevated the ratio of interleukin-4 to interferon-γ at rest and following HE. Therefore, we conclude that HIT is superior to MCT for enhancing aerobic fitness. Moreover, either HIT or MCT effectively depresses apoptosis and promotes autophagy in CD4 lymphocytes and is accompanied by increased interleukin-4/interferon-γ ratio and decreased peroxide production during HE.
Magpies can use local cues to retrieve their food caches.
Feenders, Gesa; Smulders, Tom V
2011-03-01
Much importance has been placed on the use of spatial cues by food-hoarding birds in the retrieval of their caches. In this study, we investigate whether food-hoarding birds can be trained to use local cues ("beacons") in their cache retrieval. We test magpies (Pica pica) in an active hoarding-retrieval paradigm, where local cues are always reliable, while spatial cues are not. Our results show that the birds use the local cues to retrieve their caches, even when occasionally contradicting spatial information is available. The design of our study does not allow us to test rigorously whether the birds prefer using local over spatial cues, nor to investigate the process through which they learn to use local cues. We furthermore provide evidence that magpies develop landmark preferences, which improve their retrieval accuracy. Our findings support the hypothesis that birds are flexible in their use of memory information, using a combination of the most reliable or salient information to retrieve their caches. © Springer-Verlag 2010
A Survey Of Architectural Approaches for Managing Embedded DRAM and Non-volatile On-chip Caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh; Vetter, Jeffrey S; Li, Dong
Recent trends of CMOS scaling and increasing number of on-chip cores have led to a large increase in the size of on-chip caches. Since SRAM has low density and consumes large amount of leakage power, its use in designing on-chip caches has become more challenging. To address this issue, researchers are exploring the use of several emerging memory technologies, such as embedded DRAM, spin transfer torque RAM, resistive RAM, phase change RAM and domain wall memory. In this paper, we survey the architectural approaches proposed for designing memory systems and, specifically, caches with these emerging memory technologies. To highlight theirmore » similarities and differences, we present a classification of these technologies and architectural approaches based on their key characteristics. We also briefly summarize the challenges in using these technologies for architecting caches. We believe that this survey will help the readers gain insights into the emerging memory device technologies, and their potential use in designing future computing systems.« less
Kavaliauskas, Mykolas; Aspe, Rodrigo R; Babraj, John
2015-08-01
The work-to-rest ratio during cycling-based high-intensity interval training (HIT) could be important in regulating physiological and performance adaptations. We sought to determine the effectiveness of cycling-based HIT with different work-to-rest ratios for long-distance running. Thirty-two long-distance runners (age: 39 ± 8 years; sex: 14 men, 18 women; average weekly running training volume: 25 miles) underwent baseline testing (3-km time-trial, V[Combining Dot Above]O2peak and time to exhaustion, and Wingate test) before a 2-week matched-work cycling HIT of 6 × 10-second sprints with different rest periods (30 seconds [R30], 80 seconds [R80], 120 seconds [R120], or control). Three-kilometer time trial was significantly improved in the R30 group only (3.1 ± 4.0%, p = 0.04), whereas time to exhaustion was significantly increased in the 2 groups with a lower work-to-rest ratio (R30 group 6.4 ± 6.3%, p = 0.003 vs. R80 group 4.4 ± 2.7%, p = 0.03 vs. R120 group 1.9 ± 5.0%, p = 0.2). However, improvements in average power production were significantly greater with a higher work-to-rest ratio (R30 group 0.3 ± 4.1%, p = 0.8 vs. R80 group 4.6 ± 4.2%, p = 0.03 vs. R120 group 5.3 ± 5.9%, p = 0.02), whereas peak power significantly increased only in the R80 group (8.5 ± 8.2%, p = 0.04) but not in the R30 group (4.3 ± 6.1%, p = 0.3) or in the R120 group (7.1 ± 7.9%, p = 0.09). Therefore, cycling-based HIT is an effective way to improve running performance, and the type and magnitude of adaptation is dependent on the work-to-rest ratio.
Software Exploit Prevention and Remediation via Software Memory Protection
2009-05-01
trampolines that are necessary. Trampolines are pieces of code emitted into the fragment cache to transfer con- trol back to Strata. Most control...transfer instructions (CTIs) are initially linked to trampolines (unless the transfer target already exists in the fragment cache). Once a CTI’s target...instruction becomes available in the fragment cache, the CTI is linked directly to the destination, avoiding future uses of the trampoline . This
Image matrix processor for fast multi-dimensional computations
Roberson, G.P.; Skeate, M.F.
1996-10-15
An apparatus for multi-dimensional computation is disclosed which comprises a computation engine, including a plurality of processing modules. The processing modules are configured in parallel and compute respective contributions to a computed multi-dimensional image of respective two dimensional data sets. A high-speed, parallel access storage system is provided which stores the multi-dimensional data sets, and a switching circuit routes the data among the processing modules in the computation engine and the storage system. A data acquisition port receives the two dimensional data sets representing projections through an image, for reconstruction algorithms such as encountered in computerized tomography. The processing modules include a programmable local host, by which they may be configured to execute a plurality of different types of multi-dimensional algorithms. The processing modules thus include an image manipulation processor, which includes a source cache, a target cache, a coefficient table, and control software for executing image transformation routines using data in the source cache and the coefficient table and loading resulting data in the target cache. The local host processor operates to load the source cache with a two dimensional data set, loads the coefficient table, and transfers resulting data out of the target cache to the storage system, or to another destination. 10 figs.
High Performance Analytics with the R3-Cache
NASA Astrophysics Data System (ADS)
Eavis, Todd; Sayeed, Ruhan
Contemporary data warehouses now represent some of the world’s largest databases. As these systems grow in size and complexity, however, it becomes increasingly difficult for brute force query processing approaches to meet the performance demands of end users. Certainly, improved indexing and more selective view materialization are helpful in this regard. Nevertheless, with warehouses moving into the multi-terabyte range, it is clear that the minimization of external memory accesses must be a primary performance objective. In this paper, we describe the R 3-cache, a natively multi-dimensional caching framework designed specifically to support sophisticated warehouse/OLAP environments. R 3-cache is based upon an in-memory version of the R-tree that has been extended to support buffer pages rather than disk blocks. A key strength of the R 3-cache is that it is able to utilize multi-dimensional fragments of previous query results so as to significantly minimize the frequency and scale of disk accesses. Moreover, the new caching model directly accommodates the standard relational storage model and provides mechanisms for pro-active updates that exploit the existence of query “hot spots”. The current prototype has been evaluated as a component of the Sidera DBMS, a “shared nothing” parallel OLAP server designed for multi-terabyte analytics. Experimental results demonstrate significant performance improvements relative to simpler alternatives.
Zhou, ZhangBing; Zhao, Deng; Shu, Lei; Tsang, Kim-Fung
2015-01-01
Wireless sensor networks, serving as an important interface between physical environments and computational systems, have been used extensively for supporting domain applications, where multiple-attribute sensory data are queried from the network continuously and periodically. Usually, certain sensory data may not vary significantly within a certain time duration for certain applications. In this setting, sensory data gathered at a certain time slot can be used for answering concurrent queries and may be reused for answering the forthcoming queries when the variation of these data is within a certain threshold. To address this challenge, a popularity-based cooperative caching mechanism is proposed in this article, where the popularity of sensory data is calculated according to the queries issued in recent time slots. This popularity reflects the possibility that sensory data are interested in the forthcoming queries. Generally, sensory data with the highest popularity are cached at the sink node, while sensory data that may not be interested in the forthcoming queries are cached in the head nodes of divided grid cells. Leveraging these cooperatively cached sensory data, queries are answered through composing these two-tier cached data. Experimental evaluation shows that this approach can reduce the network communication cost significantly and increase the network capability. PMID:26131665
Analysis of power gating in different hierarchical levels of 2MB cache, considering variation
NASA Astrophysics Data System (ADS)
Jafari, Mohsen; Imani, Mohsen; Fathipour, Morteza
2015-09-01
This article reintroduces power gating technique in different hierarchical levels of static random-access memory (SRAM) design including cell, row, bank and entire cache memory in 16 nm Fin field effect transistor. Different structures of SRAM cells such as 6T, 8T, 9T and 10T are used in design of 2MB cache memory. The power reduction of the entire cache memory employing cell-level optimisation is 99.7% with the expense of area and other stability overheads. The power saving of the cell-level optimisation is 3× (1.2×) higher than power gating in cache (bank) level due to its superior selectivity. The access delay times are allowed to increase by 4% in the same energy delay product to achieve the best power reduction for each supply voltages and optimisation levels. The results show the row-level power gating is the best for optimising the power of the entire cache with lowest drawbacks. Comparisons of cells show that the cells whose bodies have higher power consumption are the best candidates for power gating technique in row-level optimisation. The technique has the lowest percentage of saving in minimum energy point (MEP) of the design. The power gating also improves the variation of power in all structures by at least 70%.
Respiratory hospital admissions associated with PM10 pollution in Utah, Salt Lake, and Cache Valleys
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pope CA, I.I.I.
This study assessed the association between respiratory hospital admissions and PM10 pollution in Utah, Salt Lake, and Cache valleys during April 1985 through March 1989. Utah and Salt Lake valleys had high levels of PM10 pollution that violated both the annual and 24-h standards issued by the Environmental Protection Agency (EPA). Much lower PM10 levels occurred in the Cache Valley. Utah Valley experienced the intermittent operation of its primary source of PM10 pollution: an integrated steel mill. Bronchitis and asthma admissions for preschool-age children were approximately twice as frequent in Utah Valley when the steel mill was operating versus whenmore » it was not. Similar differences were not observed in Salt Lake or Cache valleys. Even though Cache Valley had higher smoking rates and lower temperatures in winter than did Utah Valley, per capita bronchitis and asthma admissions for all ages were approximately twice as high in Utah Valley. During the period when the steel mill was closed, differences in per capita admissions between Utah and Cache valleys narrowed considerably. Regression analysis also demonstrated a statistical association between respiratory hospital admissions and PM10 pollution. The results suggest that PM10 pollution plays a role in the incidence and severity of respiratory disease.« less
Longland, William; Ostoja, Steven M.
2013-01-01
Seeds of Indian ricegrass (Achnatherum hymenoides), a native bunchgrass common to sandy soils on arid western rangelands, are naturally dispersed by seed-caching rodent species, particularly Dipodomys spp. (kangaroo rats). These animals cache large quantities of seeds when mature seeds are available on or beneath plants and recover most of their caches for consumption during the remainder of the year. Unrecovered seeds in caches account for the vast majority of Indian ricegrass seedling recruitment. We applied three different densities of white millet (Panicum miliaceum) seeds as “diversionary foods” to plots at three Great Basin study sites in an attempt to reduce rodents' over-winter cache recovery so that more Indian ricegrass seeds would remain in soil seedbanks and potentially establish new seedlings. One year after diversionary seed application, a moderate level of Indian ricegrass seedling recruitment occurred at two of our study sites in western Nevada, although there was no recruitment at the third site in eastern California. At both Nevada sites, the number of Indian ricegrass seedlings sampled along transects was significantly greater on all plots treated with diversionary seeds than on non-seeded control plots. However, the density of diversionary seeds applied to plots had a marginally non-significant effect on seedling recruitment, and it was not correlated with recruitment patterns among plots. Results suggest that application of a diversionary seed type that is preferred by seed-caching rodents provides a promising passive restoration strategy for target plant species that are dispersed by these rodents.
Do disk drives dream of buffer cache hits?
DOE Office of Scientific and Technical Information (OSTI.GOV)
Holt, A.
1994-12-31
G.E. Moore, in his book Principia Ethica, examines the popular view of ethics that deals with {open_quotes}what we ought to do{close_quotes} as well as using ethics to cover the general inquiry: {open_quotes}what is good?{close_quotes} This paper utilises Moore`s view of Ethics to examine computer systems performance. Moore asserts that {open_quotes}good{close_quotes} in itself is indefinable. It is argued in this report that, although we describe computer systems as good (or bad) a computer system cannot be good in itself, rather a means to good! In terms of {open_quotes}what we ought to do{close_quotes} this paper looks at what actions (would) bring aboutmore » good computer system performance according to computer science and engineering literature. In particular we look at duties, responsibilities and {open_quotes}to do what is right{close_quotes} in terms of system administration, design and usage. We further argue that in order to first make ethical observations with respect computer system performance and then apply them, requires technical knowledge which is typically limited to industry specialists and experts.« less
Security Enhancement Using Cache Based Reauthentication in WiMAX Based E-Learning System
Rajagopal, Chithra; Bhuvaneshwaran, Kalaavathi
2015-01-01
WiMAX networks are the most suitable for E-Learning through their Broadcast and Multicast Services at rural areas. Authentication of users is carried out by AAA server in WiMAX. In E-Learning systems the users must be forced to perform reauthentication to overcome the session hijacking problem. The reauthentication of users introduces frequent delay in the data access which is crucial in delaying sensitive applications such as E-Learning. In order to perform fast reauthentication caching mechanism known as Key Caching Based Authentication scheme is introduced in this paper. Even though the cache mechanism requires extra storage to keep the user credentials, this type of mechanism reduces the 50% of the delay occurring during reauthentication. PMID:26351658
Security Enhancement Using Cache Based Reauthentication in WiMAX Based E-Learning System.
Rajagopal, Chithra; Bhuvaneshwaran, Kalaavathi
2015-01-01
WiMAX networks are the most suitable for E-Learning through their Broadcast and Multicast Services at rural areas. Authentication of users is carried out by AAA server in WiMAX. In E-Learning systems the users must be forced to perform reauthentication to overcome the session hijacking problem. The reauthentication of users introduces frequent delay in the data access which is crucial in delaying sensitive applications such as E-Learning. In order to perform fast reauthentication caching mechanism known as Key Caching Based Authentication scheme is introduced in this paper. Even though the cache mechanism requires extra storage to keep the user credentials, this type of mechanism reduces the 50% of the delay occurring during reauthentication.
Clayton, Nicola S; Yu, Kara Shirley; Dickinson, Anthony
2003-01-01
When Western Scrub-Jays (Aphelocoma californica) cached and recovered perishable crickets, N. S. Clayton, K. S. Yu, and A. Dickinson (2001) reported that the jays rapidly learned to search for fresh crickets after a 1-day retention interval (RI) between caching and recovery but to avoid searching for perished crickets after a 4-day RI. In the present experiments, the jays generalized their search preference for crickets to intermediate RIs and used novel information about the rate of decay of crickets presented during the RI to reverse these search preferences at recovery. The authors interpret this reversal as evidence that the birds can integrate information about the caching episode with new information presented during the RI.
An Analytic Model for the Success Rate of a Robotic Actuator System in Hitting Random Targets.
Bradley, Stuart
2015-11-20
Autonomous robotic systems are increasingly being used in a wide range of applications such as precision agriculture, medicine, and the military. These systems have common features which often includes an action by an "actuator" interacting with a target. While simulations and measurements exist for the success rate of hitting targets by some systems, there is a dearth of analytic models which can give insight into, and guidance on optimization, of new robotic systems. The present paper develops a simple model for estimation of the success rate for hitting random targets from a moving platform. The model has two main dimensionless parameters: the ratio of actuator spacing to target diameter; and the ratio of platform distance moved (between actuator "firings") to the target diameter. It is found that regions of parameter space having specified high success are described by simple equations, providing guidance on design. The role of a "cost function" is introduced which, when minimized, provides optimization of design, operating, and risk mitigation costs.
Weston, Matthew; Taylor, Kathryn L; Batterham, Alan M; Hopkins, Will G
2014-07-01
Low-volume high-intensity interval training (HIT) appears to be an efficient and practical way to develop physical fitness. Our objective was to estimate meta-analysed mean effects of HIT on aerobic power (maximum oxygen consumption [VO(2max)] in an incremental test) and sprint fitness (peak and mean power in a 30-s Wingate test). Five databases (PubMed, MEDLINE, Scopus, BIOSIS and Web of Science) were searched for original research articles published up to January 2014. Search terms included 'high intensity', 'HIT', 'sprint', 'fitness' and 'VO(2max)'. Inclusion criteria were fitness assessed pre- and post-training; training period ≥2 weeks; repetition duration 30-60 s; work/rest ratio <1.0; exercise intensity described as maximal or near maximal; adult subjects aged >18 years. The final data set consisted of 55 estimates from 32 trials for VO(2max), 23 estimates from 16 trials for peak sprint power, and 19 estimates from 12 trials for mean sprint power. Effects on fitness were analysed as percentages via log transformation. Standard errors calculated from exact p values (where reported) or imputed from errors of measurement provided appropriate weightings. Fixed effects in the meta-regression model included type of study (controlled, uncontrolled), subject characteristics (sex, training status, baseline fitness) and training parameters (number of training sessions, repetition duration, work/rest ratio). Probabilistic magnitude-based inferences for meta-analysed effects were based on standardized thresholds for small, moderate and large changes (0.2, 0.6 and 1.2, respectively) derived from between-subject standard deviations (SDs) for baseline fitness. A mean low-volume HIT protocol (13 training sessions, 0.16 work/rest ratio) in a controlled trial produced a likely moderate improvement in the VO(2max) of active non-athletic males (6.2 %; 90 % confidence limits ±3.1 %), when compared with control. There were possibly moderate improvements in the VO(2max) of sedentary males (10.0 %; ±5.1 %) and active non-athletic females (3.6 %; ±4.3 %) and a likely small increase for sedentary females (7.3 %; ±4.8 %). The effect on the VO(2max) of athletic males was unclear (2.7 %; ±4.6 %). A possibly moderate additional increase was likely for subjects with a 10 mL·kg(-1)·min(-1) lower baseline VO(2max) (3.8 %; ±2.5 %), whereas the modifying effects of sex and difference in exercise dose were unclear. The comparison of HIT with traditional endurance training was unclear (-1.6 %; ±4.3 %). Unexplained variation between studies was 2.0 % (SD). Meta-analysed effects of HIT on Wingate peak and mean power were unclear. Low-volume HIT produces moderate improvements in the aerobic power of active non-athletic and sedentary subjects. More studies are needed to resolve the unclear modifying effects of sex and HIT dose on aerobic power and the unclear effects on sprint fitness.
Mueller, Sandro Manuel; Aguayo, David; Zuercher, Matthias; Fleischmann, Oliver; Boutellier, Urs; Auer, Maria; Jung, Hans H.; Toigo, Marco
2015-01-01
Aerobic high-intensity interval training (HIT) improves cardiovascular capacity but may reduce the finite work capacity above critical power (W′) and lead to atrophy of myosin heavy chain (MyHC)-2 fibers. Since whole-body vibration may enhance indices of anaerobic performance, we examined whether side-alternating whole-body vibration as a replacement for the active rest intervals during a 4x4 min HIT prevents decreases in anaerobic performance and capacity without compromising gains in aerobic function. Thirty-three young recreationally active men were randomly assigned to conduct either conventional 4x4 min HIT, HIT with 3 min of WBV at 18 Hz (HIT+VIB18) or 30 Hz (HIT+VIB30) in lieu of conventional rest intervals, or WBV at 30 Hz (VIB30). Pre and post training, critical power (CP), W′, cellular muscle characteristics, as well as cardiovascular and neuromuscular variables were determined. W′ (−14.3%, P = 0.013), maximal voluntary torque (−8.6%, P = 0.001), rate of force development (−10.5%, P = 0.018), maximal jumping power (−6.3%, P = 0.007) and cross-sectional areas of MyHC-2A fibers (−6.4%, P = 0.044) were reduced only after conventional HIT. CP, V̇O2peak, peak cardiac output, and overall capillary-to-fiber ratio were increased after HIT, HIT+VIB18, and HIT+VIB30 without differences between groups. HIT-specific reductions in anaerobic performance and capacity were prevented by replacing active rest intervals with side-alternating whole-body vibration, notably without compromising aerobic adaptations. Therefore, competitive cyclists (and potentially other endurance-oriented athletes) may benefit from replacing the active rest intervals during aerobic HIT with side-alternating whole-body vibration. Trial Registration ClinicalTrials.gov Identifier: NCT01875146 PMID:25679998
Cache Sharing and Isolation Tradeoffs in Multicore Mixed-Criticality Systems
2015-05-01
of lockdown registers, to provide way-based partitioning. These alternatives are illustrated in Fig. 1 with respect to a quad-core ARM Cortex A9...presented a cache-partitioning scheme that allows multiple tasks to share the same cache partition on a single processor (as we do for Level-A and...sets and determined the fraction that were schedulable on our target hardware platform, the quad-core ARM Cortex A9 machine mentioned earlier, the LLC
Software-Controlled Caches in the VMP Multiprocessor
1986-03-01
programming system level that Processors is tuned for the VMP design. In this vein, we are interested in exploring how far the software support can go to ...handled in software, analogously to the handling agement of the shared program state is familiar and of virtual memory page faults. Hardware support for...ensure good behavior, as opposed to how Each cache miss results in bus traffic. Table 2 pro- vides the bus cost for the "average" cache miss. Fig
Constant time worker thread allocation via configuration caching
DOE Office of Scientific and Technical Information (OSTI.GOV)
Eichenberger, Alexandre E; O'Brien, John K. P.
Mechanisms are provided for allocating threads for execution of a parallel region of code. A request for allocation of worker threads to execute the parallel region of code is received from a master thread. Cached thread allocation information identifying prior thread allocations that have been performed for the master thread are accessed. Worker threads are allocated to the master thread based on the cached thread allocation information. The parallel region of code is executed using the allocated worker threads.
PCM-Based Durable Write Cache for Fast Disk I/O
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Zhuo; Wang, Bin; Carpenter, Patrick
2012-01-01
Flash based solid-state devices (FSSDs) have been adopted within the memory hierarchy to improve the performance of hard disk drive (HDD) based storage system. However, with the fast development of storage-class memories, new storage technologies with better performance and higher write endurance than FSSDs are emerging, e.g., phase-change memory (PCM). Understanding how to leverage these state-of-the-art storage technologies for modern computing systems is important to solve challenging data intensive computing problems. In this paper, we propose to leverage PCM for a hybrid PCM-HDD storage architecture. We identify the limitations of traditional LRU caching algorithms for PCM-based caches, and develop amore » novel hash-based write caching scheme called HALO to improve random write performance of hard disks. To address the limited durability of PCM devices and solve the degraded spatial locality in traditional wear-leveling techniques, we further propose novel PCM management algorithms that provide effective wear-leveling while maximizing access parallelism. We have evaluated this PCM-based hybrid storage architecture using applications with a diverse set of I/O access patterns. Our experimental results demonstrate that the HALO caching scheme leads to an average reduction of 36.8% in execution time compared to the LRU caching scheme, and that the SFC wear leveling extends the lifetime of PCM by a factor of 21.6.« less
Cooperation and information replication in wireless networks.
Poularakis, Konstantinos; Tassiulas, Leandros
2016-03-06
A significant portion of today's network traffic is due to recurring downloads of a few popular contents. It has been observed that replicating the latter in caches installed at network edges-close to users-can drastically reduce network bandwidth usage and improve content access delay. Such caching architectures are gaining increasing interest in recent years as a way of dealing with the explosive traffic growth, fuelled further by the downward slope in storage space price. In this work, we provide an overview of caching with a particular emphasis on emerging network architectures that enable caching at the radio access network. In this context, novel challenges arise due to the broadcast nature of the wireless medium, which allows simultaneously serving multiple users tuned into a multicast stream, and the mobility of the users who may be frequently handed off from one cell tower to another. Existing results indicate that caching at the wireless edge has a great potential in removing bottlenecks on the wired backbone networks. Taking into consideration the schedule of multicast service and mobility profiles is crucial to extract maximum benefit in network performance. © 2016 The Author(s).
Turbidity and Total Suspended Solids on the Lower Cache River Watershed, AR.
Rosado-Berrios, Carlos A; Bouldin, Jennifer L
2016-06-01
The Cache River Watershed (CRW) in Arkansas is part of one of the largest remaining bottomland hardwood forests in the US. Although wetlands are known to improve water quality, the Cache River is listed as impaired due to sedimentation and turbidity. This study measured turbidity and total suspended solids (TSS) in seven sites of the lower CRW; six sites were located on the Bayou DeView tributary of the Cache River. Turbidity and TSS levels ranged from 1.21 to 896 NTU, and 0.17 to 386.33 mg/L respectively and had an increasing trend over the 3-year study. However, a decreasing trend from upstream to downstream in the Bayou DeView tributary was noted. Sediment loading calculated from high precipitation events and mean TSS values indicate that contributions from the Cache River main channel was approximately 6.6 times greater than contributions from Bayou DeView. Land use surrounding this river channel affects water quality as wetlands provide a filter for sediments in the Bayou DeView channel.
Tier 3 batch system data locality via managed caches
NASA Astrophysics Data System (ADS)
Fischer, Max; Giffels, Manuel; Jung, Christopher; Kühn, Eileen; Quast, Günter
2015-05-01
Modern data processing increasingly relies on data locality for performance and scalability, whereas the common HEP approaches aim for uniform resource pools with minimal locality, recently even across site boundaries. To combine advantages of both, the High- Performance Data Analysis (HPDA) Tier 3 concept opportunistically establishes data locality via coordinated caches. In accordance with HEP Tier 3 activities, the design incorporates two major assumptions: First, only a fraction of data is accessed regularly and thus the deciding factor for overall throughput. Second, data access may fallback to non-local, making permanent local data availability an inefficient resource usage strategy. Based on this, the HPDA design generically extends available storage hierarchies into the batch system. Using the batch system itself for scheduling file locality, an array of independent caches on the worker nodes is dynamically populated with high-profile data. Cache state information is exposed to the batch system both for managing caches and scheduling jobs. As a result, users directly work with a regular, adequately sized storage system. However, their automated batch processes are presented with local replications of data whenever possible.
Integrating Cache Performance Modeling and Tuning Support in Parallelization Tools
NASA Technical Reports Server (NTRS)
Waheed, Abdul; Yan, Jerry; Saini, Subhash (Technical Monitor)
1998-01-01
With the resurgence of distributed shared memory (DSM) systems based on cache-coherent Non Uniform Memory Access (ccNUMA) architectures and increasing disparity between memory and processors speeds, data locality overheads are becoming the greatest bottlenecks in the way of realizing potential high performance of these systems. While parallelization tools and compilers facilitate the users in porting their sequential applications to a DSM system, a lot of time and effort is needed to tune the memory performance of these applications to achieve reasonable speedup. In this paper, we show that integrating cache performance modeling and tuning support within a parallelization environment can alleviate this problem. The Cache Performance Modeling and Prediction Tool (CPMP), employs trace-driven simulation techniques without the overhead of generating and managing detailed address traces. CPMP predicts the cache performance impact of source code level "what-if" modifications in a program to assist a user in the tuning process. CPMP is built on top of a customized version of the Computer Aided Parallelization Tools (CAPTools) environment. Finally, we demonstrate how CPMP can be applied to tune a real Computational Fluid Dynamics (CFD) application.
Clark's nutcracker spatial memory: the importance of large, structural cues.
Bednekoff, Peter A; Balda, Russell P
2014-02-01
Clark's nutcrackers, Nucifraga columbiana, cache and recover stored seeds in high alpine areas including areas where snowfall, wind, and rockslides may frequently obscure or alter cues near the cache site. Previous work in the laboratory has established that Clark's nutcrackers use spatial memory to relocate cached food. Following from aspects of this work, we performed experiments to test the importance of large, structural cues for Clark's nutcracker spatial memory. Birds were no more accurate in recovering caches when more objects were on the floor of a large experimental room nor when this room was subdivided with a set of panels. However, nutcrackers were consistently less accurate in this large room than in a small experimental room. Clark's nutcrackers probably use structural features of experimental rooms as important landmarks during recovery of cached food. This use of large, extremely stable cues may reflect the imperfect reliability of smaller, closer cues in the natural habitat of Clark's nutcrackers. This article is part of a Special Issue entitled: CO3 2013. Copyright © 2013 Elsevier B.V. All rights reserved.
Data Resilience in the dCache Storage System
Rossi, A. L.; Adeyemi, F.; Ashish, A.; ...
2017-11-23
In this study we discuss design, implementation considerations, and performance of a new Resilience Service in the dCache storage system responsible for file availability and durability functionality.
Simplifying and speeding the management of intra-node cache coherence
Blumrich, Matthias A [Ridgefield, CT; Chen, Dong [Croton on Hudson, NY; Coteus, Paul W [Yorktown Heights, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Phillip [Cortlandt Manor, NY; Hoenicke, Dirk [Ossining, NY; Ohmacht, Martin [Yorktown Heights, NY
2012-04-17
A method and apparatus for managing coherence between two processors of a two processor node of a multi-processor computer system. Generally the present invention relates to a software algorithm that simplifies and significantly speeds the management of cache coherence in a message passing parallel computer, and to hardware apparatus that assists this cache coherence algorithm. The software algorithm uses the opening and closing of put/get windows to coordinate the activated required to achieve cache coherence. The hardware apparatus may be an extension to the hardware address decode, that creates, in the physical memory address space of the node, an area of virtual memory that (a) does not actually exist, and (b) is therefore able to respond instantly to read and write requests from the processing elements.
dCache: Big Data storage for HEP communities and beyond
NASA Astrophysics Data System (ADS)
Millar, A. P.; Behrmann, G.; Bernardt, C.; Fuhrmann, P.; Litvintsev, D.; Mkrtchyan, T.; Petersen, A.; Rossi, A.; Schwank, K.
2014-06-01
With over ten years in production use dCache data storage system has evolved to match ever changing lansdcape of continually evolving storage technologies with new solutions to both existing problems and new challenges. In this paper, we present three areas of innovation in dCache: providing efficient access to data with NFS v4.1 pNFS, adoption of CDMI and WebDAV as an alternative to SRM for managing data, and integration with alternative authentication mechanisms.
Nelson, Michael E.; Mech, L. David
2011-01-01
Wolves (Canis lupus) in northeastern Minnesota cached six radio-collars (four in winter, two in spring-summer) of 202 radio-collared White-tailed Deer (Odocoileus virginianus) they killed or consumed from 1975 to 2010. A Wolf bedded on top of one collar cached in snow. We found one collar each at a Wolf den and Wolf rendezvous site, 2.5 km and 0.5 km respectively, from each deer's previous locations.
Middleton, B.; Wu, X.B.
2008-01-01
Agricultural development on floodplains contributes to hydrologic alteration and forest fragmentation, which may alter landscape-level processes. These changes may be related to shifts in the seed bank composition of floodplain wetlands. We examined the patterns of seed bank composition across a floodplain watershed by looking at the number of seeds germinating per m2 by species in 60 farmed and intact forested wetlands along the Cache River watershed in Illinois. The seed bank composition was compared above and below a water diversion (position), which artificially subdivides the watershed. Position of these wetlands represented the most variability of Axis I in a Nonmetric Multidimensional Scaling (NMS) analysis of site environmental variables and their relationship to seed bank composition (coefficient of determination for Axis 1: r2 = 0.376; Pearson correlation of position to Axis 1: r = 0.223). The 3 primary axes were also represented by other site environmental variables, including farming status (farmed or unfarmed), distance from the mouth of the river, latitude, and longitude. Spatial analysis based on Mantel correlograms showed that both water-dispersed and wind/water-dispersed seed assemblages had strong spatial structure in the upper Cache (above the water diversion), bur the spatial structure of water-dispersed seed assemblage was diminished in the lower Cache (below the water diversion), which lost floodpulsing. Bearing analysis also Suggested that water-dispersal process had a stronger influence on the overall spatial pattern of seed assemblage in the upper Cache, while wind/water-dispersal process had a stronger influence in the lower Cache. An analysis of the landscapes along the river showed that the mid-lower Cache (below the water diversion) had undergone greater land cover changes associated with agriculture than did the upper Cache watershed. Thus, the combination of forest fragmentation and hydrologic changes in the surrounding landscape may have had an influence on the seed bank composition and spatial distribution of the seed banks of the Cache River watershed. Our study suggests that the spatial pattern of seed bank composition may be influenced by landscape-level factors and processes.
Forest rodents provide directed dispersal of Jeffrey pine seeds
Briggs, J.S.; Wall, S.B.V.; Jenkins, S.H.
2009-01-01
Some species of animals provide directed dispersal of plant seeds by transporting them nonrandomly to microsites where their chances of producing healthy seedlings are enhanced. We investigated whether this mutualistic interaction occurs between granivorous rodents and Jeffrey pine (Pinus jeffreyi) in the eastern Sierra Nevada by comparing the effectiveness of random abiotic seed dispersal with the dispersal performed by four species of rodents: deer mice (Peromyscus maniculatus), yellow-pine and long-eared chipmunks (Tamias amoenus and T. quadrimaculatus), and golden-mantled ground squirrels (Spermophilus lateralis). We conducted two caching studies using radio-labeled seeds, the first with individual animals in field enclosures and the second with a community of rodents in open forest. We used artificial caches to compare the fates of seeds placed at the range of microsites and depths used by animals with the fates of seeds dispersed abiotically. Finally, we examined the distribution and survival of naturally establishing seedlings over an eight-year period.Several lines of evidence suggested that this community of rodents provided directed dispersal. Animals preferred to cache seeds in microsites that were favorable for emergence or survival of seedlings and avoided caching in microsites in which seedlings fared worst. Seeds buried at depths typical of animal caches (5–25 mm) produced at least five times more seedlings than did seeds on the forest floor. The four species of rodents differed in the quality of dispersal they provided. Small, shallow caches made by deer mice most resembled seeds dispersed by abiotic processes, whereas many of the large caches made by ground squirrels were buried too deeply for successful emergence of seedlings. Chipmunks made the greatest number of caches within the range of depths and microsites favorable for establishment of pine seedlings. Directed dispersal is an important element of the population dynamics of Jeffrey pine, a dominant tree species in the eastern Sierra Nevada. Quantifying the occurrence and dynamics of directed dispersal in this and other cases will contribute to better understanding of mutualistic coevolution of plants and animals and to more effective management of ecosystems in which directed dispersal is a keystone process.
Overview of the HIT-SI3 spheromak experiment
NASA Astrophysics Data System (ADS)
Hossack, A. C.; Jarboe, T. R.; Chandra, R. N.; Morgan, K. D.; Sutherland, D. A.; Everson, C. J.; Penna, J. M.; Nelson, B. A.
2017-10-01
The HIT-SI and HIT-SI3 spheromak experiments (a = 23 cm) study efficient, steady-state current drive for magnetic confinement plasmas using a novel method which is ideal for low aspect ratio, toroidal geometries. Sustained spheromaks show coherent, imposed plasma motion and low plasma-generated mode activity, indicating stability. Analysis of surface magnetic fields in HIT-SI indicates large n = 0 and 1 mode amplitudes and little energy in higher modes. Within measurement uncertainties all the n = 1 energy is imposed by the injectors, rather than being plasma-generated. The fluctuating field imposed by the injectors is sufficient to sustain the toroidal current through dynamo action whereas the plasma-generated field is not (Hossack et al., Phys. Plasmas, 2017). Ion Doppler spectroscopy shows coherent, imposed plasma motion inside r 10 cm in HIT-SI and a smaller volume of coherent motion in HIT-SI3. Coherent motion indicates the spheromak is stable and a lack of plasma-generated n = 1 energy indicates the maximum q is maintained below 1 for stability during sustainment. In HIT-SI3, the imposed mode structure is varied to test the plasma response (Hossack et al., Nucl. Fusion, 2017). Imposing n = 2, n = 3, or large, rotating n = 1 perturbations is correlated with transient plasma-generated activity. Work supported by the U.S. Department of Energy, Office of Science, Office of Fusion Energy Sciences, under Award Number DE-FG02-96ER54361.
Compiler-Directed File Layout Optimization for Hierarchical Storage Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ding, Wei; Zhang, Yuanrui; Kandemir, Mahmut
File layout of array data is a critical factor that effects the behavior of storage caches, and has so far taken not much attention in the context of hierarchical storage systems. The main contribution of this paper is a compiler-driven file layout optimization scheme for hierarchical storage caches. This approach, fully automated within an optimizing compiler, analyzes a multi-threaded application code and determines a file layout for each disk-resident array referenced by the code, such that the performance of the target storage cache hierarchy is maximized. We tested our approach using 16 I/O intensive application programs and compared its performancemore » against two previously proposed approaches under different cache space management schemes. Our experimental results show that the proposed approach improves the execution time of these parallel applications by 23.7% on average.« less
Compiler-Directed File Layout Optimization for Hierarchical Storage Systems
Ding, Wei; Zhang, Yuanrui; Kandemir, Mahmut; ...
2013-01-01
File layout of array data is a critical factor that effects the behavior of storage caches, and has so far taken not much attention in the context of hierarchical storage systems. The main contribution of this paper is a compiler-driven file layout optimization scheme for hierarchical storage caches. This approach, fully automated within an optimizing compiler, analyzes a multi-threaded application code and determines a file layout for each disk-resident array referenced by the code, such that the performance of the target storage cache hierarchy is maximized. We tested our approach using 16 I/O intensive application programs and compared its performancemore » against two previously proposed approaches under different cache space management schemes. Our experimental results show that the proposed approach improves the execution time of these parallel applications by 23.7% on average.« less
Efficacy of Code Optimization on Cache-Based Processors
NASA Technical Reports Server (NTRS)
VanderWijngaart, Rob F.; Saphir, William C.; Chancellor, Marisa K. (Technical Monitor)
1997-01-01
In this paper a number of techniques for improving the cache performance of a representative piece of numerical software is presented. Target machines are popular processors from several vendors: MIPS R5000 (SGI Indy), MIPS R8000 (SGI PowerChallenge), MIPS R10000 (SGI Origin), DEC Alpha EV4 + EV5 (Cray T3D & T3E), IBM RS6000 (SP Wide-node), Intel PentiumPro (Ames' Whitney), Sun UltraSparc (NERSC's NOW). The optimizations all attempt to increase the locality of memory accesses. But they meet with rather varied and often counterintuitive success on the different computing platforms. We conclude that it may be genuinely impossible to obtain portable performance on the current generation of cache-based machines. At the least, it appears that the performance of modern commodity processors cannot be described with parameters defining the cache alone.
SIDECACHE: Information access, management and dissemination framework for web services.
Doderer, Mark S; Burkhardt, Cory; Robbins, Kay A
2011-06-14
Many bioinformatics algorithms and data sets are deployed using web services so that the results can be explored via the Internet and easily integrated into other tools and services. These services often include data from other sites that is accessed either dynamically or through file downloads. Developers of these services face several problems because of the dynamic nature of the information from the upstream services. Many publicly available repositories of bioinformatics data frequently update their information. When such an update occurs, the developers of the downstream service may also need to update. For file downloads, this process is typically performed manually followed by web service restart. Requests for information obtained by dynamic access of upstream sources is sometimes subject to rate restrictions. SideCache provides a framework for deploying web services that integrate information extracted from other databases and from web sources that are periodically updated. This situation occurs frequently in biotechnology where new information is being continuously generated and the latest information is important. SideCache provides several types of services including proxy access and rate control, local caching, and automatic web service updating. We have used the SideCache framework to automate the deployment and updating of a number of bioinformatics web services and tools that extract information from remote primary sources such as NCBI, NCIBI, and Ensembl. The SideCache framework also has been used to share research results through the use of a SideCache derived web service.
A Scalable proxy cache for Grid Data Access
NASA Astrophysics Data System (ADS)
Cristian Cirstea, Traian; Just Keijser, Jan; Koeroo, Oscar Arthur; Starink, Ronald; Templon, Jeffrey Alan
2012-12-01
We describe a prototype grid proxy cache system developed at Nikhef, motivated by a desire to construct the first building block of a future https-based Content Delivery Network for grid infrastructures. Two goals drove the project: firstly to provide a “native view” of the grid for desktop-type users, and secondly to improve performance for physics-analysis type use cases, where multiple passes are made over the same set of data (residing on the grid). We further constrained the design by requiring that the system should be made of standard components wherever possible. The prototype that emerged from this exercise is a horizontally-scalable, cooperating system of web server / cache nodes, fronted by a customized webDAV server. The webDAV server is custom only in the sense that it supports http redirects (providing horizontal scaling) and that the authentication module has, as back end, a proxy delegation chain that can be used by the cache nodes to retrieve files from the grid. The prototype was deployed at Nikhef and tested at a scale of several terabytes of data and approximately one hundred fast cores of computing. Both small and large files were tested, in a number of scenarios, and with various numbers of cache nodes, in order to understand the scaling properties of the system. For properly-dimensioned cache-node hardware, the system showed speedup of several integer factors for the analysis-type use cases. These results and others are presented and discussed.
Dynamic Allocation of SPM Based on Time-Slotted Cache Conflict Graph for System Optimization
NASA Astrophysics Data System (ADS)
Wu, Jianping; Ling, Ming; Zhang, Yang; Mei, Chen; Wang, Huan
This paper proposes a novel dynamic Scratch-pad Memory allocation strategy to optimize the energy consumption of the memory sub-system. Firstly, the whole program execution process is sliced into several time slots according to the temporal dimension; thereafter, a Time-Slotted Cache Conflict Graph (TSCCG) is introduced to model the behavior of Data Cache (D-Cache) conflicts within each time slot. Then, Integer Nonlinear Programming (INP) is implemented, which can avoid time-consuming linearization process, to select the most profitable data pages. Virtual Memory System (VMS) is adopted to remap those data pages, which will cause severe Cache conflicts within a time slot, to SPM. In order to minimize the swapping overhead of dynamic SPM allocation, a novel SPM controller with a tightly coupled DMA is introduced to issue the swapping operations without CPU's intervention. Last but not the least, this paper discusses the fluctuation of system energy profit based on different MMU page size as well as the Time Slot duration quantitatively. According to our design space exploration, the proposed method can optimize all of the data segments, including global data, heap and stack data in general, and reduce the total energy consumption by 27.28% on average, up to 55.22% with a marginal performance promotion. And comparing to the conventional static CCG (Cache Conflicts Graph), our approach can obtain 24.7% energy profit on average, up to 30.5% with a sight boost in performance.
Transfer, loss and physical processing of water in hit-and-run collisions of planetary embryos
NASA Astrophysics Data System (ADS)
Burger, C.; Maindl, T. I.; Schäfer, C. M.
2018-01-01
Collisions between large, similar-sized bodies are believed to shape the final characteristics and composition of terrestrial planets. Their inventories of volatiles such as water are either delivered or at least significantly modified by such events. Besides the transition from accretion to erosion with increasing impact velocity, similar-sized collisions can also result in hit-and-run outcomes for sufficiently oblique impact angles and large enough projectile-to-target mass ratios. We study volatile transfer and loss focusing on hit-and-run encounters by means of smooth particle hydrodynamics simulations, including all main parameters: impact velocity, impact angle, mass ratio and also the total colliding mass. We find a broad range of overall water losses, up to 75% in the most energetic hit-and-run events, and confirm the much more severe consequences for the smaller body also for stripping of volatile layers. Transfer of water between projectile and target inventories is found to be mostly rather inefficient, and final water contents are dominated by pre-collision inventories reduced by impact losses, for similar pre-collision water mass fractions. Comparison with our numerical results shows that current collision outcome models are not accurate enough to reliably predict these composition changes in hit-and-run events. To also account for non-mechanical losses, we estimate the amount of collisionally vaporized water over a broad range of masses and find that these contributions are particularly important in collisions of ˜ Mars-sized bodies, with sufficiently high impact energies, but still relatively low gravity. Our results clearly indicate that the cumulative effect of several (hit-and-run) collisions can efficiently strip protoplanets of their volatile layers, especially the smaller body, as it might be common, e.g., for Earth-mass planets in systems with Super-Earths. An accurate model for stripping of volatiles that can be included in future planet formation simulations has to account for the peculiarities of hit-and-run events and track compositional changes in both large post-collision fragments.
Horizontally scaling dChache SRM with the Terracotta platform
DOE Office of Scientific and Technical Information (OSTI.GOV)
Perelmutov, T.; Crawford, M.; Moibenko, A.
2011-01-01
The dCache disk caching file system has been chosen by a majority of LHC experiments Tier 1 centers for their data storage needs. It is also deployed at many Tier 2 centers. The Storage Resource Manager (SRM) is a standardized grid storage interface and a single point of remote entry into dCache, and hence is a critical component. SRM must scale to increasing transaction rates and remain resilient against changing usage patterns. The initial implementation of the SRM service in dCache suffered from an inability to support clustered deployment, and its performance was limited by the hardware of a singlemore » node. Using the Terracotta platform, we added the ability to horizontally scale the dCache SRM service to run on multiple nodes in a cluster configuration, coupled with network load balancing. This gives site administrators the ability to increase the performance and reliability of SRM service to face the ever-increasing requirements of LHC data handling. In this paper we will describe the previous limitations of the architecture SRM server and how the Terracotta platform allowed us to readily convert single node service into a highly scalable clustered application.« less
Dispersal Mutualism Incorporated into Large-Scale, Infrequent Disturbances
Parker, V. Thomas
2015-01-01
Because of their influence on succession and other community interactions, large-scale, infrequent natural disturbances also should play a major role in mutualistic interactions. Using field data and experiments, I test whether mutualisms have been incorporated into large-scale wildfire by whether the outcomes of a mutualism depend on disturbance. In this study a seed dispersal mutualism is shown to depend on infrequent, large-scale disturbances. A dominant shrubland plant (Arctostaphylos species) produces seeds that make up a persistent soil seed bank and requires fire to germinate. In post-fire stands, I show that seedlings emerging from rodent caches dominate sites experiencing higher fire intensity. Field experiments show that rodents (Perimyscus californicus, P. boylii) do cache Arctostaphylos fruit and bury most seed caches to a sufficient depth to survive a killing heat pulse that a fire might drive into the soil. While the rodent dispersal and caching behavior itself has not changed compared to other habitats, the environmental transformation caused by wildfire converts the caching burial of seed from a dispersal process to a plant fire adaptive trait, and provides the context for stimulating subsequent life history evolution in the plant host. PMID:26151560
Reader set encoding for directory of shared cache memory in multiprocessor system
Ahn, Dnaiel; Ceze, Luis H.; Gara, Alan; Ohmacht, Martin; Xiaotong, Zhuang
2014-06-10
In a parallel processing system with speculative execution, conflict checking occurs in a directory lookup of a cache memory that is shared by all processors. In each case, the same physical memory address will map to the same set of that cache, no matter which processor originated that access. The directory includes a dynamic reader set encoding, indicating what speculative threads have read a particular line. This reader set encoding is used in conflict checking. A bitset encoding is used to specify particular threads that have read the line.
DSP code optimization based on cache
NASA Astrophysics Data System (ADS)
Xu, Chengfa; Li, Chengcheng; Tang, Bin
2013-03-01
DSP program's running efficiency on board is often lower than which via the software simulation during the program development, which is mainly resulted from the user's improper use and incomplete understanding of the cache-based memory. This paper took the TI TMS320C6455 DSP as an example, analyzed its two-level internal cache, and summarized the methods of code optimization. Processor can achieve its best performance when using these code optimization methods. At last, a specific algorithm application in radar signal processing is proposed. Experiment result shows that these optimization are efficient.
Yap, Te-Lu; Chen, Yong; Nah, Shireen A; Ong, Caroline Choo Phaik; Jacobsen, Anette; Low, Yee
2016-12-01
Our study aimed to compare the efficacy of two endoscopic techniques used for the correction of vesicoureteral reflux (VUR): subureteral transurethral injection (STING) and hydrodistension implantation technique (HIT). A systematic review was conducted using MEDLINE, Google scholar, and Cochrane databases from 1984 to 2015. Meta-analysis of the selected studies was performed to compare the extent of reflux resolution following both techniques. Six observational studies met the inclusion criteria for content. These comprised 632 ureters treated by STING and 895 ureters treated by HIT procedure. All included studies utilized dextranomer/hyaluronic acid (Deflux) as the bulking agent. The overall resolution of VUR was significantly higher in HIT (82.5%) compared to STING (71.4%) [pooled odds ratio (OR)=0.54; 95% confidence interval (CI) 0.42-0.69; P<0.0001; I 2 =8%]. A subgroup analysis showed that HIT had better outcomes than STING for both lower grade (I-III) [OR=0.43; 95% CI 0.23-0.82; P=0.01; I 2 =0%] and high-grade VUR (IV-V) [OR=0.43; 95% CI 0.20-0.91; P=0.03; I 2 =0%]. However, there was no statistical difference in the requirement of additional injections between STING and HIT groups. HIT is superior to STING technique for resolution of VUR after Deflux injection. However, more randomized trials with longer follow-up are necessary to demonstrate the benefit of HIT compared to STING procedure. Retrospective comparative studies - level III. Copyright © 2016. Published by Elsevier Inc.
Memory and the hippocampus in food-storing birds: a comparative approach.
Clayton, N S
1998-01-01
Comparative studies provide a unique source of evidence for the role of the hippocampus in learning and memory. Within birds and mammals, the hippocampal volume of scatter-hoarding species that cache food in many different locations is enlarged, relative to the remainder of the telencephalon, when compared with than that of species which cache food in one larder, or do not cache at all. Do food-storing species show enhanced memory function in association with the volumetric enlargement of the hippocampus? Comparative studies within the parids (titmice and chickadees) and corvids (jays, nutcrackers and magpies), two families of birds which show natural variation in food-storing behavior, suggest that there may be two kinds of memory specialization associated with scatter-hoarding. First, in terms of spatial memory, several scatter-hoarding species have a more accurate and enduring spatial memory, and a preference to rely more heavily upon spatial cues, than that of closely related species which store less food, or none at all. Second, some scatter-hoarding parids and corvids are also more resistant to memory interference. While the most critical component about a cache site may be its spatial location, there is mounting evidence that food-storing birds remember additional information about the contents and status of cache sites. What is the underlying neural mechanism by which the hippocampus learns and remembers cache sites? The current mammalian dogma is that the neural mechanisms of learning and memory are achieved primarily by variations in synaptic number and efficacy. Recent work on the concomitant development of food-storing, memory and the avian hippocampus illustrates that the avian hippocampus may swell or shrivel by as much as 30% in response to presence or absence of food-storing experience. Memory for food caches triggers a dramatic increase in the total number of number of neurons within the avian hippocampus by altering the rate at which these cells are born and die.
NIC atomic operation unit with caching and bandwidth mitigation
Hemmert, Karl Scott; Underwood, Keith D.; Levenhagen, Michael J.
2016-03-01
A network interface controller atomic operation unit and a network interface control method comprising, in an atomic operation unit of a network interface controller, using a write-through cache and employing a rate-limiting functional unit.
Sparse Partial Equilibrium Tables in Chemically Resolved Reactive Flow
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vitello, P; Fried, L E; Pudliner, B
2003-07-14
The detonation of an energetic material is the result of a complex interaction between kinetic chemical reactions and hydrodynamics. Unfortunately, little is known concerning the detailed chemical kinetics of detonations in energetic materials. CHEETAH uses rate laws to treat species with the slowest chemical reactions, while assuming other chemical species are in equilibrium. CHEETAH supports a wide range of elements and condensed detonation products and can also be applied to gas detonations. A sparse hash table of equation of state values, called the ''cache'' is used in CHEETAH to enhance the efficiency of kinetic reaction calculations. For large-scale parallel hydrodynamicmore » calculations, CHEETAH uses MPI communication to updates to the cache. We present here details of the sparse caching model used in the CHEETAH. To demonstrate the efficiency of modeling using a sparse cache model we consider detonations in energetic materials.« less
Shrager, Jeff; Billman, Dorrit; Convertino, Gregorio; Massar, J P; Pirolli, Peter
2010-01-01
Science is a form of distributed analysis involving both individual work that produces new knowledge and collaborative work to exchange information with the larger community. There are many particular ways in which individual and community can interact in science, and it is difficult to assess how efficient these are, and what the best way might be to support them. This paper reports on a series of experiments in this area and a prototype implementation using a research platform called CACHE. CACHE both supports experimentation with different structures of interaction between individual and community cognition and serves as a prototype for computational support for those structures. We particularly focus on CACHE-BC, the Bayes community version of CACHE, within which the community can break up analytical tasks into "mind-sized" units and use provenance tracking to keep track of the relationship between these units. Copyright © 2009 Cognitive Science Society, Inc.
Efficient image data distribution and management with application to web caching architectures
NASA Astrophysics Data System (ADS)
Han, Keesook J.; Suter, Bruce W.
2003-03-01
We present compact image data structures and associated packet delivery techniques for effective Web caching architectures. Presently, images on a web page are inefficiently stored, using a single image per file. Our approach is to use clustering to merge similar images into a single file in order to exploit the redundancy between images. Our studies indicate that a 30-50% image data size reduction can be achieved by eliminating the redundancies of color indexes. Attached to this file is new metadata to permit an easy extraction of images. This approach will permit a more efficient use of the cache, since a shorter list of cache references will be required. Packet and transmission delays can be reduced by 50% eliminating redundant TCP/IP headers and connection time. Thus, this innovative paradigm for the elimination of redundancy may provide valuable benefits for optimizing packet delivery in IP networks by reducing latency and minimizing the bandwidth requirements.
NASA Technical Reports Server (NTRS)
Liu, Yuan-Kwei
1991-01-01
The feasibility is analyzed of upgrading the Intel 386 microprocessor, which has been proposed as the baseline processor for the Space Station Freedom (SSF) Data Management System (DMS), to the more advanced i486 microprocessors. The items compared between the two processors include the instruction set architecture, power consumption, the MIL-STD-883C Class S (Space) qualification schedule, and performance. The advantages of the i486 over the 386 are (1) lower power consumption; and (2) higher floating point performance. The i486 on-chip cache does not have parity check or error detection and correction circuitry. The i486 with on-chip cache disabled, however, has lower integer performance than the 386 without cache, which is the current DMS design choice. Adding cache to the 386/386 DX memory hierachy appears to be the most beneficial change to the current DMS design at this time.
NASA Technical Reports Server (NTRS)
Liu, Yuan-Kwei
1991-01-01
The feasibility is analyzed of upgrading the Intel 386 microprocessor, which has been proposed as the baseline processor for the Space Station Freedom (SSF) Data Management System (DMS), to the more advanced i486 microprocessors. The items compared between the two processors include the instruction set architecture, power consumption, the MIL-STD-883C Class S (Space) qualification schedule, and performance. The advantages of the i486 over the 386 are (1) lower power consumption; and (2) higher floating point performance. The i486 on-chip cache does not have parity check or error detection and correction circuitry. The i486 with on-chip cache disabled, however, has lower integer performance than the 386 without cache, which is the current DMS design choice. Adding cache to the 386/387 DX memory hierarchy appears to be the most beneficial change to the current DMS design at this time.
Using Solid State Disk Array as a Cache for LHC ATLAS Data Analysis
NASA Astrophysics Data System (ADS)
Yang, W.; Hanushevsky, A. B.; Mount, R. P.; Atlas Collaboration
2014-06-01
User data analysis in high energy physics presents a challenge to spinning-disk based storage systems. The analysis is data intense, yet reads are small, sparse and cover a large volume of data files. It is also unpredictable due to users' response to storage performance. We describe here a system with an array of Solid State Disk as a non-conventional, standalone file level cache in front of the spinning disk storage to help improve the performance of LHC ATLAS user analysis at SLAC. The system uses several days of data access records to make caching decisions. It can also use information from other sources such as a work-flow management system. We evaluate the performance of the system both in terms of caching and its impact on user analysis jobs. The system currently uses Xrootd technology, but the technique can be applied to any storage system.
Is random access memory random?
NASA Technical Reports Server (NTRS)
Denning, P. J.
1986-01-01
Most software is contructed on the assumption that the programs and data are stored in random access memory (RAM). Physical limitations on the relative speeds of processor and memory elements lead to a variety of memory organizations that match processor addressing rate with memory service rate. These include interleaved and cached memory. A very high fraction of a processor's address requests can be satified from the cache without reference to the main memory. The cache requests information from main memory in blocks that can be transferred at the full memory speed. Programmers who organize algorithms for locality can realize the highest performance from these computers.
2015-06-10
This diagram, superimposed on a photo of Martian landscape, illustrates a concept called "adaptive caching," which is in development for NASA's 2020 Mars rover mission. In addition to the investigations that the Mars 2020 rover will conduct on Mars, the rover will collect carefully selected samples of Mars rock and soil and cache them to be available for possible return to Earth if a Mars sample-return mission is scheduled and flown. Each sample will be stored in a sealed tube. Adaptive caching would result in a set of samples, up to the maximum number of tubes carried on the rover, being placed on the surface at the discretion of the mission operators. The tubes holding the collected samples would not go into a surrounding container. In this illustration, green dots indicate "regions of interest," where samples might be collected. The green diamond indicates one region of interest serving as the depot for the cache. The green X at upper right represents the landing site. The solid black line indicates the rover's route during its prime mission, and the dashed black line indicates its route during an extension of the mission. The base image is a portion of the "Everest Panorama" taken by the panoramic camera on NASA's Mars Exploration Rover Spirit at the top of Husband Hill in 2005. http://photojournal.jpl.nasa.gov/catalog/PIA19150
Use of the sun as a heading indicator when caching and recovering in a wild rodent
Samson, Jamie; Manser, Marta B.
2016-01-01
A number of diurnal species have been shown to use directional information from the sun to orientate. The use of the sun in this way has been suggested to occur in either a time-dependent (relying on specific positional information) or a time-compensated manner (a compass that adjusts itself over time with the shifts in the sun’s position). However, some interplay may occur between the two where a species could also use the sun in a time-limited way, whereby animals acquire certain information about the change of position, but do not show full compensational abilities. We tested whether Cape ground squirrels (Xerus inauris) use the sun as an orientation marker to provide information for caching and recovery. This species is a social sciurid that inhabits arid, sparsely vegetated habitats in Southern Africa, where the sun is nearly always visible during the diurnal period. Due to the lack of obvious landmarks, we predicted that they might use positional cues from the sun in the sky as a reference point when caching and recovering food items. We provide evidence that Cape ground squirrels use information from the sun’s position while caching and reuse this information in a time-limited way when recovering these caches. PMID:27580797
Solutions and debugging for data consistency in multiprocessors with noncoherent caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bernstein, D.; Mendelson, B.; Breternitz, M. Jr.
1995-02-01
We analyze two important problems that arise in shared-memory multiprocessor systems. The stale data problem involves ensuring that data items in local memory of individual processors are current, independent of writes done by other processors. False sharing occurs when two processors have copies of the same shared data block but update different portions of the block. The false sharing problem involves guaranteeing that subsequent writes are properly combined. In modern architectures these problems are usually solved in hardware, by exploiting mechanisms for hardware controlled cache consistency. This leads to more expensive and nonscalable designs. Therefore, we are concentrating on softwaremore » methods for ensuring cache consistency that would allow for affordable and scalable multiprocessing systems. Unfortunately, providing software control is nontrivial, both for the compiler writer and for the application programmer. For this reason we are developing a debugging environment that will facilitate the development of compiler-based techniques and will help the programmer to tune his or her application using explicit cache management mechanisms. We extend the notion of a race condition for IBM Shared Memory System POWER/4, taking into consideration its noncoherent caches, and propose techniques for detection of false sharing problems. Identification of the stale data problem is discussed as well, and solutions are suggested.« less
NASA Astrophysics Data System (ADS)
Tanaka, Kiyoshi; Takano, Shuichi; Sugimura, Tatsuo
2000-10-01
In this work we focus on the indexed triangle strips that is an extended representation of triangle strips to improve the efficiency for geometrical transformation of vertices, and present a method to construct optimum indexed triangle strips using Genetic Algorithm (GA) for real-time visualization. The main objective of this work is how to optimally construct indexed triangle strips by improving the ratio that reuses the data stored in the cash memory and simultaneously reducing the total index numbers with GA. Simulation results verify that the average index numbers and cache miss ratio per polygon cold be small, and consequently the total visualization time required for the optimum solution obtained by this scheme could be remarkably reduced.
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-29
..., Armstrong Hall, Room 201, 14 E. Cache La Poudre, Colorado Springs, CO 80903, telephone (719) 389-6201..., Room 201, 14 E. Cache La Poudre, Colorado Springs, CO 80903, telephone (719) 389-6201, before April 29...
A set-associative, fault-tolerant cache design
NASA Technical Reports Server (NTRS)
Lamet, Dan; Frenzel, James F.
1992-01-01
The design of a defect-tolerant control circuit for a set-associative cache memory is presented. The circuit maintains the stack ordering necessary for implementing the Least Recently Used (LRU) replacement algorithm. A discussion of programming techniques for bypassing defective blocks is included.
Enhancement web proxy cache performance using Wrapper Feature Selection methods with NB and J48
NASA Astrophysics Data System (ADS)
Mahmoud Al-Qudah, Dua'a.; Funke Olanrewaju, Rashidah; Wong Azman, Amelia
2017-11-01
Web proxy cache technique reduces response time by storing a copy of pages between client and server sides. If requested pages are cached in the proxy, there is no need to access the server. Due to the limited size and excessive cost of cache compared to the other storages, cache replacement algorithm is used to determine evict page when the cache is full. On the other hand, the conventional algorithms for replacement such as Least Recently Use (LRU), First in First Out (FIFO), Least Frequently Use (LFU), Randomized Policy etc. may discard important pages just before use. Furthermore, using conventional algorithm cannot be well optimized since it requires some decision to intelligently evict a page before replacement. Hence, most researchers propose an integration among intelligent classifiers and replacement algorithm to improves replacement algorithms performance. This research proposes using automated wrapper feature selection methods to choose the best subset of features that are relevant and influence classifiers prediction accuracy. The result present that using wrapper feature selection methods namely: Best First (BFS), Incremental Wrapper subset selection(IWSS)embedded NB and particle swarm optimization(PSO)reduce number of features and have a good impact on reducing computation time. Using PSO enhance NB classifier accuracy by 1.1%, 0.43% and 0.22% over using NB with all features, using BFS and using IWSS embedded NB respectively. PSO rises J48 accuracy by 0.03%, 1.91 and 0.04% over using J48 classifier with all features, using IWSS-embedded NB and using BFS respectively. While using IWSS embedded NB fastest NB and J48 classifiers much more than BFS and PSO. However, it reduces computation time of NB by 0.1383 and reduce computation time of J48 by 2.998.
Ironside, Kirsten E; Mattson, David J; Theimer, Tad; Jansen, Brian; Holton, Brandon; Arundel, Terence; Peters, Michael; Sexton, Joseph O; Edwards, Thomas C
2017-01-01
Many studies of animal movement have focused on directed versus area-restricted movement, which rely on correlations between step-length and turn-angles and on stationarity through time to define behavioral states. Although these approaches might apply well to grazing in patchy landscapes, species that either feed for short periods on large, concentrated food sources or cache food exhibit movements that are difficult to model using the traditional metrics of turn-angle and step-length alone. We used GPS telemetry collected from a prey-caching predator, the cougar ( Puma concolor, Linnaeus ), to test whether combining metrics of site recursion, spatiotemporal clustering, speed, and turning into an index of movement using partial sums, improves the ability to identify caching behavior. The index was used to identify changes in movement characteristics over time and segment paths into behavioral classes. The identification of behaviors from the Path Identification Index (PII) was evaluated using field investigations of cougar activities at GPS locations. We tested for statistical stationarity across behaviors for use of topographic view-sheds. Changes in the frequency and duration of PII were useful for identifying seasonal activities such as migration, gestation, and denning. The comparison of field investigations of cougar activities to behavioral PII classes resulted in an overall classification accuracy of 81%. Changes in behaviors were reflected in cougars' use of topographic view-sheds, resulting in statistical nonstationarity over time, and revealed important aspects of hunting behavior. Incorporating metrics of site recursion and spatiotemporal clustering revealed the temporal structure in movements of a caching forager. The movement index PII, shows promise for identifying behaviors in species that frequently return to specific locations such as food caches, watering holes, or dens, and highlights the potential role memory and cognitive abilities play in determining animal movements.
Urhan, A Utku; Brodin, Anders
2015-05-01
Scatter hoarding birds are known for their accurate spatial memory. In a previous experiment, we tested the retrieval accuracy in marsh tits in a typical laboratory set-up for this species. We also tested the performance of humans in this experimental set-up. Somewhat unexpectedly, humans performed much better than marsh tits. In the first five attempts, humans relocated almost 90 % of the caches they had hidden 5 h earlier. Marsh tits only relocated 25 % in the first five attempts and just above 40 % in the first ten attempts. Typically, in this type of experiment, the birds will be caching and retrieving many times in the same sites in the same experimental room. This is very different from the conditions in nature where hoarding parids only cache once in a caching site. Hence, it is possible that memories from previous sessions will disturb the formation of new memories. If there is such proactive interference, the prediction is that success should decay over sessions. Here, we have designed an experiment to investigate whether there is such memory interference in this type of experiment. We allowed marsh tits and humans to cache and retrieve in three repeated sessions without prior experience of the arena. The performance did not change over sessions, and on average, marsh tits correctly visited around 25 % of the caches in the first five attempts. The corresponding success in humans was constant across sessions, and it was around 90 % on average. We conclude that the somewhat poor performance of the marsh tits did not depend on proactive memory interference. We also discuss other possible reasons for why marsh tits in general do not perform better in laboratory experiments.
The history of scatter hoarding studies.
Brodin, Anders
2010-03-27
In this review, I will present an overview of the development of the field of scatter hoarding studies. Scatter hoarding is a conspicuous behaviour and it has been observed by humans for a long time. Apart from an exceptional experimental study already published in 1720, it started with observational field studies of scatter hoarding birds in the 1940s. Driven by a general interest in birds, several ornithologists made large-scale studies of hoarding behaviour in species such as nutcrackers and boreal titmice. Scatter hoarding birds seem to remember caching locations accurately, and it was shown in the 1960s that successful retrieval is dependent on a specific part of the brain, the hippocampus. The study of scatter hoarding, spatial memory and the hippocampus has since then developed into a study system for evolutionary studies of spatial memory. In 1978, a game theoretical paper started the era of modern studies by establishing that a recovery advantage is necessary for individual hoarders for the evolution of a hoarding strategy. The same year, a combined theoretical and empirical study on scatter hoarding squirrels investigated how caches should be spaced out in order to minimize cache loss, a phenomenon sometimes called optimal cache density theory. Since then, the scatter hoarding paradigm has branched into a number of different fields: (i) theoretical and empirical studies of the evolution of hoarding, (ii) field studies with modern sampling methods, (iii) studies of the precise nature of the caching memory, (iv) a variety of studies of caching memory and its relationship to the hippocampus. Scatter hoarding has also been the subject of studies of (v) coevolution between scatter hoarding animals and the plants that are dispersed by these.
Massively parallel algorithms for trace-driven cache simulations
NASA Technical Reports Server (NTRS)
Nicol, David M.; Greenberg, Albert G.; Lubachevsky, Boris D.
1991-01-01
Trace driven cache simulation is central to computer design. A trace is a very long sequence of reference lines from main memory. At the t(exp th) instant, reference x sub t is hashed into a set of cache locations, the contents of which are then compared with x sub t. If at the t sup th instant x sub t is not present in the cache, then it is said to be a miss, and is loaded into the cache set, possibly forcing the replacement of some other memory line, and making x sub t present for the (t+1) sup st instant. The problem of parallel simulation of a subtrace of N references directed to a C line cache set is considered, with the aim of determining which references are misses and related statistics. A simulation method is presented for the Least Recently Used (LRU) policy, which regradless of the set size C runs in time O(log N) using N processors on the exclusive read, exclusive write (EREW) parallel model. A simpler LRU simulation algorithm is given that runs in O(C log N) time using N/log N processors. Timings are presented of the second algorithm's implementation on the MasPar MP-1, a machine with 16384 processors. A broad class of reference based line replacement policies are considered, which includes LRU as well as the Least Frequently Used and Random replacement policies. A simulation method is presented for any such policy that on any trace of length N directed to a C line set runs in the O(C log N) time with high probability using N processors on the EREW model. The algorithms are simple, have very little space overhead, and are well suited for SIMD implementation.
Upadhyay, Amit A.; Fleetwood, Aaron D.; Adebali, Ogun; ...
2016-04-06
Cellular receptors usually contain a designated sensory domain that recognizes the signal. Per/Arnt/Sim (PAS) domains are ubiquitous sensors in thousands of species ranging from bacteria to humans. Although PAS domains were described as intracellular sensors, recent structural studies revealed PAS-like domains in extracytoplasmic regions in several transmembrane receptors. However, these structurally defined extracellular PAS-like domains do not match sequence-derived PAS domain models, and thus their distribution across the genomic landscape remains largely unknown. Here we show that structurally defined extracellular PAS-like domains belong to the Cache superfamily, which is homologous to, but distinct from the PAS superfamily. Our newly builtmore » computational models enabled identification of Cache domains in tens of thousands of signal transduction proteins including those from important pathogens and model organisms.Moreover, we show that Cache domains comprise the dominant mode of extracellular sensing in prokaryotes.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Upadhyay, Amit A.; Fleetwood, Aaron D.; Adebali, Ogun
Cellular receptors usually contain a designated sensory domain that recognizes the signal. Per/Arnt/Sim (PAS) domains are ubiquitous sensors in thousands of species ranging from bacteria to humans. Although PAS domains were described as intracellular sensors, recent structural studies revealed PAS-like domains in extracytoplasmic regions in several transmembrane receptors. However, these structurally defined extracellular PAS-like domains do not match sequence-derived PAS domain models, and thus their distribution across the genomic landscape remains largely unknown. Here we show that structurally defined extracellular PAS-like domains belong to the Cache superfamily, which is homologous to, but distinct from the PAS superfamily. Our newly builtmore » computational models enabled identification of Cache domains in tens of thousands of signal transduction proteins including those from important pathogens and model organisms.Moreover, we show that Cache domains comprise the dominant mode of extracellular sensing in prokaryotes.« less
Memory for Multiple Cache Locations and Prey Quantities in a Food-Hoarding Songbird
Armstrong, Nicola; Garland, Alexis; Burns, K. C.
2012-01-01
Most animals can discriminate between pairs of numbers that are each less than four without training. However, North Island robins (Petroica longipes), a food-hoarding songbird endemic to New Zealand, can discriminate between quantities of items as high as eight without training. Here we investigate whether robins are capable of other complex quantity discrimination tasks. We test whether their ability to discriminate between small quantities declines with (1) the number of cache sites containing prey rewards and (2) the length of time separating cache creation and retrieval (retention interval). Results showed that subjects generally performed above-chance expectations. They were equally able to discriminate between different combinations of prey quantities that were hidden from view in 2, 3, and 4 cache sites from between 1, 10, and 60 s. Overall results indicate that North Island robins can process complex quantity information involving more than two discrete quantities of items for up to 1 min long retention intervals without training. PMID:23293622
Memory for multiple cache locations and prey quantities in a food-hoarding songbird.
Armstrong, Nicola; Garland, Alexis; Burns, K C
2012-01-01
Most animals can discriminate between pairs of numbers that are each less than four without training. However, North Island robins (Petroica longipes), a food-hoarding songbird endemic to New Zealand, can discriminate between quantities of items as high as eight without training. Here we investigate whether robins are capable of other complex quantity discrimination tasks. We test whether their ability to discriminate between small quantities declines with (1) the number of cache sites containing prey rewards and (2) the length of time separating cache creation and retrieval (retention interval). Results showed that subjects generally performed above-chance expectations. They were equally able to discriminate between different combinations of prey quantities that were hidden from view in 2, 3, and 4 cache sites from between 1, 10, and 60 s. Overall results indicate that North Island robins can process complex quantity information involving more than two discrete quantities of items for up to 1 min long retention intervals without training.
Arias, Michelle R.; Alpers, Charles N.; Marvin-DiPasquale, Mark C.; Fuller, Christopher C.; Agee, Jennifer L.; Sneed, Michelle; Morita, Andrew Y.; Salas, Antonia
2017-10-31
Cache Creek Settling Basin was constructed in 1937 to trap sediment from Cache Creek before delivery to the Yolo Bypass, a flood conveyance for the Sacramento River system that is tributary to the Sacramento–San Joaquin Delta. Sediment management options being considered by stakeholders in the Cache Creek Settling Basin include sediment excavation; however, that could expose sediments containing elevated mercury concentrations from historical mercury mining in the watershed. In cooperation with the California Department of Water Resources, the U.S. Geological Survey undertook sediment coring campaigns in 2011–12 (1) to describe lateral and vertical distributions of mercury concentrations in deposits of sediment in the Cache Creek Settling Basin and (2) to improve constraint of estimates of the rate of sediment deposition in the basin.Sediment cores were collected in the Cache Creek Settling Basin, Yolo County, California, during October 2011 at 10 locations and during August 2012 at 5 other locations. Total core depths ranged from approximately 4.6 to 13.7 meters (15 to 45 feet), with penetration to about 9.1 meters (30 feet) at most locations. Unsplit cores were logged for two geophysical parameters (gamma bulk density and magnetic susceptibility); then, selected cores were split lengthwise. One half of each core was then photographed and archived, and the other half was subsampled. Initial subsamples from the cores (20-centimeter composite samples from five predetermined depths in each profile) were analyzed for total mercury, methylmercury, total reduced sulfur, iron speciation, organic content (as the percentage of weight loss on ignition), and grain-size distribution. Detailed follow-up subsampling (3-centimeter intervals) was done at six locations along an east-west transect in the southern part of the Cache Creek Settling Basin and at one location in the northern part of the basin for analyses of total mercury; organic content; and cesium-137, which was used for dating. This report documents site characteristics; field and laboratory methods; and results of the analyses of each core section and subsample of these sediment cores, including associated quality-assurance and quality-control data.
Analysis of cache for streaming tape drive
NASA Technical Reports Server (NTRS)
Chinnaswamy, V.
1993-01-01
A tape subsystem consists of a controller and a tape drive. Tapes are used for backup, data interchange, and software distribution. The backup operation is addressed. During a backup operation, data is read from disk, processed in CPU, and then sent to tape. The processing speeds of a disk subsystem, CPU, and a tape subsystem are likely to be different. A powerful CPU can read data from a fast disk, process it, and supply the data to the tape subsystem at a faster rate than the tape subsystem can handle. On the other hand, a slow disk drive and a slow CPU may not be able to supply data fast enough to keep a tape drive busy all the time. The backup process may supply data to tape drive in bursts. Each burst may be followed by an idle period. Depending on the nature of the file distribution in the disk, the input stream to the tape subsystem may vary significantly during backup. To compensate for these differences and optimize the utilization of a tape subsystem, a cache or buffer is introduced in the tape controller. Most of the tape drives today are streaming tape drives. A streaming tape drive goes into reposition when there is no data from the controller. Once the drive goes into reposition, the controller can receive data, but it cannot supply data to the tape drive until the drive completes its reposition. A controller can also receive data from the host and send data to the tape drive at the same time. The relationship of cache size, host transfer rate, drive transfer rate, reposition, and ramp up times for optimal performance of the tape subsystem are investigated. Formulas developed will also show the advantages of cache watermarks to increase the streaming time of the tape drive, maximum loss due to insufficient cache, tradeoffs between cache and reposition times and the effectiveness of cache on a streaming tape drive due to idle times or interruptions due in host transfers. Several mathematical formulas are developed to predict the performance of the tape drive. Some examples are given illustrating the usefulness of these formulas. Finally, a summary and some conclusions are provided.
Security in the CernVM File System and the Frontier Distributed Database Caching System
NASA Astrophysics Data System (ADS)
Dykstra, D.; Blomer, J.
2014-06-01
Both the CernVM File System (CVMFS) and the Frontier Distributed Database Caching System (Frontier) distribute centrally updated data worldwide for LHC experiments using http proxy caches. Neither system provides privacy or access control on reading the data, but both control access to updates of the data and can guarantee the authenticity and integrity of the data transferred to clients over the internet. CVMFS has since its early days required digital signatures and secure hashes on all distributed data, and recently Frontier has added X.509-based authenticity and integrity checking. In this paper we detail and compare the security models of CVMFS and Frontier.
ROPEC - ROtary PErcussive Coring Drill for Mars Sample Return
NASA Technical Reports Server (NTRS)
Chu, Philip; Spring, Justin; Zacny, Kris
2014-01-01
The ROtary Percussive Coring Drill is a light weight, flight-like, five-actuator drilling system prototype designed to acquire core material from rock targets for the purposes of Mars Sample Return. In addition to producing rock cores for sample caching, the ROPEC drill can be integrated with a number of end effectors to perform functions such as rock surface abrasion, dust and debris removal, powder and regolith acquisition, and viewing of potential cores prior to caching. The ROPEC drill and its suite of end effectors have been demonstrated with a five degree of freedom Robotic Arm mounted to a mobility system with a prototype sample cache and bit storage station.
USDA-ARS?s Scientific Manuscript database
Positive interactions among individual plants (facilitation) may often enhance seedling survival in stressful environments. Many granivorous small mammal species cache groups of seeds for future consumption in shallowly buried scatterhoards, and seeds of many plant species germinate and establish ag...
78 FR 2655 - Uinta-Wasatch-Cache National Forest; Utah; Ogden Travel Plan Project
Federal Register 2010, 2011, 2012, 2013, 2014
2013-01-14
...-Wasatch-Cache National Forest; Utah; Ogden Travel Plan Project AGENCY: Forest Service, USDA. ACTION... prepare a supplement to the Ogden Travel Plan Revision Final Supplemental Environmental Impact Statement (FSEIS). The Ogden Travel Plan Revision FSEIS evaluated six alternatives for possible travel management...
Cache-site selection in Clark's Nutcracker (Nucifraga columbiana)
Teresa J. Lorenz; Kimberly A. Sullivan; Amanda V. Bakian; Carol A. Aubry
2011-01-01
Clark's Nutcracker (Nucifraga Columbiana) is one of the most specialized scatter-hoarding birds, considered a seed disperser for four species of pines (Pinus spp.), as well as an obligate coevolved mutualist of White bark Pine (P. albicaulis). Cache-site selection has not been formally studied in Clark...
Soft-core processor study for node-based architectures.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Van Houten, Jonathan Roger; Jarosz, Jason P.; Welch, Benjamin James
2008-09-01
Node-based architecture (NBA) designs for future satellite projects hold the promise of decreasing system development time and costs, size, weight, and power and positioning the laboratory to address other emerging mission opportunities quickly. Reconfigurable Field Programmable Gate Array (FPGA) based modules will comprise the core of several of the NBA nodes. Microprocessing capabilities will be necessary with varying degrees of mission-specific performance requirements on these nodes. To enable the flexibility of these reconfigurable nodes, it is advantageous to incorporate the microprocessor into the FPGA itself, either as a hardcore processor built into the FPGA or as a soft-core processor builtmore » out of FPGA elements. This document describes the evaluation of three reconfigurable FPGA based processors for use in future NBA systems--two soft cores (MicroBlaze and non-fault-tolerant LEON) and one hard core (PowerPC 405). Two standard performance benchmark applications were developed for each processor. The first, Dhrystone, is a fixed-point operation metric. The second, Whetstone, is a floating-point operation metric. Several trials were run at varying code locations, loop counts, processor speeds, and cache configurations. FPGA resource utilization was recorded for each configuration. Cache configurations impacted the results greatly; for optimal processor efficiency it is necessary to enable caches on the processors. Processor caches carry a penalty; cache error mitigation is necessary when operating in a radiation environment.« less
Sex, estradiol, and spatial memory in a food-caching corvid.
Rensel, Michelle A; Ellis, Jesse M S; Harvey, Brigit; Schlinger, Barney A
2015-09-01
Estrogens significantly impact spatial memory function in mammalian species. Songbirds express the estrogen synthetic enzyme aromatase at relatively high levels in the hippocampus and there is evidence from zebra finches that estrogens facilitate performance on spatial learning and/or memory tasks. It is unknown, however, whether estrogens influence hippocampal function in songbirds that naturally exhibit memory-intensive behaviors, such as cache recovery observed in many corvid species. To address this question, we examined the impact of estradiol on spatial memory in non-breeding Western scrub-jays, a species that routinely participates in food caching and retrieval in nature and in captivity. We also asked if there were sex differences in performance or responses to estradiol. Utilizing a combination of an aromatase inhibitor, fadrozole, with estradiol implants, we found that while overall cache recovery rates were unaffected by estradiol, several other indices of spatial memory, including searching efficiency and efficiency to retrieve the first item, were impaired in the presence of estradiol. In addition, males and females differed in some performance measures, although these differences appeared to be a consequence of the nature of the task as neither sex consistently out-performed the other. Overall, our data suggest that a sustained estradiol elevation in a food-caching bird impairs some, but not all, aspects of spatial memory on an innate behavioral task, at times in a sex-specific manner. Copyright © 2015 Elsevier Inc. All rights reserved.
SEX, ESTRADIOL, AND SPATIAL MEMORY IN A FOOD-CACHING CORVID
Rensel, Michelle A.; Ellis, Jesse M.S.; Harvey, Brigit; Schlinger, Barney A.
2015-01-01
Estrogens significantly impact spatial memory function in mammalian species. Songbirds express the estrogen synthetic enzyme aromatase at relatively high levels in the hippocampus and there is evidence from zebra finches that estrogens facilitate performance on spatial learning and/or memory tasks. It is unknown, however, whether estrogens influence hippocampal function in songbirds that naturally exhibit memory-intensive behaviors, such as cache recovery observed in many corvid species. To address this question, we examined the impact of estradiol on spatial memory in non-breeding Western scrub-jays, a species that routinely participates in food caching and retrieval in nature and in captivity. We also asked if there were sex differences in performance or responses to estradiol. Utilizing a combination of an aromatase inhibitor, fadrozole, with estradiol implants, we found that while overall cache recovery rates were unaffected by estradiol, several other indices of spatial memory, including searching efficiency and efficiency to retrieve the first item, were impaired in the presence of estradiol. In addition, males and females differed in some performance measures, although these differences appeared to be a consequence of the nature of the task as neither sex consistently out-performed the other. Overall, our data suggest that a sustained estradiol elevation in a food-caching bird impairs some, but not all, aspects of spatial memory on an innate behavioral task, at times in a sex-specific manner. PMID:26232613
Pravosudov, V V; Roth, T C; Forister, M L; Ladage, L D; Burg, T M; Braun, M J; Davidson, B S
2012-09-01
Food-caching birds rely on stored food to survive the winter, and spatial memory has been shown to be critical in successful cache recovery. Both spatial memory and the hippocampus, an area of the brain involved in spatial memory, exhibit significant geographic variation linked to climate-based environmental harshness and the potential reliance on food caches for survival. Such geographic variation has been suggested to have a heritable basis associated with differential selection. Here, we ask whether population genetic differentiation and potential isolation among multiple populations of food-caching black-capped chickadees is associated with differences in memory and hippocampal morphology by exploring population genetic structure within and among groups of populations that are divergent to different degrees in hippocampal morphology. Using mitochondrial DNA and 583 AFLP loci, we found that population divergence in hippocampal morphology is not significantly associated with neutral genetic divergence or geographic distance, but instead is significantly associated with differences in winter climate. These results are consistent with variation in a history of natural selection on memory and hippocampal morphology that creates and maintains differences in these traits regardless of population genetic structure and likely associated gene flow. Published 2012. This article is a US Government work and is in the public domain in the USA.
Two Balls' Collision of Mass Ratio 3:1
NASA Astrophysics Data System (ADS)
Ogawara, Yasuo; Hull, Michael M.
2018-04-01
Students will sometimes ask why momentum and kinetic energy concepts are both necessary. When physics teachers demonstrate situations that require both an understanding of kinetic energy and momentum, a favorite is Newton's cradle, or a comparable demonstration of two balls of equal mass hitting each other. However, in addition to the case of two balls of equal mass, if a ball hits another ball of three times the mass with equal speed, the results are also interesting, and, like the equal-mass demonstration, both kinetic energy and momentum are critical for understanding the motion.
Claire A. Zugmeyer; John L. Koprowski
2009-01-01
Severe disturbance may alter or eliminate important habitat structure that helps preserve food caches of foodhoarding species. Recent recolonization of an insect-damaged forest by the endangered Mt. Graham red squirrel (Tamiasciurus hudsonicus grahamensis) provided an opportunity to examine habitat selection for midden (cache) sites following...
Random Fill Cache Architecture (Preprint)
2014-10-01
a concrete example, we show how the cache collision attack works to extract the AES encryption keys (e.g., in the OpenSSL implementation of AES). AES...each round are implemented as table lookups for performance reasons. OpenSSL uses ten 1-KB lookup tables, five for encryption and five for decryption
76 FR 14372 - Uinta-Wasatch-Cache National Forest Resource Advisory Committee
Federal Register 2010, 2011, 2012, 2013, 2014
2011-03-16
... Street, Salt Lake City, Utah. Written comments should be sent to Loyal Clark, Uinta-Wasatch-Cache... open to the public. The following business will be conducted: (1) Review Forest Service project approval letter, (2) discuss travel budget, and (3) review new proposals. Persons who wish to bring related...
Differentiated strategies for improving streaming service quality
NASA Astrophysics Data System (ADS)
An, Hui; Chen, Xin-Meng
2005-02-01
With the explosive growth of streaming services, users are becoming more and more sensitive to its quality of service. To handle these problems, the research community focuses of the application of caching and replication techniques. But most approaches try to find specific strategies of caching of replication that suit for streaming service characteristics and to design some kind of universal policy to deal with all streaming objects. This paper explores the combination of caching and replication for improving streaming service quality and demonstrates that it makes sense to incorporate two technologies. It provides a system model and discusses some related issues of how to determining a refreshable streaming object and which refreshment policies a refreshable object should use.
Hydrologic data for the Cache Creek-Bear Thrust environmental impact statement near Jackson, Wyoming
Craig, G.S.; Ringen, B.H.; Cox, E.R.
1981-01-01
Information on the quantity and quality of surface and ground water in an area of concern for the Cache Creek-Bear Thrust Environmental Impact Statement in northwestern Wyoming is presented without interpretation. The environmental impact statement is being prepared jointly by the U.S. Geological Survey and the U.S. Forest Service and concerns proposed exploration and development of oil and gas on leased Federal land near Jackson, Wyoming. Information includes data from a gaging station on Cache Creek and from wells, springs, and miscellaneous sites on streams. Data include streamflow, chemical and suspended-sediment quality of streams, and the occurrence and chemical quality of ground water. (USGS)
NASA Astrophysics Data System (ADS)
Zecha, Stefanie; Regelous, Anette
2017-04-01
National Geoparks are restricted areas incorporating educational resources of great importance in promoting education for sustainable development, mobilizing knowledge inherent to the EarthSciences. Different methods can be used to implement the education of sustainability. Here we present possibilities for National Geoparks to support sustainability focusing on new media and EarthCaches based on the data set of the "EarthCachers International EarthCaching" conference in Goslar in October 2015. Using an empirical study designed by ourselves we collected actual information about the environmental consciousness of Earthcachers. The data set was analyzed using SPSS and statistical methods. Here we present the results and their consequences for National Geoparks.
Tschanz, JoAnn T.; Norton, Maria C.; Zandi, Peter P.; Lyketsos, Constantine G.
2014-01-01
The Cache County Study on Memory in Aging is a longitudinal, population-based study of Alzheimer's disease (AD) and other dementias. Initiated in 1995 and extending to 2013, the study has followed over 5,000 elderly residents of Cache County, Utah (USA) for over twelve years. Achieving a 90% participation rate at enrollment, and spawning two ancillary projects, the study has contributed to the literature on genetic, psychosocial and environmental risk factors for AD, late life cognitive decline, and the clinical progression of dementia after its onset. This paper describes the major study contributions to the literature on AD and dementia. PMID:24423221
Tschanz, Joann T; Norton, Maria C; Zandi, Peter P; Lyketsos, Constantine G
2013-12-01
The Cache County Study on Memory in Aging is a longitudinal, population-based study of Alzheimer's disease (AD) and other dementias. Initiated in 1995 and extending to 2013, the study has followed over 5,000 elderly residents of Cache County, Utah (USA) for over twelve years. Achieving a 90% participation rate at enrolment, and spawning two ancillary projects, the study has contributed to the literature on genetic, psychosocial and environmental risk factors for AD, late-life cognitive decline, and the clinical progression of dementia after its onset. This paper describes the major study contributions to the literature on AD and dementia.
Improving Seasonal Climate Predictability in the Colorado River Basin for Enhanced Decision Support
NASA Astrophysics Data System (ADS)
Rajagopal, S.; Mahmoud, M. I.
2016-12-01
The water resource management community is increasingly seeking skillful seasonal climate forecasts with long lead times. But predicting wet or dry climate with sufficient lead time (3 months) for a season (especially winter) in the Colorado River Basin (CRB) is a challenging problem. The typical approach taken to predicting winter climate is based on using climate indices and climate models to predict precipitation or streamflow in the Colorado River Basin. In addition to this approach; which may have a long lead time, water supply forecasts are also generated based on current observations by the Colorado River Forecast Center. Recently, the effects of coupled atmospheric-ocean phenomena such as ENSO over North America, and atmospheric circulation patterns at the 500 mb pressure level, which make the CRB wet or dry, have been studied separately. In the current work we test whether combining climate indices and circulation patterns improve predictability in the CRB. To accomplish this, the atmospheric circulation data from the Earth System Research Laboratory (ESRL) and climate indices data from the Climate Prediction Center were combined using logical functions. To quantify the skill in prediction, statistics such as the hit ratio and false alarm ratio were computed. The results from using a combination of climate indices and atmospheric circulation patterns suggest that there is an improvement in the prediction skill with hit ratios higher than 0.8, as compared to using either predictor individually (which typically produced a hit ratio of 0.6). Based on this result, there is value in using this hybrid approach when compared to a black box statistical model, as the climate index is an analog to the moisture availability and the right atmospheric circulation pattern helps in transporting that moisture to the Basin.
Resource-Efficient Data-Intensive System Designs for High Performance and Capacity
2015-09-01
76, 79, 80, and 81.] [9] Anirudh Badam, KyoungSoo Park, Vivek S. Pai, and Larry L. Peterson. HashCache: cache storage for the next billion. In Proc...Jeffrey Dean, Sanjay Ghemawat, Wilson C. Hsieh, Deborah A. Wallach, Mike Burrows , Tushar Chandra, Andrew Fikes, and Robert E. Gruber. Bigtable: A
Retrospective Cognition by Food-Caching Western Scrub-Jays
ERIC Educational Resources Information Center
de Kort, S.R.; Dickinson, A.; Clayton, N.S.
2005-01-01
Episodic-like memory, the retrospective component of cognitive time travel in animals, needs to fulfil three criteria to meet the behavioral properties of episodic memory as defined for humans. Here, we review results obtained with the cache-recovery paradigm with western scrub-jays and conclude that they fulfil these three criteria. The jays…
76 FR 16640 - Petitions for Modification of Existing Mandatory Safety Standards
Federal Register 2010, 2011, 2012, 2013, 2014
2011-03-24
... standard to permit an alternative method of compliance to allow additional outby storage caches of Self.... The petitioner further states that: (a) Additional SCSR outby storage caches will be placed a maximum of 2,000 feet apart in beltlines and return air courses; (b) these additional SCSR outby storage...
Acorn Caching in Tree Squirrels: Teaching Hypothesis Testing in the Park
ERIC Educational Resources Information Center
McEuen, Amy B.; Steele, Michael A.
2012-01-01
We developed an exercise for a university-level ecology class that teaches hypothesis testing by examining acorn preferences and caching behavior of tree squirrels (Sciurus spp.). This exercise is easily modified to teach concepts of behavioral ecology for earlier grades, particularly high school, and provides students with a theoretical basis for…
USDA-ARS?s Scientific Manuscript database
Seeds of many plant species are dispersed by seed-caching rodents that place groups of seeds in superficially-buried scatterhoard caches. A case in point is provided by an important forage plant on arid western rangelands, Indian ricegrass (Oryzopsis hymenoides), for which seedling recruitment comes...
Using shadow page cache to improve isolated drivers performance.
Zheng, Hao; Dong, Xiaoshe; Wang, Endong; Chen, Baoke; Zhu, Zhengdong; Liu, Chengzhe
2015-01-01
With the advantage of the reusability property of the virtualization technology, users can reuse various types and versions of existing operating systems and drivers in a virtual machine, so as to customize their application environment. In order to prevent users' virtualization environments being impacted by driver faults in virtual machine, Chariot examines the correctness of driver's write operations by the method of combining a driver's write operation capture and a driver's private access control table. However, this method needs to keep the write permission of shadow page table as read-only, so as to capture isolated driver's write operations through page faults, which adversely affect the performance of the driver. Based on delaying setting frequently used shadow pages' write permissions to read-only, this paper proposes an algorithm using shadow page cache to improve the performance of isolated drivers and carefully study the relationship between the performance of drivers and the size of shadow page cache. Experimental results show that, through the shadow page cache, the performance of isolated drivers can be greatly improved without impacting Chariot's reliability too much.
Using Shadow Page Cache to Improve Isolated Drivers Performance
Dong, Xiaoshe; Wang, Endong; Chen, Baoke; Zhu, Zhengdong; Liu, Chengzhe
2015-01-01
With the advantage of the reusability property of the virtualization technology, users can reuse various types and versions of existing operating systems and drivers in a virtual machine, so as to customize their application environment. In order to prevent users' virtualization environments being impacted by driver faults in virtual machine, Chariot examines the correctness of driver's write operations by the method of combining a driver's write operation capture and a driver's private access control table. However, this method needs to keep the write permission of shadow page table as read-only, so as to capture isolated driver's write operations through page faults, which adversely affect the performance of the driver. Based on delaying setting frequently used shadow pages' write permissions to read-only, this paper proposes an algorithm using shadow page cache to improve the performance of isolated drivers and carefully study the relationship between the performance of drivers and the size of shadow page cache. Experimental results show that, through the shadow page cache, the performance of isolated drivers can be greatly improved without impacting Chariot's reliability too much. PMID:25815373
Exploitation of pocket gophers and their food caches by grizzly bears
Mattson, D.J.
2004-01-01
I investigated the exploitation of pocket gophers (Thomomys talpoides) by grizzly bears (Ursus arctos horribilis) in the Yellowstone region of the United States with the use of data collected during a study of radiomarked bears in 1977-1992. My analysis focused on the importance of pocket gophers as a source of energy and nutrients, effects of weather and site features, and importance of pocket gophers to grizzly bears in the western contiguous United States prior to historical extirpations. Pocket gophers and their food caches were infrequent in grizzly bear feces, although foraging for pocket gophers accounted for about 20-25% of all grizzly bear feeding activity during April and May. Compared with roots individually excavated by bears, pocket gopher food caches were less digestible but more easily dug out. Exploitation of gopher food caches by grizzly bears was highly sensitive to site and weather conditions and peaked during and shortly after snowmelt. This peak coincided with maximum success by bears in finding pocket gopher food caches. Exploitation was most frequent and extensive on gently sloping nonforested sites with abundant spring beauty (Claytonia lanceolata) and yampah (Perdieridia gairdneri). Pocket gophers are rare in forests, and spring beauty and yampah roots are known to be important foods of both grizzly bears and burrowing rodents. Although grizzly bears commonly exploit pocket gophers only in the Yellowstone region, this behavior was probably widespread in mountainous areas of the western contiguous United States prior to extirpations of grizzly bears within the last 150 years.
Strickland, Dan; Kielstra, Brian; Ryan Norris, D
2011-12-01
Variation in habitat quality can have important consequences for fitness and population dynamics. For food-caching species, a critical determinant of habitat quality is normally the density of storable food, but it is also possible that quality is driven by the ability of habitats to preserve food items. The food-caching gray jay (Perisoreus canadensis) occupies year-round territories in the coniferous boreal and subalpine forests of North America, but does not use conifer seed crops as a source of food. Over the last 33 years, we found that the occupancy rate of territories in Algonquin Park (ON, Canada) has declined at a higher rate in territories with a lower proportion of conifers compared to those with a higher proportion. Individuals occupying territories with a low proportion of conifers were also less likely to successfully fledge young. Using chambers to simulate food caches, we conducted an experiment to examine the hypothesis that coniferous trees are better able to preserve the perishable food items stored in summer and fall than deciduous trees due to their antibacterial and antifungal properties. Over a 1-4 month exposure period, we found that mealworms, blueberries, and raisins all lost less weight when stored on spruce and pine trees compared to deciduous and other coniferous trees. Our results indicate a novel mechanism to explain how habitat quality may influence the fitness and population dynamics of food-caching animals, and has important implications for understanding range limits for boreal breeding animals.
Ironside, Kirsten E.; Mattson, David J.; Theimer, Tad; Jansen, Brian; Holton, Brandon; Arundel, Terry; Peters, Michael; Sexton, Joseph O.; Edwards, Thomas C.
2017-01-01
Relocation studies of animal movement have focused on directed versus area restricted movement, which rely on correlations between step-length and turn angles, along with a degree of stationarity through time to define behavioral states. Although these approaches may work well for grazing foraging strategies in a patchy landscape, species that do not spend a significant amount of time searching out and gathering small dispersed food items, but instead feed for short periods on large, concentrated sources or cache food result in movements that maybe difficult to analyze using turning and velocity alone. We use GPS telemetry collected from a prey-caching predator, the cougar (Puma concolor), to test whether adding additional movement metrics capturing site recursion, to the more traditional velocity and turning, improve the ability to identify behaviors. We evaluated our movement index’s ability to identify behaviors using field investigations. We further tested for statistical stationarity across behaviors for use of topographic view-sheds. We found little correlation between turn angle, velocity, tortuosity, and site fidelity and combined them into a movement index used to identify movement paths (temporally autocorrelated movements) related to fast directed movements (taxis), area restricted movements (search), and prey caching (foraging). Changes in the frequency and duration of these movements were helpful for identifying seasonal activities such as migration and denning in females. Comparison of field investigations of cougar activities to behavioral classes defined using the movement index and found an overall classification accuracy of 81%. Changes in behaviors resulted in changes in how cougars used topographic view-sheds, showing statistical non-stationarity over time. The movement index shows promise for identifying behaviors in species that frequently return to specific locations such as food caches, watering holes, or dens, and highlights the role memory and cognitive abilities may play in determining animal movements. With the addition of measures capturing site recursion the temporal structure in movements of a caching forager was revealed.
Cache and energy efficient algorithms for Nussinov's RNA Folding.
Zhao, Chunchun; Sahni, Sartaj
2017-12-06
An RNA folding/RNA secondary structure prediction algorithm determines the non-nested/pseudoknot-free structure by maximizing the number of complementary base pairs and minimizing the energy. Several implementations of Nussinov's classical RNA folding algorithm have been proposed. Our focus is to obtain run time and energy efficiency by reducing the number of cache misses. Three cache-efficient algorithms, ByRow, ByRowSegment and ByBox, for Nussinov's RNA folding are developed. Using a simple LRU cache model, we show that the Classical algorithm of Nussinov has the highest number of cache misses followed by the algorithms Transpose (Li et al.), ByRow, ByRowSegment, and ByBox (in this order). Extensive experiments conducted on four computational platforms-Xeon E5, AMD Athlon 64 X2, Intel I7 and PowerPC A2-using two programming languages-C and Java-show that our cache efficient algorithms are also efficient in terms of run time and energy. Our benchmarking shows that, depending on the computational platform and programming language, either ByRow or ByBox give best run time and energy performance. The C version of these algorithms reduce run time by as much as 97.2% and energy consumption by as much as 88.8% relative to Classical and by as much as 56.3% and 57.8% relative to Transpose. The Java versions reduce run time by as much as 98.3% relative to Classical and by as much as 75.2% relative to Transpose. Transpose achieves run time and energy efficiency at the expense of memory as it takes twice the memory required by Classical. The memory required by ByRow, ByRowSegment, and ByBox is the same as that of Classical. As a result, using the same amount of memory, the algorithms proposed by us can solve problems up to 40% larger than those solvable by Transpose.
Gonthier, G.J.; Kleiss, B.A.
1996-01-01
The U.S. Geological Survey, working in cooperation with the U.S. Army Corps of Engineers, Waterways Experiment Station, collected surface-water and ground-water data from 119 wells and 13 staff gages from September 1989 to September 1992 to describe ground-water flow patterns and water budget in the Black Swamp, a bottomland forested wetland in eastern Arkansas. The study area was between two streamflow gaging stations located about 30.5 river miles apart on the Cache River. Ground-water flow was from northwest to southeast with some diversion toward the Cache River. Hydraulic connection between the surface water and the alluvial aquifer is indicated by nearly equal changes in surface-water and ground-water levels near the Cache River. Diurnal fluctuations of hydraulic head ranged from more than 0 to 0.38 feet and were caused by evapotranspiration. Changes in hydraulic head of the alluvial aquifer beneath the wetland lagged behind stage fluctuations and created the potential for changes in ground-water movement. Differences between surface-water levels in the wetland and stage of the Cache River created a frequently occurring local ground-water flow condition in which surface water in the wetland seeped into the upper part of the alluvial aquifer and then seeped into the Cache River. When the Cache River flooded the wetland, ground water consistently seeped to the surface during falling surface-water stage and surface water seeped into the ground during rising surface-water stage. Ground-water flow was a minor component of the water budget, accounting for less than 1 percent of both inflow and outflow. Surface-water drainage from the study area through diversion canals was not accounted for in the water budget and may be the reason for a surplus of water in the budget. Even though ground-water flow volume is small compared to other water budget components, ground-water seepage to the wetland surface may still be vital to some wetland functions.
Single-hit mechanism of tumour cell killing by radiation.
Chapman, J D
2003-02-01
To review the relative importance of the single-hit mechanism of radiation killing for tumour response to 1.8-2.0 Gy day(-1) fractions and to low dose-rate brachytherapy. Tumour cell killing by ionizing radiation is well described by the linear-quadratic equation that contains two independent components distinguished by dose kinetics. Analyses of tumour cell survival curves that contain six or more dose points usually provide good estimates of the alpha- and beta-inactivation coefficients. Superior estimates of tumour cell intrinsic radiosensitivity are obtained when synchronized populations are employed. The characteristics of single-hit inactivation of tumour cells are reviewed and compared with the characteristics of beta-inactivation. Potential molecular targets associated with single-hit inactivation are discussed along with strategies for potentiating cell killing by this mechanism. The single-hit mechanism of tumour cell killing shows no dependence on dose-rate and, consequently, no evidence of sublethal damage repair. It is uniquely potentiated by high linear-energy-transfer radiation, exhibits a smaller oxygen enhancement ratio and exhibits a larger indirect effect by hydroxyl radicals than the beta-mechanism. alpha-inactivation coefficients vary slightly throughout interphase but mitotic cells exhibit extremely high alpha-coefficients in the range of those observed for lymphocytes and some repair-deficient cells. Evidence is accumulating to suggest that chromatin in compacted form could be a radiation-hypersensitive target associated with single-hit radiation killing. Analyses of tumour cell survival curves demonstrate that it is the single-hit mechanism (alpha) that determines the majority of cell killing after doses of 2Gy and that this mechanism is highly variable between tumour cell lines. The characteristics of single-hit inactivation are qualitatively and quantitatively distinct from those of beta-inactivation. Compacted chromatin in tumour cells should be further investigated as a radiation-hypersensitive target that could be modulated for therapeutic advantage.
Predictive Cache Modeling and Analysis
2011-11-01
metaheuristic /bin-packing algorithm to optimize task placement based on task communication characterization. Our previous work on task allocation showed...Cache Miss Minimization Technology To efficiently explore combinations and discover nearly-optimal task-assignment algorithms , we extended to our...it was possible to use our algorithmic techniques to decrease network bandwidth consumption by ~25%. In this effort, we adapted these existing
Statistical Inference-Based Cache Management for Mobile Learning
ERIC Educational Resources Information Center
Li, Qing; Zhao, Jianmin; Zhu, Xinzhong
2009-01-01
Supporting efficient data access in the mobile learning environment is becoming a hot research problem in recent years, and the problem becomes tougher when the clients are using light-weight mobile devices such as cell phones whose limited storage space prevents the clients from holding a large cache. A practical solution is to store the cache…
Geo-Caching: Place-Based Discovery of Virginia State Parks and Museums
ERIC Educational Resources Information Center
Gray, Howard Richard
2007-01-01
The use of Global Positioning Systems (GPS) units has exploded in recent years along with the computer technology to access this data-based information. Geo-caching is an exciting game using GPS that provides place-based information regarding the public lands, facilities and cultural heritage programs within the Virginia Parks and Museum system.…
Code of Federal Regulations, 2011 CFR
2011-07-01
.../Attainment. The area surrounding Brigham City, as described by the following Townships or the portions of the following Townships in Box Elder County: T9N 2W, T9N R1W, T8N 2W Cache County, UT (part): Cache County.... The area surrounding Grantsville, as described by the following Townships or the portions of the...
Code of Federal Regulations, 2010 CFR
2010-07-01
.../Attainment. The area surrounding Brigham City, as described by the following Townships or the portions of the following Townships in Box Elder County: T9N 2W, T9N R1W, T8N 2W Cache County, UT (part): Cache County.... The area surrounding Grantsville, as described by the following Townships or the portions of the...
Managing coherence via put/get windows
Blumrich, Matthias A [Ridgefield, CT; Chen, Dong [Croton on Hudson, NY; Coteus, Paul W [Yorktown Heights, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Hoenicke, Dirk [Ossining, NY; Ohmacht, Martin [Yorktown Heights, NY
2011-01-11
A method and apparatus for managing coherence between two processors of a two processor node of a multi-processor computer system. Generally the present invention relates to a software algorithm that simplifies and significantly speeds the management of cache coherence in a message passing parallel computer, and to hardware apparatus that assists this cache coherence algorithm. The software algorithm uses the opening and closing of put/get windows to coordinate the activated required to achieve cache coherence. The hardware apparatus may be an extension to the hardware address decode, that creates, in the physical memory address space of the node, an area of virtual memory that (a) does not actually exist, and (b) is therefore able to respond instantly to read and write requests from the processing elements.
Managing coherence via put/get windows
Blumrich, Matthias A [Ridgefield, CT; Chen, Dong [Croton on Hudson, NY; Coteus, Paul W [Yorktown Heights, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Hoenicke, Dirk [Ossining, NY; Ohmacht, Martin [Yorktown Heights, NY
2012-02-21
A method and apparatus for managing coherence between two processors of a two processor node of a multi-processor computer system. Generally the present invention relates to a software algorithm that simplifies and significantly speeds the management of cache coherence in a message passing parallel computer, and to hardware apparatus that assists this cache coherence algorithm. The software algorithm uses the opening and closing of put/get windows to coordinate the activated required to achieve cache coherence. The hardware apparatus may be an extension to the hardware address decode, that creates, in the physical memory address space of the node, an area of virtual memory that (a) does not actually exist, and (b) is therefore able to respond instantly to read and write requests from the processing elements.
NASA Astrophysics Data System (ADS)
Bauerdick, L. A. T.; Bloom, K.; Bockelman, B.; Bradley, D. C.; Dasu, S.; Dost, J. M.; Sfiligoi, I.; Tadel, A.; Tadel, M.; Wuerthwein, F.; Yagil, A.; Cms Collaboration
2014-06-01
Following the success of the XRootd-based US CMS data federation, the AAA project investigated extensions of the federation architecture by developing two sample implementations of an XRootd, disk-based, caching proxy. The first one simply starts fetching a whole file as soon as a file open request is received and is suitable when completely random file access is expected or it is already known that a whole file be read. The second implementation supports on-demand downloading of partial files. Extensions to the Hadoop Distributed File System have been developed to allow for an immediate fallback to network access when local HDFS storage fails to provide the requested block. Both cache implementations are in pre-production testing at UCSD.
Hagawane, T N; Gaikwad, R V; Kshirsagar, N A
2016-05-01
Despite advances in therapy and overall medical care, acute lung injury (ALI)/acute respiratory distress syndrome (ARDS) management remains a problem. Hence the objective of this study was to develop a rat model that mimics human ALI/ARDS. Four groups of Wistar rats, 48 per group were treated with (i) intratracheal (IT) lipopolysaccharide (LPS) (5 mg/kg) dissolved in normal saline (NS), (ii) intravenous (iv) oleic acid (OA) (250 μl/kg) suspension in bovine serum albumin (BSA), (iii) dual hit: IT LPS (2 mg/kg) dissolved in NS and iv OA (100 μl/kg) and (iv) control group: IT NS and iv BSA. From each group at set periods of time various investigations like chest x-rays, respiratory rate (RR), tidal volume (TV), total cell count, differential cell count, total protein count and cytokine levels in bronchoalveolar lavage fluid (BALF), lung wet/dry weight ratio and histopathological examination were done. It was noted that the respiratory rate, and tumour necrosis factor-α (TNF-α) levels were significantly higher at 4 h in the dual hit group as compared to LPS, OA and control groups. Interleukin-6 (IL-6) levels were significantly higher in the dual hit group as compared to LPS at 8 and 24 h, OA at 8 h and control (at all time intervals) group. IL-1β levels were significantly higher in LPS and dual hit groups at all time intervals, but not in OA and control groups. The injury induced in dual hit group was earlier and more sustained as compared to LPS and OA alone. The lung pathology and changes in respiration functions produced by the dual hit model were closer to the diagnostic criteria of ALI/ARDS in terms of clinical manifestations and pulmonary injury and the injury persisted longer as compared to LPS and OA single hit model. Therefore, the ARDS model produced by the dual hit method was closer to the diagnostic criteria of ARDS in terms of clinical manifestations and pulmonary injury.
2013-01-01
Background The adoption of health information technology has been recommended as a viable mechanism for improving quality of care and patient health outcomes. However, the capacity of health information technology (i.e., availability and use of multiple and advanced functionalities), particularly in federally qualified health centers (FQHCs) on improving quality of care is not well understood. We examined associations between health information technology (HIT) capacity at FQHCs and quality of care, measured by the receipt of discharge summary, frequency of patients receiving reminders/notifications for preventive care/follow-up care, and timely appointment for specialty care. Methods The analyses used 2009 data from the National Survey of Federally Qualified Health Centers. The study included 776 of the FQHCs that participated in the survey. We examined the extent of HIT use and tested the hypothesis that level of HIT capacity is associated with quality of care. Multivariable logistic regressions, reporting unadjusted and adjusted odds ratios, were used to examine whether ‘FQHCs’ HIT capacity’ is associated with the outcome measures. Results The results showed a positive association between health information technology capacity and quality of care. FQHCs with higher HIT capacity were significantly more likely to have improved quality of care, measured by the receipt of discharge summaries (OR=1.43; CI=1.01, 2.40), the use of a patient notification system for preventive and follow-up care (OR=1.74; CI=1.23, 2.45), and timely appointment for specialty care (OR=1.77; CI=1.24, 2.53). Conclusions Our findings highlight the promise of HIT in improving quality of care, particularly for vulnerable populations who seek care at FQHCs. The results also show that FQHCs may not be maximizing the benefits of HIT. Efforts to implement HIT must include strategies that facilitate the implementation of comprehensive and advanced functionalities, as well as promote meaningful use of these systems. Further examination of the role of health information systems in clinical decision-making and improvements in patient outcomes are needed to better understand the benefits of HIT in improving overall quality of care. PMID:23363660
Frimpong, Jemima A; Jackson, Bradford E; Stewart, LaShonda M; Singh, Karan P; Rivers, Patrick A; Bae, Sejong
2013-01-31
The adoption of health information technology has been recommended as a viable mechanism for improving quality of care and patient health outcomes. However, the capacity of health information technology (i.e., availability and use of multiple and advanced functionalities), particularly in federally qualified health centers (FQHCs) on improving quality of care is not well understood. We examined associations between health information technology (HIT) capacity at FQHCs and quality of care, measured by the receipt of discharge summary, frequency of patients receiving reminders/notifications for preventive care/follow-up care, and timely appointment for specialty care. The analyses used 2009 data from the National Survey of Federally Qualified Health Centers. The study included 776 of the FQHCs that participated in the survey. We examined the extent of HIT use and tested the hypothesis that level of HIT capacity is associated with quality of care. Multivariable logistic regressions, reporting unadjusted and adjusted odds ratios, were used to examine whether 'FQHCs' HIT capacity' is associated with the outcome measures. The results showed a positive association between health information technology capacity and quality of care. FQHCs with higher HIT capacity were significantly more likely to have improved quality of care, measured by the receipt of discharge summaries (OR=1.43; CI=1.01, 2.40), the use of a patient notification system for preventive and follow-up care (OR=1.74; CI=1.23, 2.45), and timely appointment for specialty care (OR=1.77; CI=1.24, 2.53). Our findings highlight the promise of HIT in improving quality of care, particularly for vulnerable populations who seek care at FQHCs. The results also show that FQHCs may not be maximizing the benefits of HIT. Efforts to implement HIT must include strategies that facilitate the implementation of comprehensive and advanced functionalities, as well as promote meaningful use of these systems. Further examination of the role of health information systems in clinical decision-making and improvements in patient outcomes are needed to better understand the benefits of HIT in improving overall quality of care.
The use of argatroban for carotid endarterectomy in heparin-induced thrombocytopenia.
Hallman, Sarah E; Hebbar, Latha; Robison, Jay; Uber, Walter E
2005-04-01
Heparin-induced thrombocytopenia (HIT) is a major obstacle in cardiovascular surgeries. In this case report, we used argatroban, a direct thrombin inhibitor, to achieve and maintain anticoagulation for carotid endarterectomy. Unlike heparin, the direct thrombin inhibitors bind directly to thrombin, bypassing antithrombin III and the potential to precipitate HIT. A bolus of argatroban 150 microg/kg followed by an infusion of 5 microg . kg(-1) . min(-1) was used, and adequate anticoagulation was demonstrated with multiple laboratory tests (at 28 min, prothrombin time = 29.8 s, partial thromboplastin time = 69.1 s, international normalized ratio = 3.52 s, and activated clotting time = 220 s). The surgery was successful, and the patient was discharged the next day with no postoperative neurologic sequelae or other complications. We conclude that argatroban can be used safely and successfully for carotid endarterectomy in a patient with a history of HIT.
Lu, Kai; Wang, Li; Wang, Changying; Yang, Yuan; Hu, Dayi; Ding, Rongjing
2015-08-01
The optimal aerobic exercise training (AET) protocol for patients following myocardial infarction (MI) has remained under debate. The present study therefore aimed to compare the effects of continuous moderate-intensity training (CMT) and high-intensity interval training (HIT) on cardiac functional recovery, and to investigate the potential associated mechanisms in a post-MI rat model. Female Sprague Dawley rats (8-10 weeks old) undergoing MI or sham surgery were subsequently submitted to CMT or HIT, or kept sedentary for eight weeks. Prior to and following AET, echocardiographic parameters and exercise capacity of the rats were measured. Western blotting was used to evaluate the levels of apoptosis and associated signaling pathway protein expression. The concentrations of biomarkers of oxidative stress were also determined by ELISA assay. Messenger (m)RNA levels and activity of the key enzymes for glycolysis and fatty acid oxidation, as well as the rate of adenosine triphosphate (ATP) synthesis, were also measured. Compared with the MI group, exercise capacity and cardiac function were significantly improved following AET, particularly following HIT. Left ventricular ejection fraction and fraction shortening were further improved in the MI-HIT group in comparison to that of the MI-CMT group. The two forms of AET almost equally attenuated apoptosis of the post-infarction myocardium. CMT and HIT also alleviated oxidative stress by decreasing the concentration of malondialdehyde and increasing the concentration of superoxide dismutase and glutathione peroxidase (GPx). In particular, HIT induced a greater increase in the concentration of GPx than that of CMT. AET, and HIT in particular, significantly increased the levels of mRNA and the maximal activity of phosphofructokinase-1 and carnitine palmitoyl transferase-1, as well as the maximal ratio of ATP synthesis. In addition, compared with the MI group, the expression of signaling proteins PI3K, Akt, p38mapk and AMPK was significantly altered in the MI-CMT and MI-HIT groups. HIT was superior to CMT in its ability to improve cardiac function and exercise capability in a post-MI rat model. HIT was also superior to CMT with regard to attenuating oxidative stress and improving glucolipid metabolism of the post-MI myocardium.
Janine Rice; Tim Bardsley; Pete Gomben; Dustin Bambrough; Stacey Weems; Sarah Leahy; Christopher Plunkett; Charles Condrat; Linda A. Joyce
2017-01-01
Watersheds on the Uinta-Wasatch-Cache and Ashley National Forests provide many ecosystem services, and climate change poses a risk to these services. We developed a watershed vulnerability assessment to provide scientific information for land managers facing the challenge of managing these watersheds. Literature-based information and expert elicitation is used to...
Hide And Seek GPS And Geocaching In The Classroom
ERIC Educational Resources Information Center
Lary, Lynn M.
2004-01-01
In short, geocaching is a high-tech, worldwide treasure hunt (geocaches can now be found in more than 180 countries) where a person hides a cache for others to find. Generally, the cache is some type of waterproof container that contains a log book and an assortment of goodies, such as lottery tickets, toys, photo books for cachers to fill with…
Performance assessment of EMR systems based on post-relational database.
Yu, Hai-Yan; Li, Jing-Song; Zhang, Xiao-Guang; Tian, Yu; Suzuki, Muneou; Araki, Kenji
2012-08-01
Post-relational databases provide high performance and are currently widely used in American hospitals. As few hospital information systems (HIS) in either China or Japan are based on post-relational databases, here we introduce a new-generation electronic medical records (EMR) system called Hygeia, which was developed with the post-relational database Caché and the latest platform Ensemble. Utilizing the benefits of a post-relational database, Hygeia is equipped with an "integration" feature that allows all the system users to access data-with a fast response time-anywhere and at anytime. Performance tests of databases in EMR systems were implemented in both China and Japan. First, a comparison test was conducted between a post-relational database, Caché, and a relational database, Oracle, embedded in the EMR systems of a medium-sized first-class hospital in China. Second, a user terminal test was done on the EMR system Izanami, which is based on the identical database Caché and operates efficiently at the Miyazaki University Hospital in Japan. The results proved that the post-relational database Caché works faster than the relational database Oracle and showed perfect performance in the real-time EMR system.
Dynamically programmable cache
NASA Astrophysics Data System (ADS)
Nakkar, Mouna; Harding, John A.; Schwartz, David A.; Franzon, Paul D.; Conte, Thomas
1998-10-01
Reconfigurable machines have recently been used as co- processors to accelerate the execution of certain algorithms or program subroutines. The problems with the above approach include high reconfiguration time and limited partial reconfiguration. By far the most critical problems are: (1) the small on-chip memory which results in slower execution time, and (2) small FPGA areas that cannot implement large subroutines. Dynamically Programmable Cache (DPC) is a novel architecture for embedded processors which offers solutions to the above problems. To solve memory access problems, DPC processors merge reconfigurable arrays with the data cache at various cache levels to create a multi-level reconfigurable machines. As a result DPC machines have both higher data accessibility and FPGA memory bandwidth. To solve the limited FPGA resource problem, DPC processors implemented multi-context switching (Virtualization) concept. Virtualization allows implementation of large subroutines with fewer FPGA cells. Additionally, DPC processors can parallelize the execution of several operations resulting in faster execution time. In this paper, the speedup improvement for DPC machines are shown to be 5X faster than an Altera FLEX10K FPGA chip and 2X faster than a Sun Ultral SPARC station for two different algorithms (convolution and motion estimation).
A Chemogenomic Screen Reveals Novel Snf1p/AMPK Independent Regulators of Acetyl-CoA Carboxylase.
Bozaquel-Morais, Bruno L; Madeira, Juliana B; Venâncio, Thiago M; Pacheco-Rosa, Thiago; Masuda, Claudio A; Montero-Lomeli, Monica
2017-01-01
Acetyl-CoA carboxylase (Acc1p) is a key enzyme in fatty acid biosynthesis and is essential for cell viability. To discover new regulators of its activity, we screened a Saccharomyces cerevisiae deletion library for increased sensitivity to soraphen A, a potent Acc1p inhibitor. The hits identified in the screen (118 hits) were filtered using a chemical-phenotype map to exclude those associated with pleiotropic drug resistance. This enabled the identification of 82 ORFs that are genetic interactors of Acc1p. The main functional clusters represented by these hits were "transcriptional regulation", "protein post-translational modifications" and "lipid metabolism". Further investigation of the "transcriptional regulation" cluster revealed that soraphen A sensitivity is poorly correlated with ACC1 transcript levels. We also studied the three top unknown ORFs that affected soraphen A sensitivity: SOR1 (YDL129W), SOR2 (YIL092W) and SOR3 (YJR039W). Since the C18/C16 ratio of lipid acyl lengths reflects Acc1p activity levels, we evaluated this ratio in the three mutants. Deletion of SOR2 and SOR3 led to reduced acyl lengths, suggesting that Acc1p is indeed down-regulated in these strains. Also, these mutants showed no differences in Snf1p/AMPK activation status and deletion of SNF1 in these backgrounds did not revert soraphen A sensitivity completely. Furthermore, plasmid maintenance was reduced in sor2Δ strain and this trait was shared with 18 other soraphen A sensitive hits. In summary, our screen uncovered novel Acc1p Snf1p/AMPK-independent regulators.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Perelmutov, T.; Bakken, J.; Petravick, D.
Storage Resource Managers (SRMs) are middleware components whose function is to provide dynamic space allocation and file management on shared storage components on the Grid[1,2]. SRMs support protocol negotiation and reliable replication mechanism. The SRM standard supports independent SRM implementations, allowing for a uniform access to heterogeneous storage elements. SRMs allow site-specific policies at each location. Resource Reservations made through SRMs have limited lifetimes and allow for automatic collection of unused resources thus preventing clogging of storage systems with ''orphan'' files. At Fermilab, data handling systems use the SRM management interface to the dCache Distributed Disk Cache [5,6] and themore » Enstore Tape Storage System [15] as key components to satisfy current and future user requests [4]. The SAM project offers the SRM interface for its internal caches as well.« less
The Use of Proxy Caches for File Access in a Multi-Tier Grid Environment
NASA Astrophysics Data System (ADS)
Brun, R.; Duellmann, D.; Ganis, G.; Hanushevsky, A.; Janyst, L.; Peters, A. J.; Rademakers, F.; Sindrilaru, E.
2011-12-01
The use of proxy caches has been extensively studied in the HEP environment for efficient access of database data and showed significant performance with only very moderate operational effort at higher grid tiers (T2, T3). In this contribution we propose to apply the same concept to the area of file access and analyse the possible performance gains, operational impact on site services and applicability to different HEP use cases. Base on a proof-of-concept studies with a modified XROOT proxy server we review the cache efficiency and overheads for access patterns of typical ROOT based analysis programs. We conclude with a discussion of the potential role of this new component at the different tiers of a distributed computing grid.
Evolution of magnetic disk subsystems
NASA Astrophysics Data System (ADS)
Kaneko, Satoru
1994-06-01
The higher recording density of magnetic disk realized today has brought larger storage capacity per unit and smaller form factors. If the required access performance per MB is constant, the performance of large subsystems has to be several times better. This article describes mainly the technology for improving the performance of the magnetic disk subsystems and the prospects of their future evolution. Also considered are 'crosscall pathing' which makes the data transfer channel more effective, 'disk cache' which improves performance coupling with solid state memory technology, and 'RAID' which improves the availability and integrity of disk subsystems by organizing multiple disk drives in a subsystem. As a result, it is concluded that since the performance of the subsystem is dominated by that of the disk cache, maximation of the performance of the disk cache subsystems is very important.
The Use of Proxy Caches for File Access in a Multi-Tier Grid Environment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brun, R.; Dullmann, D.; Ganis, G.
2012-04-19
The use of proxy caches has been extensively studied in the HEP environment for efficient access of database data and showed significant performance with only very moderate operational effort at higher grid tiers (T2, T3). In this contribution we propose to apply the same concept to the area of file access and analyze the possible performance gains, operational impact on site services and applicability to different HEP use cases. Base on a proof-of-concept studies with a modified XROOT proxy server we review the cache efficiency and overheads for access patterns of typical ROOT based analysis programs. We conclude with amore » discussion of the potential role of this new component at the different tiers of a distributed computing grid.« less
Managing coherence via put/get windows
DOE Office of Scientific and Technical Information (OSTI.GOV)
Blumrich, Matthias A; Chen, Dong; Coteus, Paul W
A method and apparatus for managing coherence between two processors of a two processor node of a multi-processor computer system. Generally the present invention relates to a software algorithm that simplifies and significantly speeds the management of cache coherence in a message passing parallel computer, and to hardware apparatus that assists this cache coherence algorithm. The software algorithm uses the opening and closing of put/get windows to coordinate the activated required to achieve cache coherence. The hardware apparatus may be an extension to the hardware address decode, that creates, in the physical memory address space of the node, an areamore » of virtual memory that (a) does not actually exist, and (b) is therefore able to respond instantly to read and write requests from the processing elements.« less
Tang, Wenming; Liu, Guixiong; Li, Yuzhong; Tan, Daji
2017-01-01
High data transmission efficiency is a key requirement for an ultrasonic phased array with multi-group ultrasonic sensors. Here, a novel FIFOs scheduling algorithm was proposed and the data transmission efficiency with hardware technology was improved. This algorithm includes FIFOs as caches for the ultrasonic scanning data obtained from the sensors with the output data in a bandwidth-sharing way, on the basis of which an optimal length ratio of all the FIFOs is achieved, allowing the reading operations to be switched among all the FIFOs without time slot waiting. Therefore, this algorithm enhances the utilization ratio of the reading bandwidth resources so as to obtain higher efficiency than the traditional scheduling algorithms. The reliability and validity of the algorithm are substantiated after its implementation in the field programmable gate array (FPGA) technology, and the bandwidth utilization ratio and the real-time performance of the ultrasonic phased array are enhanced. PMID:29035345
Sidhu, Swati; Datta, Aparajita
2015-01-01
Rodents affect the post-dispersal fate of seeds by acting either as on-site seed predators or as secondary dispersers when they scatter-hoard seeds. The tropical forests of north-east India harbour a high diversity of little-studied terrestrial murid and hystricid rodents. We examined the role played by these rodents in determining the seed fates of tropical evergreen tree species in a forest site in north-east India. We selected ten tree species (3 mammal-dispersed and 7 bird-dispersed) that varied in seed size and followed the fates of 10,777 tagged seeds. We used camera traps to determine the identity of rodent visitors, visitation rates and their seed-handling behavior. Seeds of all tree species were handled by at least one rodent taxon. Overall rates of seed removal (44.5%) were much higher than direct on-site seed predation (9.9%), but seed-handling behavior differed between the terrestrial rodent groups: two species of murid rodents removed and cached seeds, and two species of porcupines were on-site seed predators. In addition, a true cricket, Brachytrupes sp., cached seeds of three species underground. We found 309 caches formed by the rodents and the cricket; most were single-seeded (79%) and seeds were moved up to 19 m. Over 40% of seeds were re-cached from primary cache locations, while about 12% germinated in the primary caches. Seed removal rates varied widely amongst tree species, from 3% in Beilschmiedia assamica to 97% in Actinodaphne obovata. Seed predation was observed in nine species. Chisocheton cumingianus (57%) and Prunus ceylanica (25%) had moderate levels of seed predation while the remaining species had less than 10% seed predation. We hypothesized that seed traits that provide information on resource quantity would influence rodent choice of a seed, while traits that determine resource accessibility would influence whether seeds are removed or eaten. Removal rates significantly decreased (p < 0.001) while predation rates increased (p = 0.06) with seed size. Removal rates were significantly lower for soft seeds (p = 0.002), whereas predation rates were significantly higher on soft seeds (p = 0.01). Our results show that murid rodents play a very important role in affecting the seed fates of tropical trees in the Eastern Himalayas. We also found that the different rodent groups differed in their seed handling behavior and responses to changes in seed characteristics. PMID:26247616
2002-07-01
our general model include: (1) service user (SU), (2) service manager (SM), and (3) service cache manager ( SCM ), where the SCM is an optional...maintained by SMs that satisfy specific requirements. Where employed, the SCM operates as an intermediary, matching advertised SDs of SMs to...Directory Service Agent (optional) not applicableLookup ServiceService Cache Manager ( SCM ) Service URL Service Type Service Attributes Template URL
Environmental Impact Analysis Process, Groom Mountain Range, Lincoln County, Nevada
1985-10-01
bases clustered around springs, temporary camps, rock shelters , quarries, lithic scatters, rock art, pinyon caches, pot drops, isolates, and historic...include pinyon caches and rock shelters with associated historic artifacts and many of the spring sites. These sites provide an unusual research...Management. (b) Proposed Action: Renewed Withdrawal of Groom Mountain Range Addition to Nellis Air Force Bombing and Gunnery Range, Lincoln County, Nevada. (c
Thomas L. Foti
2001-01-01
Relationships between forest vegetation and soil were reconstructed from field notes of the 1846 Public Land Survey (PLS) along a portion of the Cache River including Black Swamp. Locations of corners were digitized long with species,diameter,and distance from section or quarter-section corners. Trees were grouped for analysis according to occurrence on groups of...
Domain Wall Fermion Inverter on Pentium 4
NASA Astrophysics Data System (ADS)
Pochinsky, Andrew
2005-03-01
A highly optimized domain wall fermion inverter has been developed as part of the SciDAC lattice initiative. By designing the code to minimize memory bus traffic, it achieves high cache reuse and performance in excess of 2 GFlops for out of L2 cache problem sizes on a GigE cluster with 2.66 GHz Xeon processors. The code uses the SciDAC QMP communication library.
Killing of a muskox, Ovibus moschatus, by two wolves, Canis lupis, and subsequent caching
Mech, L. David; Adams, Layne G.
1999-01-01
The killing of a cow Muskox (Ovibos moschatus) by two Wolves (Canis lupus) in 5 minutes during summer on Ellesmere Island is described. After two of the four feedings observed, one Wolf cached a leg and regurgitated food as far as 2.3 km away and probably farther. The implications of this behavior for deriving food-consumption estimates are discussed.
Minimizing End-to-End Interference in I/O Stacks Spanning Shared Multi-Level Buffer Caches
ERIC Educational Resources Information Center
Patrick, Christina M.
2011-01-01
This thesis presents an end-to-end interference minimizing uniquely designed high performance I/O stack that spans multi-level shared buffer cache hierarchies accessing shared I/O servers to deliver a seamless high performance I/O stack. In this thesis, I show that I can build a superior I/O stack which minimizes the inter-application interference…
Evaluating Fragment Construction Policies for SDT Systems
2006-01-01
allocates a fragment and begins translation. Once a termination condition is met, Strata emits any trampolines that are necessary. Trampolines are pieces... trampolines (unless its target previously exists in the fragment cache). Once a CTI’s target instruction becomes available in the fragment cache, the CTI is...linked directly to the destination, avoiding future uses of the trampoline . This mechanism is called Fragment Linking and avoids significant overhead
dCache, Sync-and-Share for Big Data
NASA Astrophysics Data System (ADS)
Millar, AP; Fuhrmann, P.; Mkrtchyan, T.; Behrmann, G.; Bernardt, C.; Buchholz, Q.; Guelzow, V.; Litvintsev, D.; Schwank, K.; Rossi, A.; van der Reest, P.
2015-12-01
The availability of cheap, easy-to-use sync-and-share cloud services has split the scientific storage world into the traditional big data management systems and the very attractive sync-and-share services. With the former, the location of data is well understood while the latter is mostly operated in the Cloud, resulting in a rather complex legal situation. Beside legal issues, those two worlds have little overlap in user authentication and access protocols. While traditional storage technologies, popular in HEP, are based on X.509, cloud services and sync-and-share software technologies are generally based on username/password authentication or mechanisms like SAML or Open ID Connect. Similarly, data access models offered by both are somewhat different, with sync-and-share services often using proprietary protocols. As both approaches are very attractive, dCache.org developed a hybrid system, providing the best of both worlds. To avoid reinventing the wheel, dCache.org decided to embed another Open Source project: OwnCloud. This offers the required modern access capabilities but does not support the managed data functionality needed for large capacity data storage. With this hybrid system, scientists can share files and synchronize their data with laptops or mobile devices as easy as with any other cloud storage service. On top of this, the same data can be accessed via established mechanisms, like GridFTP to serve the Globus Transfer Service or the WLCG FTS3 tool, or the data can be made available to worker nodes or HPC applications via a mounted filesystem. As dCache provides a flexible authentication module, the same user can access its storage via different authentication mechanisms; e.g., X.509 and SAML. Additionally, users can specify the desired quality of service or trigger media transitions as necessary, thus tuning data access latency to the planned access profile. Such features are a natural consequence of using dCache. We will describe the design of the hybrid dCache/OwnCloud system, report on several months of operations experience running it at DESY, and elucidate the future road-map.
Water transfer and loss in hit-and-run collisions
NASA Astrophysics Data System (ADS)
Burger, C.; Maindl, T. I.; Schäfer, C.
2017-09-01
This work focuses on transfer and loss of volatiles, like water, in hit-and-run collisions, where especially the smaller one of the colliding pair is often stripped of considerable amounts of its initial volatile content, but still survives the encounter more or less intact. We find water losses up to 75 percent in a single collision, depending on various parameters, especially velocity, impact angle and mass ratio, but also on the total colliding mass. The physical state, especially vaporization of volatiles, is found to be particularly important in collisions of approximately Mars-sized bodies, with high impact energies, but still potentially easy volatile escape.
Cao, Lin; Xiao, Zhishu; Guo, Cong; Chen, Jin
2011-09-01
Local extinction or population decline of large frugivorous vertebrates as primary seed dispersers, caused by human disturbance and habitat change, might lead to dispersal limitation of many large-seeded fruit trees. However, it is not known whether or not scatter-hoarding rodents as secondary seed dispersers can help maintain natural regeneration (e.g. seed dispersal) of these frugivore-dispersed trees in the face of the functional reduction or loss of primary seed dispersers. In the present study, we investigated how scatter-hoarding rodents affect the fate of tagged seeds of a large-seeded fruit tree (Scleropyrum wallichianum Arnott, 1838, Santalaceae) from seed fall to seedling establishment in a heavily defaunated tropical forest in the Xishuangbanna region of Yunnan Province, in southwest China, in 2007 and 2008. Our results show that: (i) rodents removed nearly all S. wallichianum seeds in both years; (ii) a large proportion (2007, 75%; 2008, 67.5%) of the tagged seeds were cached individually in the surface soil or under leaf litters; (iii) dispersal distance of primary caches was further in 2007 (19.6 ± 14.6 m) than that in 2008 (14.1 ± 11.6 m), and distance increased as rodents recovered and moved seeds from primary caches into subsequent caching sites; and (iv) part of the cached seeds (2007, 3.2%; 2008, 2%) survived to the seedling stage each year. Our study suggests that by taking roles of both primary and secondary seed dispersers, scatter-hoarding rodents can play a significant role in maintaining seedling establishment of S. wallichianum, and are able to at least partly compensate for the loss of large frugivorous vertebrates in seed dispersal. © 2011 ISZS, Blackwell Publishing and IOZ/CAS.
Schindewolf, M; Paulik, M; Kroll, H; Kaufmann, R; Wolter, M; Boehncke, W-H; Lindhoff-Last, E; Recke, A; Ludwig, R J
2018-04-23
Heparins are widely prescribed for prevention and therapy of arterial and venous thromboembolic diseases. Heparin-induced skin lesions are the most frequent adverse effects of subcutaneous heparin treatment in non-surgical patients (7.5-39.8%); no data exist on surgical patients. Commonly, they are due to a delayed-type hypersensitivity reaction (DTH), but may also be a manifestation of life-threatening heparin-induced thrombocytopenia (HIT). Lesions of both entities resemble initially. The risk of HIT is highest among heparin-anticoagulated orthopedic surgery patients. To determine incidence and causes of heparin-induced skin lesions in major orthopedic surgery patients. In a prospective cohort study consecutive patients with subcutaneous low-molecular-weight heparin (LMWH) treatment were examined for cutaneous adverse effects. Further diagnostics (skin biopsy, clinical/laboratory assessment for thrombosis, bleeding, HIT, cross-allergies) were performed. Six of 316 enrolled patients (1.9%; 95% CI 0.4%-3.4%) developed heparin-induced skin lesions. All were caused by a DTH reaction, none was due to HIT or other rare heparin-associated skin diseases. Therapeutic use (dosage) of LMWH was identified as only risk factor (odds ratio: 3.1, 95%CI: 1.4-4.9; p=0.00141). In addition to DTH, 5 thromboembolic, 4 major bleeding complications but no cases of HIT or cross-allergies were observed. Orthopedic surgery patients have - unlike non-surgical patients - a low risk for heparin-induced skin lesions during LMWH treatment; all lesions were due to a DTH reaction. The risk for DTH differs considerably between individual patient cohorts. No association with HIT was observed. These data help to tailor anticoagulatory treatment individually and to increase patient safety. This article is protected by copyright. All rights reserved. This article is protected by copyright. All rights reserved.
Increasing Human Papillomavirus Vaccine Initiation among Publically-Insured Florida Adolescents
Staras, Stephanie A. S.; Vadaparampil, Susan T.; Livingston, Melvin D.; Thompson, Lindsay A.; Sanders, Ashley H.; Shenkman, Elizabeth A.
2014-01-01
Purpose We evaluated the feasibility of a multi-level intervention to increase HPV vaccine initiation among adolescents. Methods We used a four-arm factorial quasi-experimental trial to assess feasibility and short-term, preliminary effectiveness of a health system-level, gender-specific postcard campaign and an in-clinic health information technology (HIT) system. Between August to November 2013, we tested the intervention among 11–17 year olds without prior HPV vaccine claims in Florida Medicaid or Children’s Health Insurance Program encounters (2773 girls and 3350 boys) who attended or were assigned to primary care clinics in North Central Florida. Results At least one postcard was deliverable to 95% of parents. Most parents (91% boys’ and 80% girls’) who participated in the process evaluation survey (n=162) reported seeking additional information about the vaccine after receiving the postcard. Only 8% (57 of the 1062) of adolescents assigned to a HIT provider with an office visit during the study used the HIT system. When compared with arms not containing that component, HPV vaccine initiation increased with the postcard campaign [girls Odds Ratio (OR) = 1.6, 95% Confidence Interval (CI) = 1.1–2.3 and boys = not significant], the HIT system (girls OR = 1.5, 95% CI =1.0–2.3 and boys OR = 1.4, 95% CI=1.0–2.0), and the combined HIT and postcard intervention (girls OR = 2.4, 95% CI =1.4–4.3 and boys OR = 1.6, 95% CI=1.0–2.5). Conclusions A system-level postcard campaign was feasible. Despite low recruitment to the inclinic HIT system, the intervention demonstrated short-term, preliminary effectiveness similar to prior HPV vaccine interventions. PMID:25863554
Programmable partitioning for high-performance coherence domains in a multiprocessor system
Blumrich, Matthias A [Ridgefield, CT; Salapura, Valentina [Chappaqua, NY
2011-01-25
A multiprocessor computing system and a method of logically partitioning a multiprocessor computing system are disclosed. The multiprocessor computing system comprises a multitude of processing units, and a multitude of snoop units. Each of the processing units includes a local cache, and the snoop units are provided for supporting cache coherency in the multiprocessor system. Each of the snoop units is connected to a respective one of the processing units and to all of the other snoop units. The multiprocessor computing system further includes a partitioning system for using the snoop units to partition the multitude of processing units into a plurality of independent, memory-consistent, adjustable-size processing groups. Preferably, when the processor units are partitioned into these processing groups, the partitioning system also configures the snoop units to maintain cache coherency within each of said groups.
Tschanz, Joann T; Treiber, Katherine; Norton, Maria C; Welsh-Bohmer, Kathleen A; Toone, Leslie; Zandi, Peter P; Szekely, Christine A; Lyketsos, Constantine; Breitner, John C S
2005-01-01
There are several population-based studies of aging, memory, and dementia being conducted worldwide. Of these, the Cache County Study on Memory, Health and Aging is noteworthy for its large number of "oldest-old" members. This study, which has been following an initial cohort of 5,092 seniors since 1995, has reported among its major findings the role of the Apolipoprotein E gene on modifying the risk for Alzheimer's disease (AD) in males and females and identifying pharmacologic compounds that may act to reduce AD risk. This article summarizes the major findings of the Cache County study to date, describes ongoing investigations, and reports preliminary analyses on the outcome of the oldest-old in this population, the subgroup of participants who were over age 84 at the study's inception.
NASA Astrophysics Data System (ADS)
Mulder, Watson
Heterojunction with Intrinsic Thin-layer (HIT) solar cells are an important photovoltaic technology, recently reaching record power conversion efficiencies. HIT cells hold advantages over the conventional crystalline Si solar cells, such as their fabrication at lower temperatures and their shorter fabrication time. It is important to understand the electronic characteristics and transport properties of HIT cells to continue to improve their efficiencies. The fundamental measurements of a HIT solar cell with an innovative n+/p/p+ structure are presented. We also report on a series of these HIT cells fabricated on wafers with different doping concentrations, observing the relationship between doping concentration and characteristics such as open-circuit voltage and diffusion length. Nanocrystalline Silicon-Germanium (nc-SiGe) is a useful material for photovoltaic devices and photodetectors. The material features good absorption extending to the infrared region even in thin layers. Its bandgap can be adjusted between that of Si (˜1.1 eV) and Ge (˜0.7 eV) by varying the alloy composition ratio during deposition. However, there has been very little previous work to measure and understand the defect density spectrum of nc-SiGe. Defects are responsible for controlling the recombination and thus the performance of solar cell devices. Capacitance-Frequency measurements at various temperatures are used in order to estimate the trap density profile within the bandgap of nc-SiGe.
De Strijcker, Dorien; Lapauw, Bruno; Ouwens, D Margriet; Van de Velde, Dominique; Hansen, Dominique; Petrovic, Mirko; Cuvelier, Claude; Tonoli, Cajsa; Calders, Patrick
2018-06-01
To evaluate the effect of high intensity training (HIT) on physical fitness, basal respiratory exchange ratio (bRER), insulin sensitivity and muscle histology in overweight/obese men compared to continuous aerobic training (CAT). 16 male participants with overweight/obesity (age: 42-57 years, body mass index: 28-36 kg/m2) were randomized to HIT (n=8) or CAT (n=8) for 10 weeks, twice a week. HIT was composed of 10 minutes high intensity, 10 minutes continuous aerobic, 10 minutes high intensity exercises. CAT was composed of three times 10 minutes continuous exercising. Changes in anthropometry, physical and metabolic fitness were evaluated. Muscle histology (mitochondria and lipid content) was evaluated by transmission electron microscopy (TEM). HIT showed a significant increase for peak VO2 (P=0.01), for insulin sensitivity (AUC glucose (P<0,001), AUC insulin (P<0,001), OGTT composite score (P=0.007)) and a significant decrease of bRER (P<0.001) compared to CAT. Muscle mitochondrial content was significantly increased after HIT at the subsarcolemmal (P=0.004 number and P=0.001 surface) as well as the intermyofibrillar site (P<0.001 number and P=0.001 surface). High intensity training elicits stronger beneficial effects on physical fitness, basal RER, insulin sensitivity, and muscle mitochondrial content, as compared to continuous aerobic training.
Li, Qiudan; Zhan, Yongcheng; Wang, Lei; Leischow, Scott J; Zeng, Daniel Dajun
2016-07-30
The electronic cigarette (e-cigarette) market has grown rapidly in recent years. However, causes of e-cigarette related symptoms among users and their impact on health remain uncertain. This research aims to mine the potential relationships between symptoms and e-liquid components, such as propylene glycol (PG), vegetable glycerine (VG), flavor extracts, and nicotine, using user-generated data collected from Reddit. A total of 3605 e-liquid related posts from January 1st, 2011 to June 30th, 2015 were collected from Reddit. Then the patterns of VG/PG distribution among different flavors were analyzed. Next, the relationship between throat hit, which was a typical symptom of e-cigarette use, and e-liquid components was studied. Finally, other symptoms were examined based on e-liquid components and user sentiment. We discovered 3 main sets of findings: 1) We identified three groups of flavors in terms of VG/PG ratios. Fruits, cream, and nuts flavors were similar. Sweet, menthol, and seasonings flavors were classified into one group. Tobacco and beverages flavors were the third group. 2) Throat hit was analyzed and we found that menthol and tobacco flavors, as well as high ratios of PG and nicotine level, could produce more throat hit. 3) A total of 9 systems of 25 symptoms were identified and analyzed. Components including VG/PG ratio, flavor, and nicotine could be possible reasons for these symptoms. E-liquid components shown to be associated with e-cigarette use symptomology were VG/PG ratios, flavors, and nicotine levels. Future analysis could be conducted based on the structure of e-liquid components categories built in this study. Information revealed in this study could be utilized by e-cigarette users to understand the relationship between e-liquid type and symptoms experienced, by vendors to choose appropriate recipes of e-liquid, and by policy makers to develop new regulations.
Understanding Consistency Maintenance in Service Discovery Architectures in Response to Message Loss
2002-07-01
manager (SM), and (3) service cache manager ( SCM ). The SCM is an optional element not supported by all discovery protocols. These components participate...the SCM operates as an intermediary, matching advertised SDs of SMs to requirements provided by SUs. Table 1 shows how these general concepts map...Service DescriptionService ItemService Description (SD) Directory Service Agent (optional) not applicableLookup ServiceService Cache Manager ( SCM
DOE Office of Scientific and Technical Information (OSTI.GOV)
Millar, A. P.; Behrmann, G.; Bernardt, C.
With over ten years in production use dCache data storage system has evolved to match ever changing lansdcape of continually evolving storage technologies with new solutions to both existing problems and new challenges. In this paper, we present three areas of innovation in dCache: providing efficient access to data with NFS v4.1 pNFS, adoption of CDMI and WebDAV as an alternative to SRM for managing data, and integration with alternative authentication mechanisms.
Algorithms for Data Intensive Applications on Intelligent and Smart Memories
2003-03-01
editors). Parallel Algorithms and Architectures. North Holland, 1986. [8] P. Diniz . USC ISI, Personal Communication, March, 2001. [9] M. Frigo, C. E ...hierarchy as well as the Translation Lookaside Buer TLB aect the e ectiveness of cache friendly optimizations These penalties vary among...processors and cause large variations in the e ectiveness of cache performance optimizations The area of graph problems is fundamental in a wide variety of
Effect of Spatial Locality Prefetching on Structural Locality
1991-12-01
Pollution module calculates the SLC and CAM cache pollution percentages. And finally, the Generate Reference Frequency List module produces the output...3.2.5 Generate Reference Frequency List 3.2.6 Each program module in the structure chart is mapped into an Ada package. By performing this encapsulation...call routine to generate reference -- frequency list -- end if -- end loop -- close input, output, and reference files end Cache Simulator Figure 3.5
NASA Technical Reports Server (NTRS)
Clement, Bradley J.; Estlin, Tara A.; Bornstein, Benjamin J.
2013-01-01
The Mobile Thread Task Manager (MTTM) is being applied to parallelizing existing flight software to understand the benefits and to develop new techniques and architectural concepts for adapting software to multicore architectures. It allocates and load-balances tasks for a group of threads that migrate across processors to improve cache performance. In order to balance-load across threads, the MTTM augments a basic map-reduce strategy to draw jobs from a global queue. In a multicore processor, memory may be "homed" to the cache of a specific processor and must be accessed from that processor. The MTTB architecture wraps access to data with thread management to move threads to the home processor for that data so that the computation follows the data in an attempt to avoid L2 cache misses. Cache homing is also handled by a memory manager that translates identifiers to processor IDs where the data will be homed (according to rules defined by the user). The user can also specify the number of threads and processors separately, which is important for tuning performance for different patterns of computation and memory access. MTTM efficiently processes tasks in parallel on a multiprocessor computer. It also provides an interface to make it easier to adapt existing software to a multiprocessor environment.
Cache-Oblivious parallel SIMD Viterbi decoding for sequence search in HMMER.
Ferreira, Miguel; Roma, Nuno; Russo, Luis M S
2014-05-30
HMMER is a commonly used bioinformatics tool based on Hidden Markov Models (HMMs) to analyze and process biological sequences. One of its main homology engines is based on the Viterbi decoding algorithm, which was already highly parallelized and optimized using Farrar's striped processing pattern with Intel SSE2 instruction set extension. A new SIMD vectorization of the Viterbi decoding algorithm is proposed, based on an SSE2 inter-task parallelization approach similar to the DNA alignment algorithm proposed by Rognes. Besides this alternative vectorization scheme, the proposed implementation also introduces a new partitioning of the Markov model that allows a significantly more efficient exploitation of the cache locality. Such optimization, together with an improved loading of the emission scores, allows the achievement of a constant processing throughput, regardless of the innermost-cache size and of the dimension of the considered model. The proposed optimized vectorization of the Viterbi decoding algorithm was extensively evaluated and compared with the HMMER3 decoder to process DNA and protein datasets, proving to be a rather competitive alternative implementation. Being always faster than the already highly optimized ViterbiFilter implementation of HMMER3, the proposed Cache-Oblivious Parallel SIMD Viterbi (COPS) implementation provides a constant throughput and offers a processing speedup as high as two times faster, depending on the model's size.
Graded Mirror Self-Recognition by Clark's Nutcrackers.
Clary, Dawson; Kelly, Debbie M
2016-11-04
The traditional 'mark test' has shown some large-brained species are capable of mirror self-recognition. During this test a mark is inconspicuously placed on an animal's body where it can only be seen with the aid of a mirror. If the animal increases the number of actions directed to the mark region when presented with a mirror, the animal is presumed to have recognized the mirror image as its reflection. However, the pass/fail nature of the mark test presupposes self-recognition exists in entirety or not at all. We developed a novel mirror-recognition task, to supplement the mark test, which revealed gradation in the self-recognition of Clark's nutcrackers, a large-brained corvid. To do so, nutcrackers cached food alone, observed by another nutcracker, or with a regular or blurry mirror. The nutcrackers suppressed caching with a regular mirror, a behavioural response to prevent cache theft by conspecifics, but did not suppress caching with a blurry mirror. Likewise, during the mark test, most nutcrackers made more self-directed actions to the mark with a blurry mirror than a regular mirror. Both results suggest self-recognition was more readily achieved with the blurry mirror and that self-recognition may be more broadly present among animals than currently thought.
Fox Squirrels Match Food Assessment and Cache Effort to Value and Scarcity
Delgado, Mikel M.; Nicholas, Molly; Petrie, Daniel J.; Jacobs, Lucia F.
2014-01-01
Scatter hoarders must allocate time to assess items for caching, and to carry and bury each cache. Such decisions should be driven by economic variables, such as the value of the individual food items, the scarcity of these items, competition for food items and risk of pilferage by conspecifics. The fox squirrel, an obligate scatter-hoarder, assesses cacheable food items using two overt movements, head flicks and paw manipulations. These behaviors allow an examination of squirrel decision processes when storing food for winter survival. We measured wild squirrels' time allocations and frequencies of assessment and investment behaviors during periods of food scarcity (summer) and abundance (fall), giving the squirrels a series of 15 items (alternating five hazelnuts and five peanuts). Assessment and investment per cache increased when resource value was higher (hazelnuts) or resources were scarcer (summer), but decreased as scarcity declined (end of sessions). This is the first study to show that assessment behaviors change in response to factors that indicate daily and seasonal resource abundance, and that these factors may interact in complex ways to affect food storing decisions. Food-storing tree squirrels may be a useful and important model species to understand the complex economic decisions made under natural conditions. PMID:24671221
Peters, M E; Rosenberg, P B; Steinberg, M; Norton, M C; Welsh-Bohmer, K A; Hayden, K M; Breitner, J; Tschanz, J T; Lyketsos, C G
2013-11-01
To examine the association of neuropsychiatric symptom (NPS) severity with risk of transition to all-cause dementia, Alzheimer disease (AD), and vascular dementia (VaD). Survival analysis of time to dementia, AD, or VaD onset. Population-based study. 230 participants diagnosed with cognitive impairment, no dementia (CIND) from the Cache County Study of Memory Health and Aging were followed for a mean of 3.3 years. The Neuropsychiatric Inventory (NPI) was used to quantify the presence, frequency, and severity of NPS. Chi-squared statistics, t-tests, and Cox proportional hazard ratios were used to assess associations. The conversion rate from CIND to all-cause dementia was 12% per year, with risk factors including an APOE ε4 allele, lower Mini-Mental State Examination, lower 3MS, and higher CDR sum-of-boxes. The presence of at least one NPS was a risk factor for all-cause dementia, as was the presence of NPS with mild severity. Nighttime behaviors were a risk factor for all-cause dementia and of AD, whereas hallucinations were a risk factor for VaD. These data confirm that NPS are risk factors for conversion from CIND to dementia. Of special interest is that even NPS of mild severity are a risk for all-cause dementia or AD. Copyright © 2013 American Association for Geriatric Psychiatry. Published by Elsevier Inc. All rights reserved.
Peters, ME; Rosenberg, PB; Steinberg, M; Norton, MC; Welsh-Bohmer, KA; Hayden, KM; Breitner, J; Tschanz, JT; CG, Lyketsos
2012-01-01
Objectives To examine the association of neuropsychiatric symptom (NPS) severity with risk of transition to all-cause dementia, Alzheimer's disease (AD), and vascular dementia (VaD). Design Survival analysis of time to dementia, AD, or VaD onset. Setting Population-based study. Participants 230 participants diagnosed with cognitive impairment, no dementia (CIND) from the Cache County Study of Memory Health and Aging were followed for a mean of 3.3 years. Measurements The Neuropsychiatric Inventory (NPI) was used to quantify the presence, frequency, and severity of NPS. Chi-square statistics, t-tests, and Cox proportional hazard ratios were used to assess associations. Results The conversion rate from CIND to all-cause dementia was 12% per year, with risk factors including an APOE ε4 allele, lower MMSE, lower 3MS, and higher CDR sum-of-boxes. The presence of at least one NPS was a risk factor for all-cause dementia, as was the presence of NPS with mild severity. Nighttime behaviors were a risk factor for all-cause dementia and of AD, while hallucinations were a risk factor for VaD. Conclusions These data confirm that NPS are risk factors for conversion from CIND to dementia. Of special interest is that even NPS of mild severity are a risk for all-cause dementia or AD. PMID:23567370
Effects of Methamphetamine on Vigilance and Tracking during Extended Wakefulness.
1993-09-01
the log likelihood ratio (log(p); Green & Swets, 1966; Macmillan & Creelman , 1990), was also derived from hit and false-alarm probabilities...vigilance task. Canadian Journal of Psychology, 19, 104-110. Macmillan, N.E., & Creelman , C.D. (1990). Response bias: Characteristics of detection
Violent aggression predicted by multiple pre-adult environmental hits.
Mitjans, Marina; Seidel, Jan; Begemann, Martin; Bockhop, Fabian; Moya-Higueras, Jorge; Bansal, Vikas; Wesolowski, Janina; Seelbach, Anna; Ibáñez, Manuel Ignacio; Kovacevic, Fatka; Duvar, Oguzhan; Fañanás, Lourdes; Wolf, Hannah-Ulrike; Ortet, Generós; Zwanzger, Peter; Klein, Verena; Lange, Ina; Tänzer, Andreas; Dudeck, Manuela; Penke, Lars; van Elst, Ludger Tebartz; Bittner, Robert A; Schmidmeier, Richard; Freese, Roland; Müller-Isberner, Rüdiger; Wiltfang, Jens; Bliesener, Thomas; Bonn, Stefan; Poustka, Luise; Müller, Jürgen L; Arias, Bárbara; Ehrenreich, Hannelore
2018-05-24
Early exposure to negative environmental impact shapes individual behavior and potentially contributes to any mental disease. We reported previously that accumulated environmental risk markedly decreases age at schizophrenia onset. Follow-up of matched extreme group individuals (≤1 vs. ≥3 risks) unexpectedly revealed that high-risk subjects had >5 times greater probability of forensic hospitalization. In line with longstanding sociological theories, we hypothesized that risk accumulation before adulthood induces violent aggression and criminal conduct, independent of mental illness. We determined in 6 independent cohorts (4 schizophrenia and 2 general population samples) pre-adult risk exposure, comprising urbanicity, migration, physical and sexual abuse as primary, and cannabis or alcohol as secondary hits. All single hits by themselves were marginally associated with higher violent aggression. Most strikingly, however, their accumulation strongly predicted violent aggression (odds ratio 10.5). An epigenome-wide association scan to detect differential methylation of blood-derived DNA of selected extreme group individuals yielded overall negative results. Conversely, determination in peripheral blood mononuclear cells of histone-deacetylase1 mRNA as 'umbrella mediator' of epigenetic processes revealed an increase in the high-risk group, suggesting lasting epigenetic alterations. Together, we provide sound evidence of a disease-independent unfortunate relationship between well-defined pre-adult environmental hits and violent aggression, calling for more efficient prevention.
Simulation of flow and habitat conditions under ice, Cache la Poudre River - January 2006
Waddle, Terry
2007-01-01
The objectives of this study are (1) to describe the extent and thickness of ice cover, (2) simulate depth and velocity under ice at the study site for observed and reduced flows, and (3) to quantify fish habitat in this portion of the mainstem Cache la Poudre River for the current winter release schedule as well as for similar conditions without the 0.283 m3/s winter release.
1993-08-01
on the Lempel - Ziv [44] algo- rithm. Zip is compressing a single 8,017 byte file. " RTLSim An register transfer language simulator for the Message...package. gordoni@cs.adelaide.edu.au, Wynn Vale, 5127, Australia, 1.0 edition, October 1991. [44] Ziv J. and Lempel A. "A universal algorithm for...fixed hardware algorithm . Some data caches allow the program to explicitly allocate cache lines [68]. This allocation is only useful in writing new data
DARPA Status Report - November 1988
1988-11-01
style used in the applic4#ons reference to that block was by processor j. where j It. We was influenced by it. MACH is a multiprocessor operating S call...it can be order they occurred. However. the exact time at which the treated specially in memory management , and so most of the reference wa, made is...on cache consistency performance, sophisti- peak can be explained as clinging references that occur when cated cache management schemes that take
Side Channel Attacks on STTRAM and Low Overhead Countermeasures
2017-03-20
introduce security vulnerabilities and expose the cache memory to side channel attacks. In this paper, we propose a side channel attack (SCA) model...where the adversary can monitor the supply current of the memory array to partially identify the sensi- tive cache data that is being read or written. We...propose solutions such as short retention STTRAM, obfuscation of SCA using 1-bit parity, multi-bit random write, and, neutral- izing the SCA using
1993-03-01
CLUSTER A CLUSTER B .UDP D "Orequeqes ProxyDistribute 0 Figure 4-4: HOSTALL Implementation HOST_ALL is implemented as follows. The kernel looks up the...it includes the HOSTALL request as an argument. The generic CronusHost object is managed by the Cronus Kernel. A kernel that receives a ProxyDistnbute...request uses its cached service information to send the HOSTALL request to each host in its cluster via UDP. If the kernel has no cached information
NASA Technical Reports Server (NTRS)
Wheeler, D. J.; Ridd, M. K.; Merola, J. A.
1984-01-01
A basic geographic information system (GIS) for the North Cache Soil Conservation District (SCD) was sought for selected resource problems. Since the resource management issues in the North Cache SCD are very complex, it is not feasible in the initial phase to generate all the physical, socioeconomic, and political baseline data needed for resolving all management issues. A selection of critical varables becomes essential. Thus, there are foud specific objectives: (1) assess resource management needs and determine which resource factors ae most fundamental for building a beginning data base; (2) evaluate the variety of data gathering and analysis techniques for the resource factors selected; (3) incorporate the resulting data into a useful and efficient digital data base; and (4) demonstrate the application of the data base to selected real world resoource management issues.
Sparse Partial Equilibrium Tables in Chemically Resolved Reactive Flow
NASA Astrophysics Data System (ADS)
Vitello, Peter; Fried, Laurence E.; Pudliner, Brian; McAbee, Tom
2004-07-01
The detonation of an energetic material is the result of a complex interaction between kinetic chemical reactions and hydrodynamics. Unfortunately, little is known concerning the detailed chemical kinetics of detonations in energetic materials. CHEETAH uses rate laws to treat species with the slowest chemical reactions, while assuming other chemical species are in equilibrium. CHEETAH supports a wide range of elements and condensed detonation products and can also be applied to gas detonations. A sparse hash table of equation of state values is used in CHEETAH to enhance the efficiency of kinetic reaction calculations. For large-scale parallel hydrodynamic calculations, CHEETAH uses parallel communication to updates to the cache. We present here details of the sparse caching model used in the CHEETAH coupled to an ALE hydrocode. To demonstrate the efficiency of modeling using a sparse cache model we consider detonations in energetic materials.
Cache Locality Optimization for Recursive Programs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lifflander, Jonathan; Krishnamoorthy, Sriram
We present an approach to optimize the cache locality for recursive programs by dynamically splicing--recursively interleaving--the execution of distinct function invocations. By utilizing data effect annotations, we identify concurrency and data reuse opportunities across function invocations and interleave them to reduce reuse distance. We present algorithms that efficiently track effects in recursive programs, detect interference and dependencies, and interleave execution of function invocations using user-level (non-kernel) lightweight threads. To enable multi-core execution, a program is parallelized using a nested fork/join programming model. Our cache optimization strategy is designed to work in the context of a random work stealing scheduler. Wemore » present an implementation using the MIT Cilk framework that demonstrates significant improvements in sequential and parallel performance, competitive with a state-of-the-art compile-time optimizer for loop programs and a domain- specific optimizer for stencil programs.« less
Josephson 4 K-bit cache memory design for a prototype signal processor. I - General overview
NASA Astrophysics Data System (ADS)
Henkels, W. H.; Geppert, L. M.; Kadlec, J.; Epperlein, P. W.; Beha, H.
1985-09-01
In the early stages of thg Josephson computer project conducted at an American computer company, it was recognized that a very fast cache memory was needed to complement Josephson logic. A subnanosecond access time memory was implemented experimentally on the basis of a 2.5-micron Pb-alloy technology. It was then decided to switch over to a Nb-base-electrode technology with the objective to alleviate problems with the long-term reliability and aging of Pb-based junctions. The present paper provides a general overview of the status of a 4 x 1 K-bit Josephson cache design employing a 2.5-micron Nb-edge-junction technology. Attention is given to the fabrication process and its implications, aspects of circuit design methodology, an overview of system environment and chip components, design changes and status, and various difficulties and uncertainties.
NASA Astrophysics Data System (ADS)
Graham, Jim; Jarnevich, Catherine S.; Simpson, Annie; Newman, Gregory J.; Stohlgren, Thomas J.
2011-06-01
Invasive species are a universal global problem, but the information to identify them, manage them, and prevent invasions is stored around the globe in a variety of formats. The Global Invasive Species Information Network is a consortium of organizations working toward providing seamless access to these disparate databases via the Internet. A distributed network of databases can be created using the Internet and a standard web service protocol. There are two options to provide this integration. First, federated searches are being proposed to allow users to search "deep" web documents such as databases for invasive species. A second method is to create a cache of data from the databases for searching. We compare these two methods, and show that federated searches will not provide the performance and flexibility required from users and a central cache of the datum are required to improve performance.
Graham, Jim; Jarnevich, Catherine S.; Simpson, Annie; Newman, Gregory J.; Stohlgren, Thomas J.
2011-01-01
Invasive species are a universal global problem, but the information to identify them, manage them, and prevent invasions is stored around the globe in a variety of formats. The Global Invasive Species Information Network is a consortium of organizations working toward providing seamless access to these disparate databases via the Internet. A distributed network of databases can be created using the Internet and a standard web service protocol. There are two options to provide this integration. First, federated searches are being proposed to allow users to search “deep” web documents such as databases for invasive species. A second method is to create a cache of data from the databases for searching. We compare these two methods, and show that federated searches will not provide the performance and flexibility required from users and a central cache of the datum are required to improve performance.
Welsh-Bohmer, Kathleen A; Breitner, John C S; Hayden, Kathleen M; Lyketsos, Constantine; Zandi, Peter P; Tschanz, Joann T; Norton, Maria C; Munger, Ron
2006-07-01
The Cache County Study of Memory, Health, and Aging, more commonly referred to as the "Cache County Memory Study (CCMS)" is a longitudinal investigation of aging and Alzheimer's disease (AD) based in an exceptionally long-lived population residing in northern Utah. The study begun in 1994 has followed an initial cohort of 5,092 older individuals (many over age 84) and has examined the development of cognitive impairment and dementia in relation to genetic and environmental antecedents. This article summarizes the major contributions of the CCMS towards the understanding of mild cognitive disorders and AD across the lifespan, underscoring the role of common health exposures in modifying dementia risk and trajectories of cognitive change. The study now in its fourth wave of ascertainment illustrates the role of population-based approaches in informing testable models of cognitive aging and Alzheimer's disease.
Urban Geocaching: what Happened in Lisbon during the Last Decade?
NASA Astrophysics Data System (ADS)
Nogueira Mendes, R.; Rodrigues, T.; Rodrigues, A. M.
2013-05-01
Created in 2000 in the United States of America, Geocaching has become a major phenomenon all around the world, counting actually with millions of Geocaches (or caches) that work as a recreational motivation for millions of users, called Geocachers. During the last 30 days over 5,000,000 new logs have been submitted worldwide, disseminating individual experiences, motivations, emotions and photos through the official Geocaching website (www.geocaching.com), and several official or informal national web forums. The activity itself can be compared with modern treasure hunting that uses handheld GPS, Smartphones or Tablets, WEB 2.0, wiki features and technologies to keep Geocachers engaged with their activity, in a strong social-network. All these characteristics make Geocaching an activity with a strong geographic component that deals closely with the surrounding environment where each cache has been hidden. From previous work, significance correlation has been found regarding hides and natural/rural environments, but metropolitan and urban areas like Lisbon municipality (that holds 3.23% of the total 27534 Portuguese caches), still registers the higher density of Geocaches, and logs numbers. Lacking "natural/rural" environment, Geocaching in cities tend to happen in symbolic areas, like public parks and places, sightseeing spots and historical neighborhoods. The present study looks to Geocaching within the city of Lisbon, in order to understand how it works, and if this activity reflects the city itself, promoting its image and cultural heritage. From a freely available dataset that includes all Geocaches that have been placed in Lisbon since February 2001, spatial analysis has been conducted, showing the informal preferences of this activity. Results show a non-random distribution of caches within the study area, similar to the land use distribution. Preferable locations tend to be in iconic places of the city, usually close to the Tagus River, that concentrates 25% of the total caches. Since most of these places are known to be touristic destinations, the TOP15 logged Caches were also analyzed regarding their description and logs in order to understand if Geocaching reflects tourism and if it works as a tourist promotion tool within urban environments. Final results also reflect the Geocaching performance and major trends within urban environments providing new insights regarding this activity impacts and implications.
Caching strategies for improving performance of web-based Geographic applications
NASA Astrophysics Data System (ADS)
Liu, M.; Brodzik, M.; Collins, J. A.; Lewis, S.; Oldenburg, J.
2012-12-01
The NASA Operation IceBridge mission collects airborne remote sensing measurements to bridge the gap between NASA's Ice, Cloud and Land Elevation Satellite (ICESat) mission and the upcoming ICESat-2 mission. The IceBridge Data Portal from the National Snow and Ice Data Center provides an intuitive web interface for accessing IceBridge mission observations and measurements. Scientists and users usually do not have knowledge about the individual campaigns but are interested in data collected in a specific place. We have developed a high-performance map interface to allow users to quickly zoom to an area of interest and see any Operation IceBridge overflights. The map interface consists of two layers: the user can pan and zoom on the base map layer; the flight line layer that overlays the base layer provides all the campaign missions that intersect with the current map view. The user can click on the flight campaigns and download the data as needed. The OpenGIS® Web Map Service Interface Standard (WMS) provides a simple HTTP interface for requesting geo-registered map images from one or more distributed geospatial databases. Web Feature Service (WFS) provides an interface allowing requests for geographical features across the web using platform-independent calls. OpenLayers provides vector support (points, polylines and polygons) to build a WMS/WFS client for displaying both layers on the screen. Map Server, an open source development environment for building spatially enabled internet applications, is serving the WMS and WFS spatial data to OpenLayers. Early releases of the portal displayed unacceptably poor load time performance for flight lines and the base map tiles. This issue was caused by long response times from the map server in generating all map tiles and flight line vectors. We resolved the issue by implementing various caching strategies on top of the WMS and WFS services, including the use of Squid (www.squid-cache.org) to cache frequently-used content. Our presentation includes the architectural design of the application, and how we use OpenLayers, WMS and WFS with Squid to build a responsive web application capable of efficiently displaying geospatial data to allow the user to quickly interact with the displayed information. We describe the design, implementation and performance improvement of our caching strategies, and the tools and techniques developed to assist our data caching strategies.
EarthCache as a Tool to Promote Earth-Science in Public School Classrooms
NASA Astrophysics Data System (ADS)
Gochis, E. E.; Rose, W. I.; Klawiter, M.; Vye, E. C.; Engelmann, C. A.
2011-12-01
Geoscientists often find it difficult to bridge the gap in communication between university research and what is learned in the public schools. Today's schools operate in a high stakes environment that only allow instruction based on State and National Earth Science curriculum standards. These standards are often unknown by academics or are written in a style that obfuscates the transfer of emerging scientific research to students in the classroom. Earth Science teachers are in an ideal position to make this link because they have a background in science as well as a solid understanding of the required curriculum standards for their grade and the pedagogical expertise to pass on new information to their students. As part of the Michigan Teacher Excellence Program (MiTEP), teachers from Grand Rapids, Kalamazoo, and Jackson school districts participate in 2 week field courses with Michigan Tech University to learn from earth science experts about how the earth works. This course connects Earth Science Literacy Principles' Big Ideas and common student misconceptions with standards-based education. During the 2011 field course, we developed and began to implement a three-phase EarthCache model that will provide a geospatial interactive medium for teachers to translate the material they learn in the field to the students in their standards based classrooms. MiTEP participants use GPS and Google Earth to navigate to Michigan sites of geo-significance. At each location academic experts aide participants in making scientific observations about the locations' geologic features, and "reading the rocks" methodology to interpret the area's geologic history. The participants are then expected to develop their own EarthCache site to be used as pedagogical tool bridging the gap between standards-based classroom learning, contemporary research and unique outdoor field experiences. The final phase supports teachers in integrating inquiry based, higher-level learning student activities to EarthCache sites near their own urban communities, or in regional areas such as nature preserves and National Parks. By working together, MiTEP participants are developing a network of regional EarthCache sites and shared lesson plans which explore places that are meaningful to students while simultaneously connecting them to geologic concepts they are learning in school. We believe that the MiTEP EarthCaching model will help participants emerge as leaders of inquiry style, and virtual place-based educators within their districts.
1979-02-01
classified as Porno , Lake Miwok, and Patwin. Recent surveys within the Clear Lake-Cache Creek Basin have located 28 archeological sites, some of which...additional 8,400 acre-feet annually to the Lakeport area. Porno Reservoir on Kelsey Creek, being studied by Lake County, also would supplement M&l water...project on Scotts Creek could provide 9,100 acre- feet annually of irrigation water. Also, as previously discussed, Porno Reservoir would furnish
Accessing Data Federations with CVMFS
Weitzel, Derek; Bockelman, Brian; Dykstra, Dave; ...
2017-11-23
Data federations have become an increasingly common tool for large collaborations such as CMS and Atlas to efficiently distribute large data files. Unfortunately, these typically are implemented with weak namespace semantics and a non-POSIX API. On the other hand, CVMFS has provided a POSIX-compliant read-only interface for use cases with a small working set size (such as software distribution). The metadata required for the CVMFS POSIX interface is distributed through a caching hierarchy, allowing it to scale to the level of about a hundred thousand hosts. In this paper, we will describe our contributions to CVMFS that merges the datamore » scalability of XRootD-based data federations (such as AAA) with metadata scalability and POSIX interface of CVMFS. We modified CVMFS so it can serve unmodified files without copying them to the repository server. CVMFS 2.2.0 is also able to redirect requests for data files to servers outside of the CVMFS content distribution network. Finally, we added the ability to manage authorization and authentication using security credentials such as X509 proxy certificates. We combined these modifications with the OSGs StashCache regional XRootD caching infrastructure to create a cached data distribution network. Here, we will show performance metrics accessing the data federation through CVMFS compared to direct data federation access. Additionally, we will discuss the improved user experience of providing access to a data federation through a POSIX filesystem.« less
Accessing Data Federations with CVMFS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Weitzel, Derek; Bockelman, Brian; Dykstra, Dave
Data federations have become an increasingly common tool for large collaborations such as CMS and Atlas to efficiently distribute large data files. Unfortunately, these typically are implemented with weak namespace semantics and a non-POSIX API. On the other hand, CVMFS has provided a POSIX-compliant read-only interface for use cases with a small working set size (such as software distribution). The metadata required for the CVMFS POSIX interface is distributed through a caching hierarchy, allowing it to scale to the level of about a hundred thousand hosts. In this paper, we will describe our contributions to CVMFS that merges the datamore » scalability of XRootD-based data federations (such as AAA) with metadata scalability and POSIX interface of CVMFS. We modified CVMFS so it can serve unmodified files without copying them to the repository server. CVMFS 2.2.0 is also able to redirect requests for data files to servers outside of the CVMFS content distribution network. Finally, we added the ability to manage authorization and authentication using security credentials such as X509 proxy certificates. We combined these modifications with the OSGs StashCache regional XRootD caching infrastructure to create a cached data distribution network. Here, we will show performance metrics accessing the data federation through CVMFS compared to direct data federation access. Additionally, we will discuss the improved user experience of providing access to a data federation through a POSIX filesystem.« less
Accessing Data Federations with CVMFS
NASA Astrophysics Data System (ADS)
Weitzel, Derek; Bockelman, Brian; Dykstra, Dave; Blomer, Jakob; Meusel, Ren
2017-10-01
Data federations have become an increasingly common tool for large collaborations such as CMS and Atlas to efficiently distribute large data files. Unfortunately, these typically are implemented with weak namespace semantics and a non-POSIX API. On the other hand, CVMFS has provided a POSIX-compliant read-only interface for use cases with a small working set size (such as software distribution). The metadata required for the CVMFS POSIX interface is distributed through a caching hierarchy, allowing it to scale to the level of about a hundred thousand hosts. In this paper, we will describe our contributions to CVMFS that merges the data scalability of XRootD-based data federations (such as AAA) with metadata scalability and POSIX interface of CVMFS. We modified CVMFS so it can serve unmodified files without copying them to the repository server. CVMFS 2.2.0 is also able to redirect requests for data files to servers outside of the CVMFS content distribution network. Finally, we added the ability to manage authorization and authentication using security credentials such as X509 proxy certificates. We combined these modifications with the OSGs StashCache regional XRootD caching infrastructure to create a cached data distribution network. We will show performance metrics accessing the data federation through CVMFS compared to direct data federation access. Additionally, we will discuss the improved user experience of providing access to a data federation through a POSIX filesystem.
Cache-Oblivious parallel SIMD Viterbi decoding for sequence search in HMMER
2014-01-01
Background HMMER is a commonly used bioinformatics tool based on Hidden Markov Models (HMMs) to analyze and process biological sequences. One of its main homology engines is based on the Viterbi decoding algorithm, which was already highly parallelized and optimized using Farrar’s striped processing pattern with Intel SSE2 instruction set extension. Results A new SIMD vectorization of the Viterbi decoding algorithm is proposed, based on an SSE2 inter-task parallelization approach similar to the DNA alignment algorithm proposed by Rognes. Besides this alternative vectorization scheme, the proposed implementation also introduces a new partitioning of the Markov model that allows a significantly more efficient exploitation of the cache locality. Such optimization, together with an improved loading of the emission scores, allows the achievement of a constant processing throughput, regardless of the innermost-cache size and of the dimension of the considered model. Conclusions The proposed optimized vectorization of the Viterbi decoding algorithm was extensively evaluated and compared with the HMMER3 decoder to process DNA and protein datasets, proving to be a rather competitive alternative implementation. Being always faster than the already highly optimized ViterbiFilter implementation of HMMER3, the proposed Cache-Oblivious Parallel SIMD Viterbi (COPS) implementation provides a constant throughput and offers a processing speedup as high as two times faster, depending on the model’s size. PMID:24884826
Simulation of a Real-Time Brain Computer Interface for Detecting a Self-Paced Hitting Task.
Hammad, Sofyan H; Kamavuako, Ernest N; Farina, Dario; Jensen, Winnie
2016-12-01
An invasive brain-computer interface (BCI) is a promising neurorehabilitation device for severely disabled patients. Although some systems have been shown to work well in restricted laboratory settings, their utility must be tested in less controlled, real-time environments. Our objective was to investigate whether a specific motor task could be reliably detected from multiunit intracortical signals from freely moving animals in a simulated, real-time setting. Intracortical signals were first obtained from electrodes placed in the primary motor cortex of four rats that were trained to hit a retractable paddle (defined as a "Hit"). In the simulated real-time setting, the signal-to-noise-ratio was first increased by wavelet denoising. Action potentials were detected, and features were extracted (spike count, mean absolute values, entropy, and combination of these features) within pre-defined time windows (200 ms, 300 ms, and 400 ms) to classify the occurrence of a "Hit." We found higher detection accuracy of a "Hit" (73.1%, 73.4%, and 67.9% for the three window sizes, respectively) when the decision was made based on a combination of features rather than on a single feature. However, the duration of the window length was not statistically significant (p = 0.5). Our results showed the feasibility of detecting a motor task in real time in a less restricted environment compared to environments commonly applied within invasive BCI research, and they showed the feasibility of using information extracted from multiunit recordings, thereby avoiding the time-consuming and complex task of extracting and sorting single units. © 2016 International Neuromodulation Society.
Adaptive Bloom Filter: A Space-Efficient Counting Algorithm for Unpredictable Network Traffic
NASA Astrophysics Data System (ADS)
Matsumoto, Yoshihide; Hazeyama, Hiroaki; Kadobayashi, Youki
The Bloom Filter (BF), a space-and-time-efficient hashcoding method, is used as one of the fundamental modules in several network processing algorithms and applications such as route lookups, cache hits, packet classification, per-flow state management or network monitoring. BF is a simple space-efficient randomized data structure used to represent a data set in order to support membership queries. However, BF generates false positives, and cannot count the number of distinct elements. A counting Bloom Filter (CBF) can count the number of distinct elements, but CBF needs more space than BF. We propose an alternative data structure of CBF, and we called this structure an Adaptive Bloom Filter (ABF). Although ABF uses the same-sized bit-vector used in BF, the number of hash functions employed by ABF is dynamically changed to record the number of appearances of a each key element. Considering the hash collisions, the multiplicity of a each key element on ABF can be estimated from the number of hash functions used to decode the membership of the each key element. Although ABF can realize the same functionality as CBF, ABF requires the same memory size as BF. We describe the construction of ABF and IABF (Improved ABF), and provide a mathematical analysis and simulation using Zipf's distribution. Finally, we show that ABF can be used for an unpredictable data set such as real network traffic.
A century in the making… almost here?
Adams, Katherine T
2007-12-01
Cancer immunotherapeutics, including vaccines, may hit the market by the end of this decade. They promise longer survival and perhaps indefinite remission at a more favorable cost-benefit ratio than that of oncologics now in use. They also may fuel dramatic changes in how healthcare is delivered and financed.
O'Malley, Ann S; Reschovsky, James D; Saiontz-Martinez, Cynthia
2015-01-01
Practice tools such as health information technology (HIT) have the potential to support care processes, such as communication between health care providers, and influence care for "ambulatory care-sensitive conditions" (ACSCs). ACSCs are conditions for which good outpatient care can potentially prevent the need for hospitalization. To date, associations between such primary care practice capabilities and hospitalizations for ambulatory care-sensitive conditions have been primarily limited to smaller, local studies or unique delivery systems rather than nationally representative studies of primary care physicians in the United States. We analyzed a nationally representative sample of 1,819 primary care physicians who responded to the Center for Studying Health System Change's Physician Survey. We linked 3 years of Medicare claims (2007 to 2009) with these primary care physician survey respondents. This linkage resulted in the identification of 123,760 beneficiaries with one or more of 4 ambulatory care-sensitive chronic conditions (diabetes, chronic obstructive pulmonary disease, asthma, and congestive heart failure) for whom these physicians served as the usual provider. Key independent variables of interest were physicians' practice capabilities, including communication with specialists, use of care managers, participation in quality and performance measurement, use of patient registries, and HIT use. The dependent variable was a summary measure of ambulatory care-sensitive hospitalizations for one or more of these 4 conditions. Higher provider-reported levels of communication between primary care and specialist physicians were associated with lower rates of potentially avoidable hospitalizations. While there was no significant main effect between HIT use and ACSC hospitalizations, the associations between interspecialty communication and ACSC hospitalizations were magnified in the presence of higher HIT use. For example, patients in practices with both the highest level of interspecialty communication and the highest level of HIT use had lower odds of ambulatory care-sensitive hospitalizations than did those in practices with lower interspecialty communication and high HIT use (adjusted odds ratio, 0.70; 95% confidence limits, 0.59, 0.82). Greater primary care and specialist communication is associated with reduced hospitalizations for ambulatory care-sensitive conditions. This effect was magnified in the presence of higher provider-reported HIT use, suggesting that coordination of care with support from HIT is important in the treatment of ambulatory care-sensitive conditions. © Copyright 2015 by the American Board of Family Medicine.
A performance model for GPUs with caches
Dao, Thanh Tuan; Kim, Jungwon; Seo, Sangmin; ...
2014-06-24
To exploit the abundant computational power of the world's fastest supercomputers, an even workload distribution to the typically heterogeneous compute devices is necessary. While relatively accurate performance models exist for conventional CPUs, accurate performance estimation models for modern GPUs do not exist. This paper presents two accurate models for modern GPUs: a sampling-based linear model, and a model based on machine-learning (ML) techniques which improves the accuracy of the linear model and is applicable to modern GPUs with and without caches. We first construct the sampling-based linear model to predict the runtime of an arbitrary OpenCL kernel. Based on anmore » analysis of NVIDIA GPUs' scheduling policies we determine the earliest sampling points that allow an accurate estimation. The linear model cannot capture well the significant effects that memory coalescing or caching as implemented in modern GPUs have on performance. We therefore propose a model based on ML techniques that takes several compiler-generated statistics about the kernel as well as the GPU's hardware performance counters as additional inputs to obtain a more accurate runtime performance estimation for modern GPUs. We demonstrate the effectiveness and broad applicability of the model by applying it to three different NVIDIA GPU architectures and one AMD GPU architecture. On an extensive set of OpenCL benchmarks, on average, the proposed model estimates the runtime performance with less than 7 percent error for a second-generation GTX 280 with no on-chip caches and less than 5 percent for the Fermi-based GTX 580 with hardware caches. On the Kepler-based GTX 680, the linear model has an error of less than 10 percent. On an AMD GPU architecture, Radeon HD 6970, the model estimates with 8 percent of error rates. As a result, the proposed technique outperforms existing models by a factor of 5 to 6 in terms of accuracy.« less
The Acceleration of Structural Microarchitectural Simulation via Scheduling
2006-11-01
193 viii List of Tables 1.1 Size of Intel R ©Processors...Table 1.1 shows the total and estimated non-cache transistor counts in succeeding generations of Intel R ©microprocessors. (Cache array transistors are...Intel486TM 1989 1,200,000 800,000 Intel R ©Pentium R © 1993 3,100,000 2,300,000 Intel R ©Pentium R ©II 1997 7,500,000 5,500,000 Intel R ©Pentium R ©III 1999
A High-Precision Counter Using the DSP Technique
2004-09-01
DSP is not good enough to process all the 1-second samples. The cache memory is also not sufficient to store all the sampling data. So we cut the...sampling number in a cycle is not good enough to achieve an accuracy less than 2×10-11. For this reason, a correlation operation is performed for... not good enough to process all the 1-second samples. The cache memory is also not sufficient to store all the sampling data. We will solve this
Compiler-directed cache management in multiprocessors
NASA Technical Reports Server (NTRS)
Cheong, Hoichi; Veidenbaum, Alexander V.
1990-01-01
The necessity of finding alternatives to hardware-based cache coherence strategies for large-scale multiprocessor systems is discussed. Three different software-based strategies sharing the same goals and general approach are presented. They consist of a simple invalidation approach, a fast selective invalidation scheme, and a version control scheme. The strategies are suitable for shared-memory multiprocessor systems with interconnection networks and a large number of processors. Results of trace-driven simulations conducted on numerical benchmark routines to compare the performance of the three schemes are presented.
Developing Tools and Technologies to Meet MSR Planetary Protection Requirements
NASA Technical Reports Server (NTRS)
Lin, Ying
2013-01-01
This paper describes the tools and technologies that need to be developed for a Caching Rover mission in order to meet the overall Planetary Protection requirements for future Mars Sample Return (MSR) campaign. This is the result of an eight-month study sponsored by the Mars Exploration Program Office. The goal of this study is to provide a future MSR project with a focused technology development plan for achieving the necessary planetary protection and sample integrity capabilities for a Mars Caching Rover mission.
Neural strategies for selective attention distinguish fast-action video game players.
Krishnan, Lavanya; Kang, Albert; Sperling, George; Srinivasan, Ramesh
2013-01-01
We investigated the psychophysical and neurophysiological differences between fast-action video game players (specifically first person shooter players, FPS) and non-action players (role-playing game players, RPG) in a visual search task. We measured both successful detections (hit rates) and steady-state visually evoked EEG potentials (SSVEPs). Search difficulty was varied along two dimensions: number of adjacent attended and ignored regions (1, 2 and 4), and presentation rate of novel search arrays (3, 8.6 and 20 Hz). Hit rates decreased with increasing presentation rates and number of regions, with the FPS players performing on average better than the RPG players. The largest differences in hit rate, between groups, occurred when four regions were simultaneously attended. We computed signal-to-noise ratio (SNR) of SSVEPs and used partial least squares regression to model hit rates, SNRs and their relationship at 3 Hz and 8.6 Hz. The following are the most significant results: RPG players' parietal responses to the attended 8.6 Hz flicker were predictive of hit rate and were positively correlated with it, indicating attentional signal enhancement. FPS players' parietal responses to the ignored 3 Hz flicker were predictive of hit rate and were positively correlated with it, indicating distractor suppression. Consistent with these parietal responses, RPG players' frontal responses to the attended 8.6 Hz flicker, increased as task difficulty increased with number of regions; FPS players' frontal responses to the ignored 3 Hz flicker increased with number of regions. Thus the FPS players appear to employ an active suppression mechanism to deploy selective attention simultaneously to multiple interleaved regions, while RPG primarily use signal enhancement. These results suggest that fast-action gaming can affect neural strategies and the corresponding networks underlying attention, presumably by training mechanisms of distractor suppression.
2011-01-01
Introduction The pan-HDAC inhibitor (HDACI) suberoylanilide hydroxamic acid (SAHA) has previously shown to be a radio-sensitizer to conventional photon radiotherapy (XRT) in pediatric sarcoma cell lines. Here, we investigate its effect on the response of two sarcoma cell lines and a normal tissue cell line to heavy ion irradiation (HIT). Materials and methods Clonogenic assays after different doses of heavy ions were performed. DNA damage and repair were evaluated by measuring γH2AX via flow-cytometry. Apoptosis and cell cycle analysis were also measured via flow cytometry. Protein expression of repair proteins, p53 and p21 were measured using immunoblot analysis. Changes of nuclear architecture after treatment with SAHA and HIT were observed in one of the sarcoma cell lines via light microscopy after staining towards chromatin and γH2AX. Results Corresponding with previously reported photon data, SAHA lead to an increase of sensitivity to heavy ions along with an increase of DSB and apoptosis in the two sarcoma cell lines. In contrast, in the osteoblast cell line (hFOB 1.19), the combination of SAHA and HIT showed a significant radio-protective effect. Laser scanning microscopy revealed no significant morphologic changes after HIT compared to the combined treatment with SAHA. Immunoblot analysis revealed no significant up or down regulation of p53. However, p21 was significantly increased by SAHA and combination treatment as compared to HIT only in the two sarcoma cell lines - again in contrast to the osteoblast cell line. Changes in the repair kinetics of DSB p53-independent apoptosis with p21 involvement may be part of the underlying mechanisms for radio-sensitization by SAHA. Conclusion Our in vitro data suggest an increase of the therapeutic ratio by the combination of SAHA with HIT in infantile sarcoma cell lines. PMID:21933400
Plasma response to sustainment with imposed-dynamo current drive in HIT-SI and HIT-SI3
NASA Astrophysics Data System (ADS)
Hossack, A. C.; Jarboe, T. R.; Chandra, R. N.; Morgan, K. D.; Sutherland, D. A.; Penna, J. M.; Everson, C. J.; Nelson, B. A.
2017-07-01
The helicity injected torus—steady inductive (HIT-SI) program studies efficient, steady-state current drive for magnetic confinement plasmas using a novel experimental method. Stable, high-beta spheromaks have been sustained using steady, inductive current drive. Externally induced loop voltage and magnetic flux are oscillated together so that helicity and power injection are always positive, sustaining the edge plasma current indefinitely. Imposed-dynamo current drive (IDCD) theory further shows that the entire plasma current is sustained. The method is ideal for low aspect ratio, toroidal geometries with closed flux surfaces. Experimental studies of spheromak plasmas sustained with IDCD have shown stable magnetic profiles with evidence of pressure confinement. New measurements show coherent motion of a stable spheromak in response to the imposed perturbations. On the original device two helicity injectors were mounted on either side of the spheromak and the injected mode spectrum was predominantly n = 1. Coherent, rigid motion indicates that the spheromak is stable and a lack of plasma-generated n = 1 energy indicates that the maximum q is maintained below 1 during sustainment. Results from the HIT-SI3 device are also presented. Three inductive helicity injectors are mounted on one side of the spheromak flux conserver. Varying the relative injector phasing changes the injected mode spectrum which includes n = 2, 3, and higher modes.
Early post-tsunami disaster medical assistance to Banda Aceh: a personal account.
Garner, Alan A; Harrison, Ken
2006-02-01
The south Asian tsunami on 26 December, 2004, saw Australia deploy civilian teams to an international disaster in large numbers for the first time. The logistics of supporting such teams in both a self sustainability capacity and medical equipment had not previously been planned for or tested. For the first Australian team deployed to Banda Aceh, which arrived on the fourth day after the tsunami, equipment sourced from the New South Wales Fire Brigades Urban Search and Rescue (US&R) cache supplied all food, water, tents, generators and sleeping equipment. The medical equipment was largely sourced from the CareFlight US&R medical cache. There were significant deficits in surgical equipment as the medical cache had not been designed to provide a stand alone surgical capability. This resulted in the need for substantial improvisation by the surgical teams during the deployment. Despite this, the team performed nearly 140 major procedures in austere circumstances and significantly contributed to the early international response to this major humanitarian disaster.
Multicore and GPU algorithms for Nussinov RNA folding
2014-01-01
Background One segment of a RNA sequence might be paired with another segment of the same RNA sequence due to the force of hydrogen bonds. This two-dimensional structure is called the RNA sequence's secondary structure. Several algorithms have been proposed to predict an RNA sequence's secondary structure. These algorithms are referred to as RNA folding algorithms. Results We develop cache efficient, multicore, and GPU algorithms for RNA folding using Nussinov's algorithm. Conclusions Our cache efficient algorithm provides a speedup between 1.6 and 3.0 relative to a naive straightforward single core code. The multicore version of the cache efficient single core algorithm provides a speedup, relative to the naive single core algorithm, between 7.5 and 14.0 on a 6 core hyperthreaded CPU. Our GPU algorithm for the NVIDIA C2050 is up to 1582 times as fast as the naive single core algorithm and between 5.1 and 11.2 times as fast as the fastest previously known GPU algorithm for Nussinov RNA folding. PMID:25082539
EMR Database Upgrade from MUMPS to CACHE: Lessons Learned.
Alotaibi, Abduallah; Emshary, Mshary; Househ, Mowafa
2014-01-01
Over the past few years, Saudi hospitals have been implementing and upgrading Electronic Medical Record Systems (EMRs) to ensure secure data transfer and exchange between EMRs.This paper focuses on the process and lessons learned in upgrading the MUMPS database to a the newer Caché database to ensure the integrity of electronic data transfer within a local Saudi hospital. This paper examines the steps taken by the departments concerned, their action plans and how the change process was managed. Results show that user satisfaction was achieved after the upgrade was completed. The system was stable and offered better healthcare quality to patients as a result of the data exchange. Hardware infrastructure upgrades improved scalability and software upgrades to Caché improved stability. The overall performance was enhanced and new functions were added (CPOE) during the upgrades. The essons learned were: 1) Involve higher management; 2) Research multiple solutions available in the market; 3) Plan for a variety of implementation scenarios.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nicolae, Bogdan; Riteau, Pierre; Keahey, Kate
Storage elasticity on IaaS clouds is a crucial feature in the age of data-intensive computing, especially when considering fluctuations of I/O throughput. This paper provides a transparent solution that automatically boosts I/O bandwidth during peaks for underlying virtual disks, effectively avoiding over-provisioning without performance loss. The authors' proposal relies on the idea of leveraging short-lived virtual disks of better performance characteristics (and thus more expensive) to act during peaks as a caching layer for the persistent virtual disks where the application data is stored. Furthermore, they introduce a performance and cost prediction methodology that can be used both independently tomore » estimate in advance what trade-off between performance and cost is possible, as well as an optimization technique that enables better cache size selection to meet the desired performance level with minimal cost. The authors demonstrate the benefits of their proposal both for microbenchmarks and for two real-life applications using large-scale experiments.« less
The Science of Computing: Virtual Memory
NASA Technical Reports Server (NTRS)
Denning, Peter J.
1986-01-01
In the March-April issue, I described how a computer's storage system is organized as a hierarchy consisting of cache, main memory, and secondary memory (e.g., disk). The cache and main memory form a subsystem that functions like main memory but attains speeds approaching cache. What happens if a program and its data are too large for the main memory? This is not a frivolous question. Every generation of computer users has been frustrated by insufficient memory. A new line of computers may have sufficient storage for the computations of its predecessor, but new programs will soon exhaust its capacity. In 1960, a longrange planning committee at MIT dared to dream of a computer with 1 million words of main memory. In 1985, the Cray-2 was delivered with 256 million words. Computational physicists dream of computers with 1 billion words. Computer architects have done an outstanding job of enlarging main memories yet they have never kept up with demand. Only the shortsighted believe they can.
Methods for compressible fluid simulation on GPUs using high-order finite differences
NASA Astrophysics Data System (ADS)
Pekkilä, Johannes; Väisälä, Miikka S.; Käpylä, Maarit J.; Käpylä, Petri J.; Anjum, Omer
2017-08-01
We focus on implementing and optimizing a sixth-order finite-difference solver for simulating compressible fluids on a GPU using third-order Runge-Kutta integration. Since graphics processing units perform well in data-parallel tasks, this makes them an attractive platform for fluid simulation. However, high-order stencil computation is memory-intensive with respect to both main memory and the caches of the GPU. We present two approaches for simulating compressible fluids using 55-point and 19-point stencils. We seek to reduce the requirements for memory bandwidth and cache size in our methods by using cache blocking and decomposing a latency-bound kernel into several bandwidth-bound kernels. Our fastest implementation is bandwidth-bound and integrates 343 million grid points per second on a Tesla K40t GPU, achieving a 3 . 6 × speedup over a comparable hydrodynamics solver benchmarked on two Intel Xeon E5-2690v3 processors. Our alternative GPU implementation is latency-bound and achieves the rate of 168 million updates per second.
Zhang, Mingming; Steele, Michael A; Yi, Xianfeng
2013-11-01
The question of how tannin affects feeding and hoarding preferences of rodents still remains poorly understood, in part, because it is difficult to control for other seed traits when considering the sole effect of tannin. Here, we constructed a series of artificial 'seeds' with different tannin levels, made from wheat flour, peanut powder and hydrolysable tannins, to determine the direct effects of tannin on both feeding and hoarding preferences. We first presented 'seeds' to individual rodents of two species (Tamias sibiricus and Apodemus peninsulae) confined in semi-natural enclosures and then monitored patterns of seed dispersal and consumption by free-ranging animals in a temperate forest in the Xiaoxing'an Mountains, Heilongjiang Province of China. Our results showed that small rodents displayed a significant preference for low-tannin 'seeds' for both consumption and caching in both captive and field experiments. Moreover, our two-year study consistently showed that tannin concentration was significantly and negatively correlated with the number of cached 'seeds' at both the individual and population levels. Seed size, compared with tannin concentrations, appeared to have little effect on dispersal distances and the number of 'seeds' cached. Low-tannin 'seeds' tended to be dispersed greater distances by rodents in the field than those with higher levels of tannin. These results failed to support those of previous reports indicating that acorns containing higher tannins are more likely to be cached by food hoarding animals. Copyright © 2013 Elsevier B.V. All rights reserved.
Thieving rodents as substitute dispersers of megafaunal seeds.
Jansen, Patrick A; Hirsch, Ben T; Emsens, Willem-Jan; Zamora-Gutierrez, Veronica; Wikelski, Martin; Kays, Roland
2012-07-31
The Neotropics have many plant species that seem to be adapted for seed dispersal by megafauna that went extinct in the late Pleistocene. Given the crucial importance of seed dispersal for plant persistence, it remains a mystery how these plants have survived more than 10,000 y without their mutualist dispersers. Here we present support for the hypothesis that secondary seed dispersal by scatter-hoarding rodents has facilitated the persistence of these large-seeded species. We used miniature radio transmitters to track the dispersal of reputedly megafaunal seeds by Central American agoutis, which scatter-hoard seeds in shallow caches in the soil throughout the forest. We found that seeds were initially cached at mostly short distances and then quickly dug up again. However, rather than eating the recovered seeds, agoutis continued to move and recache the seeds, up to 36 times. Agoutis dispersed an estimated 35% of seeds for >100 m. An estimated 14% of the cached seeds survived to the next year, when a new fruit crop became available to the rodents. Serial video-monitoring of cached seeds revealed that the stepwise dispersal was caused by agoutis repeatedly stealing and recaching each other's buried seeds. Although previous studies suggest that rodents are poor dispersers, we demonstrate that communities of rodents can in fact provide highly effective long-distance seed dispersal. Our findings suggest that thieving scatter-hoarding rodents could substitute for extinct megafaunal seed dispersers of tropical large-seeded trees.
Storageless and caching Tier-2 models in the UK context
NASA Astrophysics Data System (ADS)
Cadellin Skipsey, Samuel; Dewhurst, Alastair; Crooks, David; MacMahon, Ewan; Roy, Gareth; Smith, Oliver; Mohammed, Kashif; Brew, Chris; Britton, David
2017-10-01
Operational and other pressures have lead to WLCG experiments moving increasingly to a stratified model for Tier-2 resources, where “fat” Tier-2s (“T2Ds”) and “thin” Tier-2s (“T2Cs”) provide different levels of service. In the UK, this distinction is also encouraged by the terms of the current GridPP5 funding model. In anticipation of this, testing has been performed on the implications, and potential implementation, of such a distinction in our resources. In particular, this presentation presents the results of testing of storage T2Cs, where the “thin” nature is expressed by the site having either no local data storage, or only a thin caching layer; data is streamed or copied from a “nearby” T2D when needed by jobs. In OSG, this model has been adopted successfully for CMS AAA sites; but the network topology and capacity in the USA is significantly different to that in the UK (and much of Europe). We present the result of several operational tests: the in-production University College London (UCL) site, which runs ATLAS workloads using storage at the Queen Mary University of London (QMUL) site; the Oxford site, which has had scaling tests performed against T2Ds in various locations in the UK (to test network effects); and the Durham site, which has been testing the specific ATLAS caching solution of “Rucio Cache” integration with ARC’s caching layer.
On the Efficacy of Source Code Optimizations for Cache-Based Systems
NASA Technical Reports Server (NTRS)
VanderWijngaart, Rob F.; Saphir, William C.
1998-01-01
Obtaining high performance without machine-specific tuning is an important goal of scientific application programmers. Since most scientific processing is done on commodity microprocessors with hierarchical memory systems, this goal of "portable performance" can be achieved if a common set of optimization principles is effective for all such systems. It is widely believed, or at least hoped, that portable performance can be realized. The rule of thumb for optimization on hierarchical memory systems is to maximize temporal and spatial locality of memory references by reusing data and minimizing memory access stride. We investigate the effects of a number of optimizations on the performance of three related kernels taken from a computational fluid dynamics application. Timing the kernels on a range of processors, we observe an inconsistent and often counterintuitive impact of the optimizations on performance. In particular, code variations that have a positive impact on one architecture can have a negative impact on another, and variations expected to be unimportant can produce large effects. Moreover, we find that cache miss rates - as reported by a cache simulation tool, and confirmed by hardware counters - only partially explain the results. By contrast, the compiler-generated assembly code provides more insight by revealing the importance of processor-specific instructions and of compiler maturity, both of which strongly, and sometimes unexpectedly, influence performance. We conclude that it is difficult to obtain performance portability on modern cache-based computers, and comment on the implications of this result.
On the Efficacy of Source Code Optimizations for Cache-Based Systems
NASA Technical Reports Server (NTRS)
VanderWijngaart, Rob F.; Saphir, William C.; Saini, Subhash (Technical Monitor)
1998-01-01
Obtaining high performance without machine-specific tuning is an important goal of scientific application programmers. Since most scientific processing is done on commodity microprocessors with hierarchical memory systems, this goal of "portable performance" can be achieved if a common set of optimization principles is effective for all such systems. It is widely believed, or at least hoped, that portable performance can be realized. The rule of thumb for optimization on hierarchical memory systems is to maximize temporal and spatial locality of memory references by reusing data and minimizing memory access stride. We investigate the effects of a number of optimizations on the performance of three related kernels taken from a computational fluid dynamics application. Timing the kernels on a range of processors, we observe an inconsistent and often counterintuitive impact of the optimizations on performance. In particular, code variations that have a positive impact on one architecture can have a negative impact on another, and variations expected to be unimportant can produce large effects. Moreover, we find that cache miss rates-as reported by a cache simulation tool, and confirmed by hardware counters-only partially explain the results. By contrast, the compiler-generated assembly code provides more insight by revealing the importance of processor-specific instructions and of compiler maturity, both of which strongly, and sometimes unexpectedly, influence performance. We conclude that it is difficult to obtain performance portability on modern cache-based computers, and comment on the implications of this result.
Two Balls' Collision of Mass Ratio 3:1
ERIC Educational Resources Information Center
Ogawara, Yasuo; Hull, Michael M.
2018-01-01
Students will sometimes ask why momentum and kinetic energy concepts are both necessary. When physics teachers demonstrate situations that require both an understanding of kinetic energy and momentum, a favorite is Newton's cradle, or a comparable demonstration of two balls of equal mass hitting each other. However, in addition to the case of two…
Substituted 2'-hydroxychalcones were found to give an equilibrium mixture of the starting chalcone and the corresponding flavanone in 4:6 - 1:3 ratio in the presence of various supports and MW irradiation. MW irradiation of 2'-hydroxychalcones in the presence of DBU gave two hit...
Aftermath of early Hit-and-Run collisions in the Inner Solar System
NASA Astrophysics Data System (ADS)
Sarid, Gal; Stewart, Sarah T.; Leinhardt, zoe M.
2015-08-01
Planet formation epoch, in the terrestrial planet region and the asteroid belt, was characterized by a vigorous dynamical environment that was conducive to giant impacts among planetary embryos and asteroidal parent bodies, leading to diverse outcomes. Among these the greatest potential for producing diverse end-members lies is the erosive Hit-and-Run regime (small mass ratios, off-axis oblique impacts and non-negligible ejected mass), which is also more probable in terms of the early dynamical encounter configuration in the inner solar system. This collision regime has been invoked to explain outstanding issues, such as planetary volatile loss records, origin of the Moon and mantle stripping from Mercury and some of the larger asteroids (Vesta, Psyche).We performed and analyzed a set of simulations of Hit-and-Run events, covering a large range of mass ratios (1-20), impact parameters (0.25-0.96, for near head-on to barely grazing) and impact velocities (~1.5-5 times the mutual escape velocity, as dependent on the mass ratio). We used an SPH code with tabulated EOS and a nominal simlated time >1 day, to track the collisional shock processing and the provenance of material components. of collision debris. Prior to impact runs, all bodies were allowed to initially settle to negligible particle velocities in isolation, within ~20 simulated hrs. The total number of particles involved in each of our collision simulations was between (1-3 x 105). Resulting configurations include stripped mantles, melting/vaporization of rock and/or iron cores and strong variations of asteroid parent bodies fromcanonical chondritic composition.In the context of large planetary formation simulations, velocity and impact angle distributions are necessary to asses impact probabilities. The mass distribution and interaction within planetary embryo and asteroid swarms depends both on gravitational dynamics and the applied fragmentation mechanism. We will present results pertaining to general projectile remnant scaling relations, constitution of ejected unbound material and the composition of variedcollision remnants, which become available to seed the asteroid belt.
2006-06-14
Robert Graybill . A Raw hoard for the use of this project was provided by the Computer Architecture Croup at the Massachusetts Institute of Technology...simulator is presented by MIT as being an accurate model of the Raw chip, we have found that it does not accurately model the board. Our comparison...G4 processor, model 7410. with a 32 kbyte level-1 cache on-chip and a 2 Mbyte L2 cache connected through a 250 MH/ bus [12]. Each node has 256 Mbyte
Miyaoka, Masashi; Kikuti, Yara Y; Carreras, Joaquim; Ikoma, Haruka; Hiraiwa, Shinichiro; Ichiki, Akifumi; Kojima, Minoru; Ando, Kiyoshi; Yokose, Tomoyuki; Sakai, Rika; Hoshikawa, Masahiro; Tomita, Naoto; Miura, Ikuo; Takata, Katsuyoshi; Yoshino, Tadashi; Takizawa, Jun; Bea, Silvia; Campo, Elias; Nakamura, Naoya
2018-02-01
Most high-grade B-cell lymphomas with MYC and BCL2 and/or BCL6 rearrangements are aggressive B-cell lymphomas. Occasional double-hit follicular lymphomas have been described but the clinicopathological features of these tumors are not well known. To clarify the characteristics of double-hit follicular lymphomas, we analyzed 10 cases of double-hit follicular lymphomas and 15 cases of high-grade B-cell lymphomas with MYC and BCL2 and/or BCL6 rearrangements for clinicopathological and genome-wide copy-number alterations and copy-neutral loss-of-heterozygosity profiles. For double-hit follicular lymphomas, the median age was 67.5 years (range: 48-82 years). The female/male ratio was 2.3. Eight patients presented with advanced clinical stage. The median follow-up time was 20 months (range: 1-132 months). At the end of the follow-up, 8 patients were alive, 2 patients were dead including 1 patient with diffuse large B-cell lymphoma transformation. Rearrangements of MYC/BCL2, MYC/BCL6, and MYC/BCL2/BCL6 were seen in 8, 1, and 1 cases, respectively. The partner of MYC was IGH in 6 cases. There were no cases of histological grade 1, 4 cases of grade 2, 5 cases of grade 3a, and 1 case of grade 3b. Two cases of grade 3a exhibited immunoblast-like morphology. Immunohistochemistry demonstrated 9 cases with ≥50% MYC-positive cells. There was significant difference in MYC intensity (P=0.00004) and MIB-1 positivity (P=0.001) between double-hit follicular lymphomas and high-grade B-cell lymphomas with MYC and BCL2 and/or BCL6 rearrangements. The genome profile of double-hit follicular lymphomas was comparable with conventional follicular lymphomas (GSE67385, n=198) with characteristic gains of 2p25.3-p11.1, 7p22.3-q36.3, 12q11-q24.33, and loss of 18q21.32-q23 (P<0.05). In comparison with high-grade B-cell lymphomas with MYC and BCL2 and/or BCL6 rearrangements, double-hit follicular lymphomas had fewer copy-number alterations and minimal common region of gain at 2p16.1 (70%), locus also significant against conventional follicular lymphomas (P=0.0001). In summary, double-hit follicular lymphomas tended to be high-grade histology, high MYC protein expression, high MYC/IGH fusion, and minimal common region of gain at 2p16.1. Double-hit follicular lymphomas seemed to be a different disease from high-grade B-cell lymphomas with MYC and BCL2 and/or BCL6 rearrangements and have an indolent clinical behavior similar to follicular lymphomas without MYC rearrangement.
Domagalski, Joseph L.; Alpers, Charles N.; Slotton, Darrell G.; Suchanek, Thomas H.; Ayers, Shaun M.
2004-01-01
Concentrations and mass loads of total mercury and methylmercury in streams draining abandoned mercury mines and near geothermal discharge in Cache Creek Basin, California, were measured during a 17-month period from January 2000 through May 2001. Rainfall and runoff averages during the study period were lower than long-term averages. Mass loads of mercury and methylmercury from upstream sources to downstream receiving waters, such as San Francisco Bay, were generally the highest during or after winter rainfall events. During the study period, mass loads of mercury and methylmercury from geothermal sources tended to be greater than those from abandoned mining areas because of a lack of large precipitation events capable of mobilizing significant amounts of either mercury-laden sediment or dissolved mercury and methylmercury from mine waste. Streambed sediments of Cache Creek are a source of mercury and methylmercury to downstream receiving bodies of water such as the Delta of the San Joaquin and Sacramento Rivers. Much of the mercury in these sediments was deposited over the last 150 years by erosion and stream discharge from abandoned mines or by continuous discharges from geothermal areas. Several geochemical constituents were useful as natural tracers for mining and geothermal areas. These constituents included aqueous concentrations of boron, chloride, lithium, and sulfate, and the stable isotopes of hydrogen and oxygen in water. Stable isotopes of water in areas draining geothermal discharges were enriched with more oxygen-18 relative to oxygen-16 than meteoric waters, whereas the enrichment by stable isotopes of water from much of the runoff from abandoned mines was similar to that of meteoric water. Geochemical signatures from stable isotopes and trace-element concentrations may be useful as tracers of total mercury or methylmercury from specific locations; however, mercury and methylmercury are not conservatively transported. A distinct mixing trend of trace elements and stable isotopes of hydrogen and oxygen from geothermal waters was apparent in Sulphur Creek and lower Bear Creek (tributaries to Cache Creek), but the signals are lost upon mixing with Cache Creek because of dilution.
What makes specialized food-caching mountain chickadees successful city slickers?
Kozlovsky, Dovid Y; Weissgerber, Emily A; Pravosudov, Vladimir V
2017-05-31
Anthropogenic environments are a dominant feature of the modern world; therefore, understanding which traits allow animals to succeed in these urban environments is especially important. Overall, generalist species are thought to be most successful in urban environments, with better general cognition and less neophobia as suggested critical traits. It is less clear, however, which traits would be favoured in urban environments in highly specialized species. Here, we compared highly specialized food-caching mountain chickadees living in an urban environment (Reno, NV, USA) with those living in their natural environment to investigate what makes this species successful in the city. Using a 'common garden' paradigm, we found that urban mountain chickadees tended to explore a novel environment faster and moved more frequently, were better at novel problem-solving, had better long-term spatial memory retention and had a larger telencephalon volume compared with forest chickadees. There were no significant differences between urban and forest chickadees in neophobia, food-caching rates, spatial memory acquisition, hippocampus volume, or the total number of hippocampal neurons. Our results partially support the idea that some traits associated with behavioural flexibility and innovation are associated with successful establishment in urban environments, but differences in long-term spatial memory retention suggest that even this trait specialized for food-caching may be advantageous. Our results highlight the importance of environmental context, species biology, and temporal aspects of invasion in understanding how urban environments are associated with behavioural and cognitive phenotypes and suggest that there is likely no one suite of traits that makes urban animals successful. © 2017 The Author(s).
Wang, Bo; Ives, Anthony R
2017-03-01
Individual variation in seed size and seed production is high in many plant species. How does this variation affect seed-dispersing animals and, in turn, the fitness of individual plants? In this study, we first surveyed intraspecific variation in seed mass and production in a population of a Chinese white pine, Pinus armandii. For 134 target trees investigated in 2012, there was very high variation in seed size, with mean seed mass varying among trees almost tenfold, from 0.038 to 0.361 g. Furthermore, 30 of the 134 trees produced seeds 2 years later, and for these individuals there was a correlation in seed mass of 0.59 between years, implying consistent differences among individuals. For a subset of 67 trees, we monitored the foraging preferences of scatter-hoarding rodents on a total of 15,301 seeds: 8380 were ignored, 3184 were eaten in situ, 2651 were eaten after being cached, and 395 were successfully dispersed (cached and left intact). At the scale of individual seeds, seed mass affected almost every decision that rodents made to eat, remove, and cache individual seeds. At the level of individual trees, larger seeds had increased probabilities of both predation and successful dispersal: the effects of mean seed size on costs (predation) and benefits (caching) balanced out. Thus, despite seed size affecting rodent decisions, variation among trees in dispersal success associated with mean seed size was small once seeds were harvested. This might explain, at least in part, the maintenance of high variation in mean seed mass among tree individuals.
PEM public key certificate cache server
NASA Astrophysics Data System (ADS)
Cheung, T.
1993-12-01
Privacy Enhanced Mail (PEM) provides privacy enhancement services to users of Internet electronic mail. Confidentiality, authentication, message integrity, and non-repudiation of origin are provided by applying cryptographic measures to messages transferred between end systems by the Message Transfer System. PEM supports both symmetric and asymmetric key distribution. However, the prevalent implementation uses a public key certificate-based strategy, modeled after the X.509 directory authentication framework. This scheme provides an infrastructure compatible with X.509. According to RFC 1422, public key certificates can be stored in directory servers, transmitted via non-secure message exchanges, or distributed via other means. Directory services provide a specialized distributed database for OSI applications. The directory contains information about objects and then provides structured mechanisms for accessing that information. Since directory services are not widely available now, a good approach is to manage certificates in a centralized certificate server. This document describes the detailed design of a centralized certificate cache serve. This server manages a cache of certificates and a cache of Certificate Revocation Lists (CRL's) for PEM applications. PEMapplications contact the server to obtain/store certificates and CRL's. The server software is programmed in C and ELROS. To use this server, ISODE has to be configured and installed properly. The ISODE library 'libisode.a' has to be linked together with this library because ELROS uses the transport layer functions provided by 'libisode.a.' The X.500 DAP library that is included with the ELROS distribution has to be linked in also, since the server uses the DAP library functions to communicate with directory servers.
Initial Performance Results on IBM POWER6
NASA Technical Reports Server (NTRS)
Saini, Subbash; Talcott, Dale; Jespersen, Dennis; Djomehri, Jahed; Jin, Haoqiang; Mehrotra, Piysuh
2008-01-01
The POWER5+ processor has a faster memory bus than that of the previous generation POWER5 processor (533 MHz vs. 400 MHz), but the measured per-core memory bandwidth of the latter is better than that of the former (5.7 GB/s vs. 4.3 GB/s). The reason for this is that in the POWER5+, the two cores on the chip share the L2 cache, L3 cache and memory bus. The memory controller is also on the chip and is shared by the two cores. This serializes the path to memory. For consistently good performance on a wide range of applications, the performance of the processor, the memory subsystem, and the interconnects (both latency and bandwidth) should be balanced. Recognizing this, IBM has designed the Power6 processor so as to avoid the bottlenecks due to the L2 cache, memory controller and buffer chips of the POWER5+. Unlike the POWER5+, each core in the POWER6 has its own L2 cache (4 MB - double that of the Power5+), memory controller and buffer chips. Each core in the POWER6 runs at 4.7 GHz instead of 1.9 GHz in POWER5+. In this paper, we evaluate the performance of a dual-core Power6 based IBM p6-570 system, and we compare its performance with that of a dual-core Power5+ based IBM p575+ system. In this evaluation, we have used the High- Performance Computing Challenge (HPCC) benchmarks, NAS Parallel Benchmarks (NPB), and four real-world applications--three from computational fluid dynamics and one from climate modeling.
Hothem, Roger L.; Trejo, Bonnie S.; Bauer, Marissa L.; Crayon, John J.
2008-01-01
To evaluate mercury (Hg) and other element exposure in cliff swallows (Petrochelidon pyrrhonota), eggs were collected from 16 sites within the mining-impacted Cache Creek watershed, Colusa, Lake, and Yolo counties, California, USA, in 1997-1998. Nestlings were collected from seven sites in 1998. Geometric mean total Hg (THg) concentrations ranged from 0.013 to 0.208 ??g/g wet weight (ww) in cliff swallow eggs and from 0.047 to 0.347 ??g/g ww in nestlings. Mercury detected in eggs generally followed the spatial distribution of Hg in the watershed based on proximity to both anthropogenic and natural sources. Mean Hg concentrations in samples of eggs and nestlings collected from sites near Hg sources were up to five and seven times higher, respectively, than in samples from reference sites within the watershed. Concentrations of other detected elements, including aluminum, beryllium, boron, calcium, manganese, strontium, and vanadium, were more frequently elevated at sites near Hg sources. Overall, Hg concentrations in eggs from Cache Creek were lower than those reported in eggs of tree swallows (Tachycineta bicolor) from highly contaminated locations in North America. Total Hg concentrations were lower in all Cache Creek egg samples than adverse effects levels established for other species. Total Hg concentrations in bullfrogs (Rana catesbeiana) and foothill yellow-legged frogs (Rana boylii) collected from 10 of the study sites were both positively correlated with THg concentrations in cliff swallow eggs. Our data suggest that cliff swallows are reliable bioindicators of environmental Hg. ?? Springer Science+Business Media, LLC 2007.
Planetary Sample Caching System Design Options
NASA Technical Reports Server (NTRS)
Collins, Curtis; Younse, Paulo; Backes, Paul
2009-01-01
Potential Mars Sample Return missions would aspire to collect small core and regolith samples using a rover with a sample acquisition tool and sample caching system. Samples would need to be stored in individual sealed tubes in a canister that could be transfered to a Mars ascent vehicle and returned to Earth. A sample handling, encapsulation and containerization system (SHEC) has been developed as part of an integrated system for acquiring and storing core samples for application to future potential MSR and other potential sample return missions. Requirements and design options for the SHEC system were studied and a recommended design concept developed. Two families of solutions were explored: 1)transfer of a raw sample from the tool to the SHEC subsystem and 2)transfer of a tube containing the sample to the SHEC subsystem. The recommended design utilizes sample tool bit change out as the mechanism for transferring tubes to and samples in tubes from the tool. The SHEC subsystem design, called the Bit Changeout Caching(BiCC) design, is intended for operations on a MER class rover.
Architectural Techniques For Managing Non-volatile Caches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mittal, Sparsh
As chip power dissipation becomes a critical challenge in scaling processor performance, computer architects are forced to fundamentally rethink the design of modern processors and hence, the chip-design industry is now at a major inflection point in its hardware roadmap. The high leakage power and low density of SRAM poses serious obstacles in its use for designing large on-chip caches and for this reason, researchers are exploring non-volatile memory (NVM) devices, such as spin torque transfer RAM, phase change RAM and resistive RAM. However, since NVMs are not strictly superior to SRAM, effective architectural techniques are required for making themmore » a universal memory solution. This book discusses techniques for designing processor caches using NVM devices. It presents algorithms and architectures for improving their energy efficiency, performance and lifetime. It also provides both qualitative and quantitative evaluation to help the reader gain insights and motivate them to explore further. This book will be highly useful for beginners as well as veterans in computer architecture, chip designers, product managers and technical marketing professionals.« less
[PACS: storage and retrieval of digital radiological image data].
Wirth, S; Treitl, M; Villain, S; Lucke, A; Nissen-Meyer, S; Mittermaier, I; Pfeifer, K-J; Reiser, M
2005-08-01
Efficient handling of both picture archiving and retrieval is a crucial factor when new PACS installations as well as technical upgrades are planned. For a large PACS installation for 200 actual studies, the number, modality,and body region of available priors were evaluated. In addition, image access time of 100 CT studies from hard disk (RAID), magneto-optic disk (MOD), and tape archives (TAPE) were accessed. For current examinations priors existed in 61.1% with an averaged quantity of 7.7 studies. Thereof 56.3% were within 0-3 months, 84.9% within 12 months, 91.7% within 24 months, and 96.2% within 36 months. On average, access to images from the hard disk cache was more than 100 times faster then from MOD or TAPE. Since only PACS RAID provides online image access, at least current imaging of the past 12 months should be available from cache. An accurate prefetching mechanism facilitates effective use of the expensive online cache area. For that, however, close interaction of PACS, RIS, and KIS is an indispensable prerequisite.
Empirical study of parallel LRU simulation algorithms
NASA Technical Reports Server (NTRS)
Carr, Eric; Nicol, David M.
1994-01-01
This paper reports on the performance of five parallel algorithms for simulating a fully associative cache operating under the LRU (Least-Recently-Used) replacement policy. Three of the algorithms are SIMD, and are implemented on the MasPar MP-2 architecture. Two other algorithms are parallelizations of an efficient serial algorithm on the Intel Paragon. One SIMD algorithm is quite simple, but its cost is linear in the cache size. The two other SIMD algorithm are more complex, but have costs that are independent on the cache size. Both the second and third SIMD algorithms compute all stack distances; the second SIMD algorithm is completely general, whereas the third SIMD algorithm presumes and takes advantage of bounds on the range of reference tags. Both MIMD algorithm implemented on the Paragon are general and compute all stack distances; they differ in one step that may affect their respective scalability. We assess the strengths and weaknesses of these algorithms as a function of problem size and characteristics, and compare their performance on traces derived from execution of three SPEC benchmark programs.
Experience with HEP analysis on mounted filesystems.
NASA Astrophysics Data System (ADS)
Fuhrmann, Patrick; Gasthuber, Martin; Kemp, Yves; Ozerov, Dmitry
2012-12-01
We present results on different approaches on mounted filesystems in use or under investigation at DESY. dCache, established since long as a storage system for physics data has implemented the NFS v4.1/pNFS protocol. New performance results will be shown with the most current version of the dCache server. In addition to the native usage of the mounted filesystem in a LAN environment, the results are given for the performance of the dCache NFS v4.1/pNFS in WAN case. Several commercial vendors are currently in alpha or beta phase of adding the NFS v4.1/pNFS protocol to their storage appliances. We will test some of these vendor solutions for their readiness for HEP analysis. DESY has recently purchased an IBM Sonas system. We will present the result of a thorough performance evaluation using the native protocols NFS (v3 or v4) and GPFS. As the emphasis is on the usability for end user analysis, we will use latest ROOT versions and current end user analysis code for benchmark scenarios.
Multiple channel data acquisition system
Crawley, H. Bert; Rosenberg, Eli I.; Meyer, W. Thomas; Gorbics, Mark S.; Thomas, William D.; McKay, Roy L.; Homer, Jr., John F.
1990-05-22
A multiple channel data acquisition system for the transfer of large amounts of data from a multiplicity of data channels has a plurality of modules which operate in parallel to convert analog signals to digital data and transfer that data to a communications host via a FASTBUS. Each module has a plurality of submodules which include a front end buffer (FEB) connected to input circuitry having an analog to digital converter with cache memory for each of a plurality of channels. The submodules are interfaced with the FASTBUS via a FASTBUS coupler which controls a module bus and a module memory. The system is triggered to effect rapid parallel data samplings which are stored to the cache memories. The cache memories are uploaded to the FEBs during which zero suppression occurs. The data in the FEBs is reformatted and compressed by a local processor during transfer to the module memory. The FASTBUS coupler is used by the communications host to upload the compressed and formatted data from the module memory. The local processor executes programs which are downloaded to the module memory through the FASTBUS coupler.
Multiple channel data acquisition system
Crawley, H.B.; Rosenberg, E.I.; Meyer, W.T.; Gorbics, M.S.; Thomas, W.D.; McKay, R.L.; Homer, J.F. Jr.
1990-05-22
A multiple channel data acquisition system for the transfer of large amounts of data from a multiplicity of data channels has a plurality of modules which operate in parallel to convert analog signals to digital data and transfer that data to a communications host via a FASTBUS. Each module has a plurality of submodules which include a front end buffer (FEB) connected to input circuitry having an analog to digital converter with cache memory for each of a plurality of channels. The submodules are interfaced with the FASTBUS via a FASTBUS coupler which controls a module bus and a module memory. The system is triggered to effect rapid parallel data samplings which are stored to the cache memories. The cache memories are uploaded to the FEBs during which zero suppression occurs. The data in the FEBs is reformatted and compressed by a local processor during transfer to the module memory. The FASTBUS coupler is used by the communications host to upload the compressed and formatted data from the module memory. The local processor executes programs which are downloaded to the module memory through the FASTBUS coupler. 25 figs.
Chenoweth, Meghan J; Ware, Jennifer J; Zhu, Andy Z X; Cole, Christopher B; Cox, Lisa Sanderson; Nollen, Nikki; Ahluwalia, Jasjit S; Benowitz, Neal L; Schnoll, Robert A; Hawk, Larry W; Cinciripini, Paul M; George, Tony P; Lerman, Caryn; Knight, Joanne; Tyndale, Rachel F
2018-03-01
The activity of CYP2A6, the major nicotine-inactivating enzyme, is measurable in smokers using the nicotine metabolite ratio (NMR; 3'hydroxycotinine/cotinine). Due to its role in nicotine clearance, the NMR is associated with smoking behaviours and response to pharmacotherapies. The NMR is highly heritable (~80%), and on average lower in African Americans (AA) versus whites. We previously identified several reduce and loss-of-function CYP2A6 variants common in individuals of African descent. Our current aim was to identify novel genetic influences on the NMR in AA smokers using genome-wide approaches. Genome-wide association study (GWAS). Multiple sites within Canada and the United States. AA smokers from two clinical trials: Pharmacogenetics of Nicotine Addiction Treatment (PNAT)-2 (NCT01314001; n = 504) and Kick-it-at-Swope (KIS)-3 (NCT00666978; n = 450). Genome-wide SNP genotyping, the NMR (phenotype) and population substructure and NMR covariates. Meta-analysis revealed three independent chromosome 19 signals (rs12459249, rs111645190 and rs185430475) associated with the NMR. The top overall hit, rs12459249 (P = 1.47e-39; beta = 0.59 per C (versus T) allele, SE = 0.045), located ~9.5 kb 3' of CYP2A6, remained genome-wide significant after controlling for the common (~10% in AA) non-functional CYP2A6*17 allele. In contrast, rs111645190 and rs185430475 were not genome-wide significant when controlling for CYP2A6*17. In total, 96 signals associated with the NMR were identified; many were not found in prior NMR GWASs in individuals of European descent. The top hits were also associated with the NMR in a third cohort of AA (KIS2; n = 480). None of the hits were in UGT or OCT2 genes. Three independent chromosome 19 signals account for ~20% of the variability in the nicotine metabolite ratio in African American smokers. The hits identified may contribute to inter-ethnic variability in nicotine metabolism, smoking behaviours and tobacco-related disease risk. © 2017 Society for the Study of Addiction.
Acoustic emission intensity analysis of corrosion in prestressed concrete piles
NASA Astrophysics Data System (ADS)
Vélez, William; Matta, Fabio; Ziehl, Paul
2014-02-01
Corrosion of steel strands in prestressed concrete (PC) bridges may lead to substantial damage or collapse well before the end of the design life. Acoustic Emission (AE) is a suitable nondestructive technique to detect and locate corrosion in reinforced and prestressed concrete, which is key to prioritize inspection and maintenance. An effective tool to analyze damage-related AE data is intensity analysis (IA), which is based on two data trends, namely Severity (average signal strength of high amplitude hits) and Historic Index (ratio of the average signal strength of the most recent hits to the average of all hits). IA criteria for corrosion assessment in PC were recently proposed based on empirical evidence from accelerated corrosion tests. In this paper, AE data from prestressed and non-prestressed concrete pile specimens exposed to salt water wet-dry cycling for over 600 days are used to analyze the relation between Severity and Historic Index and actual corrosion. Evidence of corrosion is gained from the inspection of decommissioned specimens. The selection of suitable J and K parameters for IA is discussed, and an IA chart with updated corrosion criteria for PC piles is presented.
Hit discovery and hit-to-lead approaches.
Keseru, György M; Makara, Gergely M
2006-08-01
Hit discovery technologies range from traditional high-throughput screening to affinity selection of large libraries, fragment-based techniques and computer-aided de novo design, many of which have been extensively reviewed. Development of quality leads using hit confirmation and hit-to-lead approaches present their own challenges, depending on the hit discovery method used to identify the initial hits. In this paper, we summarize common industry practices adopted to tackle hit-to-lead challenges and review how the advantages and drawbacks of different hit discovery techniques could affect the various issues hit-to-lead groups face.
Parsons, Brendon A; Marney, Luke C; Siegler, W Christopher; Hoggard, Jamin C; Wright, Bob W; Synovec, Robert E
2015-04-07
Comprehensive two-dimensional (2D) gas chromatography coupled with time-of-flight mass spectrometry (GC × GC-TOFMS) is a versatile instrumental platform capable of collecting highly informative, yet highly complex, chemical data for a variety of samples. Fisher-ratio (F-ratio) analysis applied to the supervised comparison of sample classes algorithmically reduces complex GC × GC-TOFMS data sets to find class distinguishing chemical features. F-ratio analysis, using a tile-based algorithm, significantly reduces the adverse effects of chromatographic misalignment and spurious covariance of the detected signal, enhancing the discovery of true positives while simultaneously reducing the likelihood of detecting false positives. Herein, we report a study using tile-based F-ratio analysis whereby four non-native analytes were spiked into diesel fuel at several concentrations ranging from 0 to 100 ppm. Spike level comparisons were performed in two regimes: comparing the spiked samples to the nonspiked fuel matrix and to each other at relative concentration factors of two. Redundant hits were algorithmically removed by refocusing the tiled results onto the original high resolution pixel level data. To objectively limit the tile-based F-ratio results to only features which are statistically likely to be true positives, we developed a combinatorial technique using null class comparisons, called null distribution analysis, by which we determined a statistically defensible F-ratio cutoff for the analysis of the hit list. After applying null distribution analysis, spiked analytes were reliably discovered at ∼1 to ∼10 ppm (∼5 to ∼50 pg using a 200:1 split), depending upon the degree of mass spectral selectivity and 2D chromatographic resolution, with minimal occurrence of false positives. To place the relevance of this work among other methods in this field, results are compared to those for pixel and peak table-based approaches.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dykstra, D.; Blomer, J.
Both the CernVM File System (CVMFS) and the Frontier Distributed Database Caching System (Frontier) distribute centrally updated data worldwide for LHC experiments using http proxy caches. Neither system provides privacy or access control on reading the data, but both control access to updates of the data and can guarantee the authenticity and integrity of the data transferred to clients over the internet. CVMFS has since its early days required digital signatures and secure hashes on all distributed data, and recently Frontier has added X.509-based authenticity and integrity checking. In this paper we detail and compare the security models of CVMFSmore » and Frontier.« less
Efficient implementation of parallel three-dimensional FFT on clusters of PCs
NASA Astrophysics Data System (ADS)
Takahashi, Daisuke
2003-05-01
In this paper, we propose a high-performance parallel three-dimensional fast Fourier transform (FFT) algorithm on clusters of PCs. The three-dimensional FFT algorithm can be altered into a block three-dimensional FFT algorithm to reduce the number of cache misses. We show that the block three-dimensional FFT algorithm improves performance by utilizing the cache memory effectively. We use the block three-dimensional FFT algorithm to implement the parallel three-dimensional FFT algorithm. We succeeded in obtaining performance of over 1.3 GFLOPS on an 8-node dual Pentium III 1 GHz PC SMP cluster.
2016-11-09
Total Number: Sub Contractors (DD882) Names of Personnel receiving masters degrees Names of personnel receiving PHDs Names of other research staff...Broadcom 5720 QP 1Gb Network Daughter Card (2) Intel Xeon E5-2680 v3 2.5GHz, 30M Cache, 9.60GT/s QPI, Turbo, HT , 12C/24T (120W...Broadcom 5720 QP 1Gb Network Daughter Card (2) Intel Xeon E5-2680 v3 2.5GHz, 30M Cache, 9.60GT/s QPI, Turbo, HT , 12C/24T (120W
Rytuba, James J.; Hothem, Roger L.; Brussee, Brianne E.; Goldstein, Daniel; May, Jason T.
2015-01-01
The Cache Creek watershed lies within California's North Coast Range, an area with abundant geologic sources of mercury (Hg) and a long history of Hg contamination (Rytuba, 2000). Bear Creek, Cache Creek, and the North Fork of Cache Creek are the major streams of the Cache Creek watershed, encompassing 2978 km2. The Cache Creek watershed contains soils naturally enriched in Hg as well as natural springs (both hot and cold) with varying levels of aqueous Hg (Domagalski and others, 2004, Suchanek and others, 2004, Holloway and others 2009). All three tributaries are known to be significant sources of anthropogenically derived Hg from historic mines, both Hg and gold (Au), and associated ore storage/processing sites and facilities (Slotton and others, 1995, 2004; CVRWQCB, 2003; Schwarzbach and others, 2001; Gassel and others, 2005; Suchanek and others., 2004, 2008a, 2009). Historically, two of the primary sources of mercury contamination in the upper part of Bear Creek have been the Rathburn and Petray Hg Mines. The Rathburn Hg mine was discovered and initially mined in the early 1890s. The Rathburn and the more recently developed Petray open pit mines are localized along fault zones in serpentinite that has been altered and cut by quartz and chalcedony veins. Cold saline-carbonate springs are located perepheral to the Hg deposits and effluent from the springs locally has high concentrations of Hg (Slowey and Rytuba, 2008). Several ephemeral tributaries to Bear Creek drain the mine area which is located on federal land managed by the U.S. Bureau of Land Management (USBLM). The USBLM requested that the U.S. Geological Survey (USGS) measure and characterize Hg and other geochemical constituents in sediment, water, and biota to establish baseline information prior to remediation of the Rathburn and Petray mines. Samples sites were established in Bear Creek upstream and downstream from the mine area. This report is made in response to the USBLM request, the lead agency mandated to conduct a Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA) - Removal Site Investigation (RSI). The RSI applies to the possible removal of Hg-contaminated mine waste from Bear Creek. This report summarizes data obtained from field sampling of water, sediment, and biota in Bear Creek, above input from the mine area and downstream from the Rathburn-Petray mine area to the confluence with Cache Creek. Our results permit a preliminary assessment of the chemical constituents that could elevate levels of monomethyl Hg (MMeHg) in Bear Creek and its uptake by biota and provide baseline information for comparison to conditions after mine remediation is completed.
Warkentin, Theodore E
2016-10-28
HIT is an acquired antibody-mediated disorder strongly associated with thrombosis, including microthrombosis secondary to disseminated intravascular dissemination (DIC). The clinical features of HIT are reviewed from the perspective of the 4Ts scoring system for HIT, which emphasises its characteristic timing of onset of thrombocytopenia. HIT antibodies recognize multimolecular complexes of platelet factor 4 (PF4)/heparin. However, a subset of HIT sera recognise PF4 bound to platelet chondroitin sulfate; these antibodies activate platelets in vitro and in vivo even in the absence of heparin, thus explaining: delayed-onset HIT (where HIT begins or worsens after stopping heparin); persisting HIT (where HIT takes several weeks to recover); spontaneous HIT syndrome (a disorder clinically and serologically resembling HIT but without proximate heparin exposure); and fondaparinux-associated HIT (four distinct syndromes featuring thrombocytopenia that begins or worsens during treatment with fondaparinux), with a new patient case presented with ongoing thrombocytopenia (and fatal haemorrhage) during treatment of HIT with fondaparinux, with fondaparinux-dependent platelet activation induced by patient serum ("fondaparinux cross-reactivity"). Ironically, despite existence of fondaparinux-associated HIT, this pentasaccharide anticoagulant is a frequent treatment for HIT (including one used by the author). HIT can be confused with other disorders, including those with a) timing similar to HIT (e. g. abciximab-associated thrombocytopenia of delayed-onset); b) combined thrombocytopenia/thrombosis (e. g. symmetrical peripheral gangrene secondary to acute DIC and shock liver); and c) both timing of onset and thrombosis (e. g. warfarin-associated venous limb gangrene complicating cancer-associated DIC). By understanding clinical and pathophysiological similarities and differences between HIT and non-HIT mimicking disorders, the clinician is better able to make the correct diagnosis.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Luo, Y.; Cameron, K.W.
1998-11-24
Workload characterization has been proven an essential tool to architecture design and performance evaluation in both scientific and commercial computing areas. Traditional workload characterization techniques include FLOPS rate, cache miss ratios, CPI (cycles per instruction or IPC, instructions per cycle) etc. With the complexity of sophisticated modern superscalar microprocessors, these traditional characterization techniques are not powerful enough to pinpoint the performance bottleneck of an application on a specific microprocessor. They are also incapable of immediately demonstrating the potential performance benefit of any architectural or functional improvement in a new processor design. To solve these problems, many people rely on simulators,more » which have substantial constraints especially on large-scale scientific computing applications. This paper presents a new technique of characterizing applications at the instruction level using hardware performance counters. It has the advantage of collecting instruction-level characteristics in a few runs virtually without overhead or slowdown. A variety of instruction counts can be utilized to calculate some average abstract workload parameters corresponding to microprocessor pipelines or functional units. Based on the microprocessor architectural constraints and these calculated abstract parameters, the architectural performance bottleneck for a specific application can be estimated. In particular, the analysis results can provide some insight to the problem that only a small percentage of processor peak performance can be achieved even for many very cache-friendly codes. Meanwhile, the bottleneck estimation can provide suggestions about viable architectural/functional improvement for certain workloads. Eventually, these abstract parameters can lead to the creation of an analytical microprocessor pipeline model and memory hierarchy model.« less
Junqueira, Daniela R; Zorzela, Liliane M; Perini, Edson
2017-04-21
Heparin-induced thrombocytopenia (HIT) is an adverse drug reaction presenting as a prothrombotic disorder related to antibody-mediated platelet activation. It is a paradoxical immune reaction resulting in thrombin generation in vivo, which leads to a hypercoagulable state and the potential to initiate venous or arterial thrombosis. A number of factors are thought to influence the incidence of HIT including the type and preparation of heparin (unfractionated heparin (UFH) or low molecular weight heparin (LMWH)) and the heparin-exposed patient population, with the postoperative patient population at higher risk.Although LMWH has largely replaced UFH as a front-line therapy, there is evidence supporting a lack of superiority of LMWH compared with UFH regarding prevention of deep vein thrombosis and pulmonary embolism following surgery, and similar frequencies of bleeding have been described with LMWH and UFH. The decision as to which of these two preparations of heparin to use may thus be influenced by harmful effects such as HIT. We therefore sought to determine the relative impact of UFH and LMWH on HIT in postoperative patients receiving thromboembolism prophylaxis. This is an update of a review first published in 2012. The objective of this review was to compare the incidence of heparin-induced thrombocytopenia (HIT) and HIT complicated by venous thromboembolism in postoperative patients exposed to unfractionated heparin (UFH) versus low molecular weight heparin (LMWH). For this update, the Cochrane Vascular Information Specialist searched the Specialised Register (May 2016), CENTRAL (2016, Issue 4) and trials registries. The authors searched Lilacs (June 2016) and additional trials were sought from reference lists of relevant publications. We included randomised controlled trials (RCTs) in which participants were postoperative patients allocated to receive prophylaxis with UFH or LMWH, in a blinded or unblinded fashion. Studies were excluded if they did not use the accepted definition of HIT. This was defined as a relative reduction in the platelet count of 50% or greater from the postoperative peak (even if the platelet count at its lowest remained greater than 150 x 10 9 /L) occurring within five to 14 days after the surgery, with or without a thrombotic event occurring in this timeframe. Additionally, we required circulating antibodies associated with the syndrome to have been investigated through laboratory assays. Two review authors independently extracted data and assessed the risk of bias. Disagreements were resolved by consensus with participation of a third author. In this update, we included three trials involving 1398 postoperative participants. Participants were submitted to general surgical procedures, minor and major, and the minimum mean age was 49 years. Pooled analysis showed a significant reduction in the risk of HIT with LMWH compared with UFH (risk ratio (RR) 0.23, 95% confidence interval (CI) 0.07 to 0.73); low-quality evidence. The number needed to treat for an additional beneficial outcome (NNTB) was 59. The risk of HIT was consistently reduced comparing participants undergoing major surgical procedures exposed to LMWH or UFH (RR 0.22, 95% CI 0.06 to 0.75); low-quality evidence. The occurrence of HIT complicated by venous thromboembolism was significantly lower in participants receiving LMWH compared with UFH (RR 0.22, 95% CI 0.06 to 0.84); low-quality evidence. The NNTB was 75. Arterial thrombosis occurred in only one participant who received UFH. There were no amputations or deaths documented. Although limited evidence is available, it appears that HIT induced by both types of heparins is common in people undergoing major surgical procedures (incidence greater than 1% and less than 10%). This updated review demonstrated low-quality evidence of a lower incidence of HIT, and HIT complicated by venous thromboembolism, in postoperative patients undergoing thromboprophylaxis with LMWH compared with UFH. Similarily, the risk of HIT in people undergoing major surgical procedures was lower when treated with LMWH compared to UFH (low-quality evidence). The quality of the evidence was downgraded due to concerns about the risk of bias in the included studies and imprecision of the study results. These findings may support current clinical use of LMWH over UFH as front-line heparin therapy. However, our conclusions are limited and there was an unexpected paucity of RCTs including HIT as an outcome. To address the scarcity of clinically-relevant information on HIT, HIT must be included as a core harmful outcome in future RCTs of heparin.
Method for improving instrument response
Hahn, David W.; Hencken, Kenneth R.; Johnsen, Howard A.; Flower, William L.
2000-01-01
This invention pertains generally to a method for improving the accuracy of particle analysis under conditions of discrete particle loading and particularly to a method for improving signal-to-noise ratio and instrument response in laser spark spectroscopic analysis of particulate emissions. Under conditions of low particle density loading (particles/m.sup.3) resulting from low overall metal concentrations and/or large particle size uniform sampling can not be guaranteed. The present invention discloses a technique for separating laser sparks that arise from sample particles from those that do not; that is, a process for systematically "gating" the instrument response arising from "sampled" particles from those responses which do not, is dislosed as a solution to his problem. The disclosed approach is based on random sampling combined with a conditional analysis of each pulse. A threshold value is determined for the ratio of the intensity of a spectral line for a given element to a baseline region. If the threshold value is exceeded, the pulse is classified as a "hit" and that data is collected and an average spectrum is generated from an arithmetic average of "hits". The true metal concentration is determined from the averaged spectrum.
Holloway, J.M.; Goldhaber, M.B.; Morrison, J.M.
2009-01-01
Historic Hg mining in the Cache Creek watershed in the Central California Coast Range has contributed to the downstream transport of Hg to the San Francisco Bay-Delta. Different aspects of Hg mobilization in soils, including pedogenesis, fluvial redistribution of sediment, volatilization and eolian transport were considered. The greatest soil concentrations (>30 mg Hg kg-1) in Cache Creek are associated with mineralized serpentinite, the host rock for Hg deposits. Upland soils with non-mineralized serpentine and sedimentary parent material also had elevated concentrations (0.9-3.7 mg Hg kg-1) relative to the average concentration in the region and throughout the conterminous United States (0.06 mg kg-1). Erosion of soil and destabilized rock and mobilization of tailings and calcines into surrounding streams have contributed to Hg-rich alluvial soil forming in wetlands and floodplains. The concentration of Hg in floodplain sediment shows sediment dispersion from low-order catchments (5.6-9.6 mg Hg kg-1 in Sulphur Creek; 0.5-61 mg Hg kg-1 in Davis Creek) to Cache Creek (0.1-0.4 mg Hg kg-1). These sediments, deposited onto the floodplain during high-flow storm events, yield elevated Hg concentrations (0.2-55 mg Hg kg-1) in alluvial soils in upland watersheds. Alluvial soils within the Cache Creek watershed accumulate Hg from upstream mining areas, with concentrations between 0.06 and 0.22 mg Hg kg-1 measured in soils ~90 km downstream from Hg mining areas. Alluvial soils have accumulated Hg released through historic mining activities, remobilizing this Hg to streams as the soils erode.
Sample Acquisition and Caching architecture for the Mars Sample Return mission
NASA Astrophysics Data System (ADS)
Zacny, K.; Chu, P.; Cohen, J.; Paulsen, G.; Craft, J.; Szwarc, T.
This paper presents a Mars Sample Return (MSR) Sample Acquisition and Caching (SAC) study developed for the three rover platforms: MER, MER+, and MSL. The study took into account 26 SAC requirements provided by the NASA Mars Exploration Program Office. For this SAC architecture, the reduction of mission risk was chosen by us as having greater priority than mass or volume. For this reason, we selected a “ One Bit per Core” approach. The enabling technology for this architecture is Honeybee Robotics' “ eccentric tubes” core breakoff approach. The breakoff approach allows the drill bits to be relatively small in diameter and in turn lightweight. Hence, the bits could be returned to Earth with the cores inside them with only a modest increase to the total returned mass, but a significant decrease in complexity. Having dedicated bits allows a reduction in the number of core transfer steps and actuators. It also alleviates the bit life problem, eliminates cross contamination, and aids in hermetic sealing. An added advantage is faster drilling time, lower power, lower energy, and lower Weight on Bit (which reduces Arm preload requirements). Drill bits are based on the BigTooth bit concept, which allows re-use of the same bit multiple times, if necessary. The proposed SAC consists of a 1) Rotary-Percussive Core Drill, 2) Bit Storage Carousel, 3) Cache, 4) Robotic Arm, and 5) Rock Abrasion and Brushing Bit (RABBit), which is deployed using the Drill. The system also includes PreView bits (for viewing of cores prior to caching) and Powder bits for acquisition of regolith or cuttings. The SAC total system mass is less than 22 kg for MER and MER+ size rovers and less than 32 kg for the MSL-size rover.
Lucas, Jeffrey R.; Freeberg, Todd M.; Egbert, Jeremy; Schwabl, Hubert
2006-01-01
We tested for hormonal and behavioral differences between Carolina chickadees (Poecile carolinensis) taken from a disturbed (recently logged) forest, an undisturbed forest, or a residential site. We measured fecal corticosterone and body mass levels in the field, and fecal corticosterone, body mass, and caching behavior in an aviary experiment. In the field, birds from the disturbed forest exhibited significantly higher fecal corticosterone levels than birds from either the undisturbed forest or from the residential site. Birds from the disturbed forest also exhibited lower body mass than those from the undisturbed forest but higher body mass than those from the residential site. Our aviary results suggest that these physiological differences between field sites are the result of short-term responses to ecological factors: Neither body mass nor fecal corticosterone levels varied between birds captured at different sites. Aviary sample sizes were sufficient to detect seasonal variation in fecal corticosterone (lowest in summer), body mass (highest in spring), and rate of gain in body mass (highest in winter). Under “closed-economy” aviary conditions (all food available from a feeder in the aviary), there were no site differences in the percent of seeds taken from the feeder that were cached. However, under “open-economy” conditions (food occasionally available ad libitum), significantly fewer seeds were cached by birds from the disturbed forest compared to the undisturbed or residential sites. On average, there was only a two-fold difference in population-levels of fecal corticosterone. This difference is about the same as an increase in fecal corticosterone induced by a two-hour increase in food deprivation, and can not be considered to be an acute stress response to disturbance. PMID:16458312
Kuprewicz, Erin K.
2015-01-01
Scatter hoarding of seeds by animals contributes significantly to forest-level processes, including plant recruitment and forest community composition. However, the potential positive and negative effects of caching on seed survival, germination success, and seedling survival have rarely been assessed through experimental studies. Here, I tested the hypothesis that seed burial mimicking caches made by scatter hoarding Central American agoutis (Dasyprocta punctate) enhances seed survival, germination, and growth by protecting seeds from seed predators and providing favorable microhabitats for germination. In a series of experiments, I used simulated agouti seed caches to assess how hoarding affects seed predation by ground-dwelling invertebrates and vertebrates for four plant species. I tracked germination and seedling growth of intact and beetle-infested seeds and, using exclosures, monitored the effects of mammals on seedling survival through time. All experiments were conducted over three years in a lowland wet forest in Costa Rica. The majority of hoarded palm seeds escaped predation by both invertebrates and vertebrates while exposed seeds suffered high levels of infestation and removal. Hoarding had no effect on infestation rates of D. panamensis, but burial negatively affected germination success by preventing endocarp dehiscence. Non-infested palm seeds had higher germination success and produced larger seedlings than infested seeds. Seedlings of A. alatum and I. deltoidea suffered high mortality by seed-eating mammals. Hoarding protected most seeds from predators and enhanced germination success (except for D. panamensis) and seedling growth, although mammals killed many seedlings of two plant species; all seedling deaths were due to seed removal from the plant base. Using experimental caches, this study shows that scatter hoarding is beneficial to most seeds and may positively affect plant propagation in tropical forests, although tradeoffs in seed survival do exist. PMID:25970832
Buck, Harleah G; Harkness, Karen; Ali, Muhammad Usman; Carroll, Sandra L; Kryworuchko, Jennifer; McGillion, Michael
2017-04-01
Caregivers (CGs) contribute important assistance with heart failure (HF) self-care, including daily maintenance, symptom monitoring, and management. Until CGs' contributions to self-care can be quantified, it is impossible to characterize it, account for its impact on patient outcomes, or perform meaningful cost analyses. The purpose of this study was to conduct psychometric testing and item reduction on the recently developed 34-item Caregiver Contribution to Heart Failure Self-care (CACHS) instrument using classical and item response theory methods. Fifty CGs (mean age 63 years ±12.84; 70% female) recruited from a HF clinic completed the CACHS in 2014 and results evaluated using classical test theory and item response theory. Items would be deleted for low (<.05) or high (>.95) endorsement, low (<.3) or high (>.7) corrected item-total correlations, significant pairwise correlation coefficients, floor or ceiling effects, relatively low latent trait and item information function levels (<1.5 and p > .5), and differential item functioning. After analysis, 14 items were excluded, resulting in a 20-item instrument (self-care maintenance eight items; monitoring seven items; and management five items). Most items demonstrated moderate to high discrimination (median 2.13, minimum .77, maximum 5.05), and appropriate item difficulty (-2.7 to 1.4). Internal consistency reliability was excellent (Cronbach α = .94, average inter-item correlation = .41) with no ceiling effects. The newly developed 20-item version of the CACHS is supported by rigorous instrument development and represents a novel instrument to measure CGs' contribution to HF self-care. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Malek, Esmaiel; Davis, Tess; Martin, Randal S.; Silva, Philip J.
2006-02-01
Logan, Utah, USA, had the nation's worst air pollution on 15 January, 2004. The high concentration of PM 2.5 (particulates smaller than 2.5 μm in diameter) in the air resulted from geographical, meteorological, and environmental aspects of Cache Valley. A strong inversion (increase of temperature with height) and light precipitation and/or wind were the major causes for trapping pollutants in the air. Other meteorological factors enhancing the inversion were: the prolonged high atmospheric surface pressure, a snow-covered surface which plunged temperatures to as low as - 23.6 °C on January 23rd and high reflection of solar radiation (up to about 80%), which caused less solar radiation absorption during the day throughout the most part of January 2004. Among non-meteorological factors are Cache Valley's small-basin geographical structure which traps air, with no big body of water to help the air circulation (as a result of differential heating and cooling rates for land and water), motor vehicle emissions, and existence of excess ammonia gas as a byproduct of livestock manure and urine. Concentration of PM 2.5 was monitored in downtown Logan. On January 15, 2004, the 24-h, filter-based concentration reached about 132.5 μg per cubic meter of air, an astonishingly high value compared to the values of 65 μg m - 3 and over, indicating a health alert for everyone. These tiny particles in the air have an enormous impact on health, aggravating heart and lung disease, triggering asthma and even death. The causes of this inversion and some suggestions to alleviate the wintertime particle concentration in Cache Valley will be addressed in this article.
Turtle: identifying frequent k-mers with cache-efficient algorithms.
Roy, Rajat Shuvro; Bhattacharya, Debashish; Schliep, Alexander
2014-07-15
Counting the frequencies of k-mers in read libraries is often a first step in the analysis of high-throughput sequencing data. Infrequent k-mers are assumed to be a result of sequencing errors. The frequent k-mers constitute a reduced but error-free representation of the experiment, which can inform read error correction or serve as the input to de novo assembly methods. Ideally, the memory requirement for counting should be linear in the number of frequent k-mers and not in the, typically much larger, total number of k-mers in the read library. We present a novel method that balances time, space and accuracy requirements to efficiently extract frequent k-mers even for high-coverage libraries and large genomes such as human. Our method is designed to minimize cache misses in a cache-efficient manner by using a pattern-blocked Bloom filter to remove infrequent k-mers from consideration in combination with a novel sort-and-compact scheme, instead of a hash, for the actual counting. Although this increases theoretical complexity, the savings in cache misses reduce the empirical running times. A variant of method can resort to a counting Bloom filter for even larger savings in memory at the expense of false-negative rates in addition to the false-positive rates common to all Bloom filter-based approaches. A comparison with the state-of-the-art shows reduced memory requirements and running times. The tools are freely available for download at http://bioinformatics.rutgers.edu/Software/Turtle and http://figshare.com/articles/Turtle/791582. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
WENGREEN, H.; NELSON, C.; MUNGER, R.G.; CORCORAN, C.
2013-01-01
Objective To examine associations between frequency of ready-to-eat-cereal (RTEC) consumption and cognitive function among elderly men and women of the Cache County Study on Memory and Aging in Utah. Design A population-based prospective cohort study established in Cache County, Utah in 1995. Setting and Participants 3831 men and women > 65 years of age who were living in Cache County, Utah in 1995. Measurement Diet was assessed using a 142-item food frequency questionnaire at baseline. Cognitive function was assessed using an adapted version of the Modified Mini-Mental State examination (3MS) at baseline and three subsequent interviews over 11 years. RTEC consumption was defined as daily, weekly, or infrequent use. Results In multivariable models, more frequent RTEC consumption was not associated with a cognitive benefit. Those consuming RTEC weekly but less than daily scored higher on their baseline 3MS than did those consuming RTEC more or less frequently (91.7, 90.6, 90.6, respectively; p-value <0.001). This association was maintained across 11 years of observation such that those consuming RTEC weekly but less than daily declined on average 3.96 points compared to an average 5.13 and 4.57 point decline for those consuming cereal more or less frequently (p-value = 0.0009). Conclusion Those consuming RTEC at least daily had poorer cognitive performance at baseline and over 11 years of follow-up compared to those who consumed cereal more or less frequently. RTEC is a nutrient dense food, but should not replace the consumption of other healthy foods in the diets’ of elderly people. Associations between RTEC consumption, dietary patterns, and cognitive function deserve further study. PMID:21369668
Kind, Tobias; Fiehn, Oliver
2007-01-01
Background Structure elucidation of unknown small molecules by mass spectrometry is a challenge despite advances in instrumentation. The first crucial step is to obtain correct elemental compositions. In order to automatically constrain the thousands of possible candidate structures, rules need to be developed to select the most likely and chemically correct molecular formulas. Results An algorithm for filtering molecular formulas is derived from seven heuristic rules: (1) restrictions for the number of elements, (2) LEWIS and SENIOR chemical rules, (3) isotopic patterns, (4) hydrogen/carbon ratios, (5) element ratio of nitrogen, oxygen, phosphor, and sulphur versus carbon, (6) element ratio probabilities and (7) presence of trimethylsilylated compounds. Formulas are ranked according to their isotopic patterns and subsequently constrained by presence in public chemical databases. The seven rules were developed on 68,237 existing molecular formulas and were validated in four experiments. First, 432,968 formulas covering five million PubChem database entries were checked for consistency. Only 0.6% of these compounds did not pass all rules. Next, the rules were shown to effectively reducing the complement all eight billion theoretically possible C, H, N, S, O, P-formulas up to 2000 Da to only 623 million most probable elemental compositions. Thirdly 6,000 pharmaceutical, toxic and natural compounds were selected from DrugBank, TSCA and DNP databases. The correct formulas were retrieved as top hit at 80–99% probability when assuming data acquisition with complete resolution of unique compounds and 5% absolute isotope ratio deviation and 3 ppm mass accuracy. Last, some exemplary compounds were analyzed by Fourier transform ion cyclotron resonance mass spectrometry and by gas chromatography-time of flight mass spectrometry. In each case, the correct formula was ranked as top hit when combining the seven rules with database queries. Conclusion The seven rules enable an automatic exclusion of molecular formulas which are either wrong or which contain unlikely high or low number of elements. The correct molecular formula is assigned with a probability of 98% if the formula exists in a compound database. For truly novel compounds that are not present in databases, the correct formula is found in the first three hits with a probability of 65–81%. Corresponding software and supplemental data are available for downloads from the authors' website. PMID:17389044
Tew, Garry A; Weston, Matthew; Kothmann, Elke; Batterham, Alan M; Gray, Joanne; Kerr, Karen; Martin, Denis; Nawaz, Shah; Yates, David; Danjoux, Gerard
2014-01-10
In patients with large abdominal aortic aneurysm (AAA), open surgical or endovascular aneurysm repair procedures are often used to minimise the risk of aneurysm-related rupture and death; however, aneurysm repair itself carries a high risk. Low cardiopulmonary fitness is associated with an increased risk of early post-operative complications and death following elective AAA repair. Therefore, fitness should be enhanced before aneurysm repair. High-intensity interval exercise training (HIT) is a potent, time-efficient strategy for enhancing cardiopulmonary fitness. Here, we describe a feasibility study for a definitive trial of a pre-operative HIT intervention to improve post-operative outcomes in patients undergoing elective AAA repair. A minimum of 50 patients awaiting elective repair of a 5.5-7.0 cm infrarenal AAA will be allocated by minimisation to HIT or usual care control in a 1:1 ratio. The patients allocated to HIT will complete three hospital-based exercise sessions per week, for 4 weeks. Each session will include 2 or 4 min of high-intensity stationary cycling followed by the same duration of easy cycling or passive recovery, repeated until a total of 16 min of high-intensity exercise is accumulated. Outcomes to be assessed before randomisation and 24-48 h before aneurysm repair include cardiopulmonary fitness, maximum AAA diameter and health-related quality of life. In the post-operative period, we will record destination (ward or critical care unit), organ-specific morbidity, mortality and the durations of critical care and hospital stay. Twelve weeks after the discharge, participants will be interviewed to reassess quality of life and determine post-discharge healthcare utilisation. The costs associated with the exercise intervention and healthcare utilisation will be calculated. Ethics approval was secured through Sunderland Research Ethics Committee. The findings of the trial will be disseminated through peer-reviewed journals, and national and international presentations. Current Controlled Trials ISRCTN09433624.
The Screening Compound Collection: A Key Asset for Drug Discovery.
Boss, Christoph; Hazemann, Julien; Kimmerlin, Thierry; von Korff, Modest; Lüthi, Urs; Peter, Oliver; Sander, Thomas; Siegrist, Romain
2017-10-25
In this case study on an essential instrument of modern drug discovery, we summarize our successful efforts in the last four years toward enhancing the Actelion screening compound collection. A key organizational step was the establishment of the Compound Library Committee (CLC) in September 2013. This cross-functional team consisting of computational scientists, medicinal chemists and a biologist was endowed with a significant annual budget for regular new compound purchases. Based on an initial library analysis performed in 2013, the CLC developed a New Library Strategy. The established continuous library turn-over mode, and the screening library size of 300'000 compounds were maintained, while the structural library quality was increased. This was achieved by shifting the selection criteria from 'druglike' to 'leadlike' structures, enriching for non-flat structures, aiming for compound novelty, and increasing the ratio of higher cost 'Premium Compounds'. Novel chemical space was gained by adding natural compounds, macrocycles, designed and focused libraries to the collection, and through mutual exchanges of proprietary compounds with agrochemical companies. A comparative analysis in 2016 provided evidence for the positive impact of these measures. Screening the improved library has provided several highly promising hits, including a macrocyclic compound, that are currently followed up in different Hit-to-Lead and Lead Optimization programs. It is important to state that the goal of the CLC was not to achieve higher HTS hit rates, but to increase the chances of identified hits to serve as the basis of successful early drug discovery programs. The experience gathered so far legitimates the New Library Strategy.
Data Movement Dominates: Final Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jacob, Bruce L.
Over the past three years in this project, what we have observed is that the primary reason for data movement in large-scale systems is that the per-node capacity is not large enough—i.e., one of the solutions to the data-movement problem (certainly not the only solution that is required, but a significant one nonetheless) is to increase per-node capacity so that inter-node traffic is reduced. This unfortunately is not as simple as it sounds. Today’s main memory systems for datacenters, enterprise computing systems, and supercomputers, fail to provide high per-socket capacity [Dirik & Jacob 2009; Cooper-Balis et al. 2012], except atmore » extremely high price points (factors of 10–100x the cost/bit of consumer main-memory systems) [Stokes 2008]. The reason is that our choice of technology for today’s main memory systems—i.e., DRAM, which we have used as a main-memory technology since the 1970s [Jacob et al. 2007]—can no longer keep up with our needs for density and price per bit. Main memory systems have always been built from the cheapest, densest, lowest-power memory technology available, and DRAM is no longer the cheapest, the densest, nor the lowest-power storage technology out there. It is now time for DRAM to go the way that SRAM went: move out of the way for a cheaper, slower, denser storage technology, and become a cache instead. This inflection point has happened before, in the context of SRAM yielding to DRAM. There was once a time that SRAM was the storage technology of choice for all main memories [Tomasulo 1967; Thornton 1970; Kidder 1981]. However, once DRAM hit volume production in the 1970s and 80s, it supplanted SRAM as a main memory technology because it was cheaper, and it was denser. It also happened to be lower power, but that was not the primary consideration of the day. At the time, it was recognized that DRAM was much slower than SRAM, but it was only at the supercomputer level (For instance the Cray X-MP in the 1980s and its follow-on, the Cray Y-MP, in the 1990s) that could one afford to build ever- larger main memories out of SRAM—the reasoning for moving to DRAM was that an appropriately designed memory hierarchy, built of DRAM as main memory and SRAM as a cache, would approach the performance of SRAM, at the price-per-bit of DRAM [Mashey 1999]. Today it is quite clear that, were one to build an entire multi-gigabyte main memory out of SRAM instead of DRAM, one could improve the performance of almost any computer system by up to an order of magnitude—but this option is not even considered, because to build that system would be prohibitively expensive. It is now time to revisit the same design choice in the context of modern technologies and modern systems. For reasons both technical and economic, we can no longer afford to build ever-larger main memory systems out of DRAM. Flash memory, on the other hand, is significantly cheaper and denser than DRAM and therefore should take its place. While it is true that flash is significantly slower than DRAM, one can afford to build much larger main memories out of flash than out of DRAM, and we show that an appropriately designed memory hierarchy, built of flash as main memory and DRAM as a cache, will approach the performance of DRAM, at the price-per-bit of flash. In our studies as part of this project, we have investigated Non-Volatile Main Memory (NVMM), a new main-memory architecture for large-scale computing systems, one that is specifically designed to address the weaknesses described previously. In particular, it provides the following features: non-volatility: The bulk of the storage is comprised of NAND flash, and in this organization DRAM is used only as a cache, not as main memory. Furthermore, the flash is journaled, which means that operations such as checkpoint/restore are already built into the system. 1+ terabytes of storage per socket: SSDs and DRAM DIMMs have roughly the same form factor (several square inches of PCB surface area), and terabyte SSDs are now commonplace. performance approaching that of DRAM: DRAM is used as a cache to the flash system. price-per-bit approaching that of NAND: Flash is currently well under $0.50 per gigabyte; DDR3 SDRAM is currently just over $10 per gigabyte [Newegg 2014]. Even today, one can build an easily affordable main memory system with a terabyte or more of NAND storage per CPU socket (which would be extremely expensive were one to use DRAM), and our cycle- accurate, full-system experiments show that this can be done at a performance point that lies within a factor of two of DRAM.« less
Code of Federal Regulations, 2010 CFR
2010-10-01
... 42 Public Health 5 2010-10-01 2010-10-01 false Approval of the State Medicaid HIT plan, the HIT... Health CENTERS FOR MEDICARE & MEDICAID SERVICES, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED... Requirements Specific to the Medicaid Program § 495.344 Approval of the State Medicaid HIT plan, the HIT PAPD...
Accelerating a Particle-in-Cell Simulation Using a Hybrid Counting Sort
NASA Astrophysics Data System (ADS)
Bowers, K. J.
2001-11-01
In this article, performance limitations of the particle advance in a particle-in-cell (PIC) simulation are discussed. It is shown that the memory subsystem and cache-thrashing severely limit the speed of such simulations. Methods to implement a PIC simulation under such conditions are explored. An algorithm based on a counting sort is developed which effectively eliminates PIC simulation cache thrashing. Sustained performance gains of 40 to 70 percent are measured on commodity workstations for a minimal 2d2v electrostatic PIC simulation. More complete simulations are expected to have even better results as larger simulations are usually even more memory subsystem limited.
Introduction of Virtualization Technology to Multi-Process Model Checking
NASA Technical Reports Server (NTRS)
Leungwattanakit, Watcharin; Artho, Cyrille; Hagiya, Masami; Tanabe, Yoshinori; Yamamoto, Mitsuharu
2009-01-01
Model checkers find failures in software by exploring every possible execution schedule. Java PathFinder (JPF), a Java model checker, has been extended recently to cover networked applications by caching data transferred in a communication channel. A target process is executed by JPF, whereas its peer process runs on a regular virtual machine outside. However, non-deterministic target programs may produce different output data in each schedule, causing the cache to restart the peer process to handle the different set of data. Virtualization tools could help us restore previous states of peers, eliminating peer restart. This paper proposes the application of virtualization technology to networked model checking, concentrating on JPF.
Performance of hashed cache data migration schemes on multicomputers
NASA Technical Reports Server (NTRS)
Hiranandani, Seema; Saltz, Joel; Mehrotra, Piyush; Berryman, Harry
1991-01-01
After conducting an examination of several data-migration mechanisms which permit an explicit and controlled mapping of data to memory, a set of schemes for storage and retrieval of off-processor array elements is experimentally evaluated and modeled. All schemes considered have their basis in the use of hash tables for efficient access of nonlocal data. The techniques in question are those of hashed cache, partial enumeration, and full enumeration; in these, nonlocal data are stored in hash tables, so that the operative difference lies in the amount of memory used by each scheme and in the retrieval mechanism used for nonlocal data.
Huang, Wenting; Medeiros, L Jeffrey; Lin, Pei; Wang, Wei; Tang, Guilin; Khoury, Joseph; Konoplev, Sergej; Yin, C Cameron; Xu, Jie; Oki, Yasuhiro; Li, Shaoying
2018-05-21
High-grade B-cell lymphomas with MYC, BCL2, and BCL6 rearrangements (triple hit lymphoma) are uncommon. We studied the clinicopathologic features of 40 patients with triple hit lymphoma and compared them to 157 patients with MYC/BCL2 double hit lymphoma and 13 patients with MYC/BCL6 double hit lymphoma. The triple hit lymphoma group included 25 men and 15 women with a median age of 61 years (range, 34-85). Nine patients had a history of B-cell lymphoma. Histologically, 23 (58%) cases were diffuse large B-cell lymphoma and 17 cases had features of B-cell lymphoma, unclassifiable, with features intermediate between diffuse large B-cell lymphoma and Burkitt lymphoma. Most cases of triple hit lymphoma were positive for CD10 (100%), BCL2 (95%), BCL6 (82%), MYC (74%), and 71% with MYC and BCL2 coexpression. P53 was overexpressed in 29% of triple hit lymphoma cases. The clinicopathological features of triple hit lymphoma patients were similar to patients with MYC/BCL2 and MYC/BCL6 double hit lymphoma, except that triple hit lymphoma cases were more often CD10 positive compared with MYC/BCL6 double hit lymphoma (p < 0.05). Induction chemotherapy used was similar for patients with triple hit lymphoma and double hit lymphoma and overall survival in triple hit lymphoma patients was 17.6 months, similar to the overall survival of patients with double hit lymphoma (p = 0.67). Patients with triple hit lymphoma showing P53 overexpression had significantly worse overall survival compared with those without P53 overexpression (p = 0.04). On the other hand, double expressor status and prior history of B-cell lymphoma did not correlate with overall survival. In conclusion, most patients with triple hit lymphoma have an aggressive clinical course and poor prognosis and these tumors have a germinal center B-cell immunophenotype, similar to patients with double hit lymphomas. P53 expression is a poor prognostic factor in patients with triple hit lymphoma.
Cognitive caching promotes flexibility in task switching: evidence from event-related potentials.
Lange, Florian; Seer, Caroline; Müller, Dorothea; Kopp, Bruno
2015-12-08
Time-consuming processes of task-set reconfiguration have been shown to contribute to the costs of switching between cognitive tasks. We describe and probe a novel mechanism serving to reduce the costs of task-set reconfiguration. We propose that when individuals are uncertain about the currently valid task, one task set is activated for execution while other task sets are maintained at a pre-active state in cognitive cache. We tested this idea by assessing an event-related potential (ERP) index of task-set reconfiguration in a three-rule task-switching paradigm involving varying degrees of task uncertainty. In high-uncertainty conditions, two viable tasks were equally likely to be correct whereas in low-uncertainty conditions, one task was more likely than the other. ERP and performance measures indicated substantial costs of task-set reconfiguration when participants were required to switch away from a task that had been likely to be correct. In contrast, task-set-reconfiguration costs were markedly reduced when the previous task set was chosen under high task uncertainty. These results suggest that cognitive caching of alternative task sets adds to human cognitive flexibility under high task uncertainty.
Optimizing transformations of stencil operations for parallel cache-based architectures
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bassetti, F.; Davis, K.
This paper describes a new technique for optimizing serial and parallel stencil- and stencil-like operations for cache-based architectures. This technique takes advantage of the semantic knowledge implicity in stencil-like computations. The technique is implemented as a source-to-source program transformation; because of its specificity it could not be expected of a conventional compiler. Empirical results demonstrate a uniform factor of two speedup. The experiments clearly show the benefits of this technique to be a consequence, as intended, of the reduction in cache misses. The test codes are based on a 5-point stencil obtained by the discretization of the Poisson equation andmore » applied to a two-dimensional uniform grid using the Jacobi method as an iterative solver. Results are presented for a 1-D tiling for a single processor, and in parallel using 1-D data partition. For the parallel case both blocking and non-blocking communication are tested. The same scheme of experiments has bee n performed for the 2-D tiling case. However, for the parallel case the 2-D partitioning is not discussed here, so the parallel case handled for 2-D is 2-D tiling with 1-D data partitioning.« less
KENWARD, BEN; SCHLOEGL, CHRISTIAN; RUTZ, CHRISTIAN; WEIR, ALEXANDER A. S.; BUGNYAR, THOMAS; KACELNIK, ALEX
2015-01-01
New Caledonian crows (Corvus moneduloides) are prolific tool users in captivity and in the wild, and have an inherited predisposition to express tool-oriented behaviours. To further understand the evolution and development of tool use, we compared the development of object manipulation in New Caledonian crows and common ravens (Corvus corax), which do not routinely use tools. We found striking qualitative similarities in the ontogeny of tool-oriented behaviour in New Caledonian crows and food-caching behaviour in ravens. Given that the common ancestor of New Caledonian crows and ravens was almost certainly a caching species, we therefore propose that the basic action patterns for tool use in New Caledonian crows may have their evolutionary origins in caching behaviour. Noncombinatorial object manipulations had similar frequencies in the two species. However, frequencies of object combinations that are precursors to functional behaviour increased in New Caledonian crows and decreased in ravens throughout the study period, ending 6 weeks post-fledging. These quantitative observations are consistent with the hypothesis that New Caledonian crows develop tool-oriented behaviour because of an increased motivation to perform object combinations that facilitate the necessary learning. PMID:25892825
Research on mixed network architecture collaborative application model
NASA Astrophysics Data System (ADS)
Jing, Changfeng; Zhao, Xi'an; Liang, Song
2009-10-01
When facing complex requirements of city development, ever-growing spatial data, rapid development of geographical business and increasing business complexity, collaboration between multiple users and departments is needed urgently, however conventional GIS software (such as Client/Server model or Browser/Server model) are not support this well. Collaborative application is one of the good resolutions. Collaborative application has four main problems to resolve: consistency and co-edit conflict, real-time responsiveness, unconstrained operation, spatial data recoverability. In paper, application model called AMCM is put forward based on agent and multi-level cache. AMCM can be used in mixed network structure and supports distributed collaborative. Agent is an autonomous, interactive, initiative and reactive computing entity in a distributed environment. Agent has been used in many fields such as compute science and automation. Agent brings new methods for cooperation and the access for spatial data. Multi-level cache is a part of full data. It reduces the network load and improves the access and handle of spatial data, especially, in editing the spatial data. With agent technology, we make full use of its characteristics of intelligent for managing the cache and cooperative editing that brings a new method for distributed cooperation and improves the efficiency.
Observations of territorial breeding common ravens caching eggs of greater sage-grouse
Howe, Kristy B.; Coates, Peter S.
2015-01-01
Previous investigations using continuous video monitoring of greater sage-grouse Centrocercus urophasianus nests have unambiguously identified common ravens Corvus corax as an important egg predator within the western United States. The quantity of greater sage-grouse eggs an individual common raven consumes during the nesting period and the extent to which common ravens actively hunt greater sage-grouse nests are largely unknown. However, some evidence suggests that territorial breeding common ravens, rather than nonbreeding transients, are most likely responsible for nest depredations. We describe greater sage-grouse egg depredation observations obtained opportunistically from three common raven nests located in Idaho and Nevada where depredated greater sage-grouse eggs were found at or in the immediate vicinity of the nest site, including the caching of eggs in nearby rock crevices. We opportunistically monitored these nests by counting and removing depredated eggs and shell fragments from the nest sites during each visit to determine the extent to which the common raven pairs preyed on greater sage-grouse eggs. To our knowledge, our observations represent the first evidence that breeding, territorial pairs of common ravens cache greater sage-grouse eggs and are capable of depredating multiple greater sage-grouse nests.
WriteShield: A Pseudo Thin Client for Prevention of Information Leakage
NASA Astrophysics Data System (ADS)
Kirihata, Yasuhiro; Sameshima, Yoshiki; Onoyama, Takashi; Komoda, Norihisa
While thin-client systems are diffusing as an effective security method in enterprises and organizations, there is a new approach called pseudo thin-client system. In this system, local disks of clients are write-protected and user data is forced to save on the central file server to realize the same security effect of conventional thin-client systems. Since it takes purely the software-based simple approach, it does not require the hardware enhancement of network and servers to reduce the installation cost. However there are several problems such as no write control to external media, memory depletion possibility, and lower security because of the exceptional write permission to the system processes. In this paper, we propose WriteShield, a pseudo thin-client system which solves these issues. In this system, the local disks are write-protected with volume filter driver and it has a virtual cache mechanism to extend the memory cache size for the write protection. This paper presents design and implementation details of WriteShield. Besides we describe the security analysis and simulation evaluation of paging algorithms for virtual cache mechanism and measure the disk I/O performance to verify its feasibility in the actual environment.
Cognitive caching promotes flexibility in task switching: evidence from event-related potentials
Lange, Florian; Seer, Caroline; Müller, Dorothea; Kopp, Bruno
2015-01-01
Time-consuming processes of task-set reconfiguration have been shown to contribute to the costs of switching between cognitive tasks. We describe and probe a novel mechanism serving to reduce the costs of task-set reconfiguration. We propose that when individuals are uncertain about the currently valid task, one task set is activated for execution while other task sets are maintained at a pre-active state in cognitive cache. We tested this idea by assessing an event-related potential (ERP) index of task-set reconfiguration in a three-rule task-switching paradigm involving varying degrees of task uncertainty. In high-uncertainty conditions, two viable tasks were equally likely to be correct whereas in low-uncertainty conditions, one task was more likely than the other. ERP and performance measures indicated substantial costs of task-set reconfiguration when participants were required to switch away from a task that had been likely to be correct. In contrast, task-set-reconfiguration costs were markedly reduced when the previous task set was chosen under high task uncertainty. These results suggest that cognitive caching of alternative task sets adds to human cognitive flexibility under high task uncertainty. PMID:26643146
Spontaneous heparin-induced thrombocytopenia (HIT) syndrome: HIT without any heparin exposure.
Miyata, Shigeki
2016-01-01
Heparin-induced thrombocytopenia (HIT) is a pro-thrombotic side effect of heparin therapy caused by HIT antibodies with platelet-activating properties. Recent advances in understanding of spontaneous HIT syndrome, which can occur even without any heparin exposure despite its clinical and serological characteristics being similar to those of HIT, reveal the following HIT clinical features atypical for an immune-mediated disease. Heparin-naïve patients can develop IgG antibodies as early as day 4, as in a secondary immune response. Evidence for an anamnestic response upon heparin re-exposure is lacking. In addition, HIT antibodies are relatively short-lived, unlike those in a secondary immune response. Antigen immunoassays are commonly used worldwide for serological diagnosis of HIT. However, such assays do not indicate whether HIT antibodies have platelet-activating properties, leading to low diagnostic specificity for HIT. The detection of platelet-activating antibodies using a washed platelet activation assay is crucial for making a HIT diagnosis. These atypical clinical and serological features should be carefully considered while appropriately diagnosing HIT, which leads to appropriate therapy such as immediate administration of an alternative anticoagulant for preventing thromboembolic events and re-administration of heparin during surgery involving cardiopulmonary bypass when HIT antibodies are no longer detectable.
Beiderlinden, Martin; Werner, Patrick; Bahlmann, Astrid; Kemper, Johann; Brezina, Tobias; Schäfer, Maximilian; Görlinger, Klaus; Seidel, Holger; Kienbaum, Peter; Treschan, Tanja A
2018-02-09
Argatroban or lepirudin anticoagulation therapy in patients with heparin induced thrombocytopenia (HIT) or HIT suspect is typically monitored using the activated partial thromboplastin time (aPTT). Although aPTT correlates well with plasma levels of argatroban and lepirudin in healthy volunteers, it might not be the method of choice in critically ill patients. However, in-vivo data is lacking for this patient population. Therefore, we studied in vivo whether ROTEM or global clotting times would provide an alternative for monitoring the anticoagulant intensity effects in critically ill patients. This study was part of the double-blind randomized trial "Argatroban versus Lepirudin in critically ill patients (ALicia)", which compared critically ill patients treated with argatroban or lepirudin. Following institutional review board approval and written informed consent, for this sub-study blood of 35 critically ill patients was analysed. Before as well as 12, 24, 48 and 72 h after initiation of argatroban or lepirudin infusion, blood was analysed for aPTT, aPTT ratios, thrombin time (TT), INTEM CT,INTEM CT ratios, EXTEM CT, EXTEM CT ratios and maximum clot firmness (MCF) and correlated with the corresponding plasma concentrations of the direct thrombin inhibitor. To reach a target aPTT of 1.5 to 2 times baseline, median [IQR] plasma concentrations of 0.35 [0.01-1.2] μg/ml argatroban and 0.17 [0.1-0.32] μg/ml lepirudin were required. For both drugs, there was no significant correlation between aPTT and aPTT ratios and plasma concentrations. INTEM CT, INTEM CT ratios, EXTEM CT, EXTEM CT ratios, TT and TT ratios correlated significantly with plasma concentrations of both drugs. Additionally, agreement between argatroban plasma levels and EXTEM CT and EXTEM CT ratios were superior to agreement between argatroban plasma levels and aPTT in the Bland Altman analysis. MCF remained unchanged during therapy with both drugs. In critically ill patients, TT and ROTEM parameters may provide better correlation to argatroban and lepirudin plasma concentrations than aPTT. ClinicalTrials.gov , NCT00798525 , registered on 25 Nov 2008.
Minet, V; Bailly, N; Douxfils, J; Osselaer, J C; Laloy, J; Chatelain, C; Elalamy, I; Chatelain, B; Dogné, J M; Mullier, F
2013-09-01
Early diagnosis of immune heparin-induced thrombocytopenia (HIT) is challenging. HemosIL® AcuStar HIT and heparin-induced multiple electrode aggregometry (HIMEA) were recently proposed as rapid diagnostic methods. We conducted a study to assess performances of AcuStar HIT-IgG (PF4-H) and AcuStar HIT-Ab (PF4-H). The secondary objective was to compare the performances of the combination of Acustar HIT and HIMEA with standardised clinical diagnosis. Sera of 104 suspected HIT patients were retrospectively tested with AcuStar HIT. HIMEA was performed on available sera (n=81). The clinical diagnosis was established by analysing in a standardized manner the patient's medical records. These tests were also compared with PF4-Enhanced®, LTA, and SRA in subsets of patients. Thresholds were determined using ROC curve analysis with clinical outcome as reference. Using the recommended thresholds (1.00AU), the negative predictive value (NPV) of HIT-IgG and HIT-Ab were 100.0% (95% CI: 95.9%-100.0% and 95.7%-100.0%). The positive predictive value (PPV) were 64.3% (95% CI: 35.1%-87.2.2%) and 45.0% (95% CI: 23.2%-68.6%), respectively. Using our thresholds (HIT-IgG: 2.89AU, HIT-Ab: 9.41AU), NPV of HIT-IgG and HIT-Ab were 100.0% (95% CI: 96.0%-100.0% and 96.1%-100.0%). PPV were 75.0% (95% CI: 42.7%-94.5%) and 81.8% (95% CI: 48.3%-97.7%), respectively. Of the 79 patients with a medium-high pretest probability score, 67 were negative using HIT-IgG (PF4-H) test at our thresholds. HIMEA was performed on HIT-IgG positive patients. Using this combination, only one patient on 79 was incorrectly diagnosed. Acustar HIT showed good performances to exclude the diagnosis of HIT. Combination with HIMEA improves PPV. Copyright © 2013 Elsevier Ltd. All rights reserved.
Heparin-induced thrombocytopenia: real-world issues.
Linkins, Lori-Ann; Warkentin, Theodore E
2011-09-01
Heparin-induced thrombocytopenia (HIT) is a prothrombotic drug reaction caused by platelet-activating antibodies. HIT sera often activate platelets without needing heparin-such heparin-"independent" platelet activation can be associated with HIT beginning or worsening despite stopping heparin ("delayed-onset HIT"). We address important issues in HIT diagnosis and therapy, using a recent cohort of HIT patients to illustrate influences of heparin type; triggers for HIT investigation; serological features of heparin-independent platelet activation; and treatment. In our cohort of recent HIT cases ( N = 13), low-molecular-weight heparin (dalteparin) was a common causative agent ( N = 8, 62%); most patients were diagnosed after HIT-thrombosis had occurred; and danaparoid was the most frequently selected treatment. Heparin-independent platelet activation was common (7/13 [54%]) and predicted slower platelet count recovery (>1 week) among evaluable patients (5/5 vs 1/6; P = 0.015). In our experience with argatroban-treated patients, HIT-associated consumptive coagulopathy confounds anticoagulant monitoring. Our observations provide guidance on practical aspects of HIT diagnosis and management. Thieme Medical Publishers.
Zhu, Tian; Cao, Shuyi; Su, Pin-Chih; Patel, Ram; Shah, Darshan; Chokshi, Heta B; Szukala, Richard; Johnson, Michael E; Hevener, Kirk E
2013-09-12
A critical analysis of virtual screening results published between 2007 and 2011 was performed. The activity of reported hit compounds from over 400 studies was compared to their hit identification criteria. Hit rates and ligand efficiencies were calculated to assist in these analyses, and the results were compared with factors such as the size of the virtual library and the number of compounds tested. A series of promiscuity, druglike, and ADMET filters were applied to the reported hits to assess the quality of compounds reported, and a careful analysis of a subset of the studies that presented hit optimization was performed. These data allowed us to make several practical recommendations with respect to selection of compounds for experimental testing, definition of hit identification criteria, and general virtual screening hit criteria to allow for realistic hit optimization. A key recommendation is the use of size-targeted ligand efficiency values as hit identification criteria.
Schneider, Uwe; Vasi, Fabiano; Besserer, Jürgen
2016-01-01
When fractionation schemes for hypofractionation and stereotactic body radiotherapy are considered, a reliable cell survival model at high dose is needed for calculating doses of similar biological effectiveness. An alternative to the LQ-model is the track-event theory which is based on the probabilities for one- and two two-track events. A one-track-event (OTE) is always represented by at least two simultaneous double strand breaks. A two-track-event (TTE) results in one double strand break. Therefore at least two two-track-events on the same or different chromosomes are necessary to produce an event which leads to cell sterilization. It is obvious that the probabilities of OTEs and TTEs must somehow depend on the geometrical structure of the chromatin. In terms of the track-event theory the ratio ε of the probabilities of OTEs and TTEs includes the geometrical dependence and is obtained in this work by simple Monte Carlo simulations. For this work it was assumed that the anchors of loop forming chromatin are most sensitive to radiation induced cell deaths. Therefore two adjacent tetranucleosomes representing the loop anchors were digitized. The probability ratio ε of OTEs and TTEs was factorized into a radiation quality dependent part and a geometrical part: ε = εion ∙ εgeo. εgeo was obtained for two situations, by applying Monte Carlo simulation for DNA on the tetranucleosomes itself and for linker DNA. Low energy electrons were represented by randomly distributed ionizations and high energy electrons by ionizations which were simulated on rays. εion was determined for electrons by using results from nanodosimetric measurements. The calculated ε was compared to the ε obtained from fits of the track event model to 42 sets of experimental human cell survival data. When the two tetranucleosomes are in direct contact and the hits are randomly distributed εgeo and ε are 0.12 and 0.85, respectively. When the hits are simulated on rays εgeo and ε are 0.10 and 0.71. For the linker-DNA εgeo and ε for randomly distributed hits are 0.010 and 0.073, and for hits on rays 0.0058 and 0.041, respectively. The calculated ε fits the experimentally obtained ε = 0.64±0.32 best for hits on the tetranucleosome when they are close to each other both, for high and low energy electrons. The parameter εgeo of the track event model was obtained by pure geometrical considerations of the chromatin structure and is 0.095 ± 0.022. It can be used as a fixed parameter in the track-event theory.
Ridge, Perry G; Maxwell, Taylor J; Foutz, Spencer J; Bailey, Matthew H; Corcoran, Christopher D; Tschanz, JoAnn T; Norton, Maria C; Munger, Ronald G; O'Brien, Elizabeth; Kerber, Richard A; Cawthon, Richard M; Kauwe, John S K
2014-01-01
The mitochondria are essential organelles and are the location of cellular respiration, which is responsible for the majority of ATP production. Each cell contains multiple mitochondria, and each mitochondrion contains multiple copies of its own circular genome. The ratio of mitochondrial genomes to nuclear genomes is referred to as mitochondrial copy number. Decreases in mitochondrial copy number are known to occur in many tissues as people age, and in certain diseases. The regulation of mitochondrial copy number by nuclear genes has been studied extensively. While mitochondrial variation has been associated with longevity and some of the diseases known to have reduced mitochondrial copy number, the role that the mitochondrial genome itself has in regulating mitochondrial copy number remains poorly understood. We analyzed the complete mitochondrial genomes from 1007 individuals randomly selected from the Cache County Study on Memory Health and Aging utilizing the inferred evolutionary history of the mitochondrial haplotypes present in our dataset to identify sequence variation and mitochondrial haplotypes associated with changes in mitochondrial copy number. Three variants belonging to mitochondrial haplogroups U5A1 and T2 were significantly associated with higher mitochondrial copy number in our dataset. We identified three variants associated with higher mitochondrial copy number and suggest several hypotheses for how these variants influence mitochondrial copy number by interacting with known regulators of mitochondrial copy number. Our results are the first to report sequence variation in the mitochondrial genome that causes changes in mitochondrial copy number. The identification of these variants that increase mtDNA copy number has important implications in understanding the pathological processes that underlie these phenotypes.
Camp, Christopher L; Wang, Dean; Sinatro, Alec S; D'Angelo, John; Coleman, Struan H; Dines, Joshua S; Fealy, Stephen; Conte, Stan
2018-05-01
Although batters are frequently hit by pitch (HBP) in baseball, the effect of HBP injuries remains undefined in the literature. To determine the effect of HBP injuries in terms of time out of play, injury patterns resulting in the greatest time out of play, and the value of protective gear such as helmets and elbow pads. Descriptive laboratory study. Based on the Major League Baseball (MLB) Health and Injury Tracking System, all injuries to batters HBP during the 2011-2015 MLB and Minor League Baseball (MiLB) seasons were identified and analyzed. Video analysis was performed on all HBP events from the 2015 MLB season. Multivariate stepwise regression analysis was utilized to determine the predictive capacity of multiple variables (velocity, pitch type, location, etc) on injury status and severity. A total of 2920 HBP injuries resulted in 24,624 days missed (DM) over the 5 seasons. MLB HBP injuries occurred at a rate of 1 per 2554 plate appearances (1 per 9780 pitches thrown). Mean DM per injury were 8.4 (11.7 for MLB vs 8.0 for MiLB, P < .001). Surgery was required for 3.1% of MLB injuries and 1.2% of MiLB injuries ( P = .005). The most common body regions injured were the hand/fingers (n = 638, 21.8%), head/face (n = 497, 17.0%), and elbow (n = 440, 15.7%), and there were 146 (5.0%) concussions. Injury rates and mean DM correlated with velocity in a near linear fashion. Players hit in the head/face (odds ratio, 28.7) or distal upper extremity (odds ratio, 6.4) were more likely to be injured than players HBP in other locations. Players with an unprotected elbow missed 1.7 more days (95% CI, -4.1 to 7.6) than those with an elbow protector ( P = .554) when injured after being HBP. Although HBP injuries occur infrequently in the course of normal play, they collectively represent a significant source of time out of play. The most common body regions injured include the hands/fingers and head/face, and batters hit in these locations are significantly more likely to be injured. After contusions, concussions were the most common injury diagnosis.
Multiple core computer processor with globally-accessible local memories
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shalf, John; Donofrio, David; Oliker, Leonid
A multi-core computer processor including a plurality of processor cores interconnected in a Network-on-Chip (NoC) architecture, a plurality of caches, each of the plurality of caches being associated with one and only one of the plurality of processor cores, and a plurality of memories, each of the plurality of memories being associated with a different set of at least one of the plurality of processor cores and each of the plurality of memories being configured to be visible in a global memory address space such that the plurality of memories are visible to two or more of the plurality ofmore » processor cores.« less
Technical Report Interchange Through Synchronized OAI Caches
NASA Technical Reports Server (NTRS)
Liu, Xiaming; Maly, Kurt; Zubair, Mohammad; Tang, Rong; Padshah, Mohammad Imran; Roncaglia, George; Rocker, JoAnne; Nelson, Michael; vonOfenheim, William; Luce, Richard
2002-01-01
The Technical Report Interchange project is a cooperative experimental effort between NASA Langley Research Center, Los Alamos National Laboratory, Air Force Research Laboratory, Sandia National Laboratory and Old Dominion University to allow for the integration of technical reports. This is accomplished using the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH) and having each site cache the metadata from the other participating sites. Each site also implements additional software to ingest the OAI-PMH harvested metadata into their native digital library (DL). This allows the users at each site to see an increased technical report collection through the familiar DL interfaces and tale advantage of whatever valued added are provided by the native DL.
Flood of December 1987 in central and eastern Arkansas
Neely, B.L.
1990-01-01
Heavy rain fell across Arkansas during December 24-28, 1987. During this period, 6 to 12 inches of rain fell in a 100-mile wide belt extending roughly from Texarkana to West Memphis, Arkansas. The intense rainfall produced flooding throughout much of central and eastern Arkansas. Peak discharges associated with the flood had recurrence intervals of 100 years at two gaging stations. Peak stages, discharges, and recurrence intervals for this flood are documented profiles for 41 gaging stations. Also included in this report are flood profiles for Bayou Meto, Bayou Two Prairie, Cache River, L'Anguille River, and flood hydrographs for gaging stations on Bayou Meto near Lonoke and Cache River at Patterson. (USGS)
NASA Astrophysics Data System (ADS)
Johnson, W. N.; Herrick, W. V.; Grundmann, W. J.
1984-10-01
For the first time, VLSI technology is used to compress the full functinality and comparable performance of the VAX 11/780 super-minicomputer into a 1.2 M transistor microprocessor chip set. There was no subsetting of the 304 instruction set and the 17 data types, nor reduction in hardware support for the 4 Gbyte virtual memory management architecture. The chipset supports an integral 8 kbyte memory cache, a 13.3 Mbyte/s system bus, and sophisticated multiprocessing. High performance is achieved through microcode optimizations afforded by the large control store, tightly coupled address and data caches, the use of internal and external 32 bit datapaths, the extensive aplication of both microlevel and macrolevel pipelining, and the use of specialized hardware assists.
Milanović, Zoran; Sporiš, Goran; Weston, Matthew
2015-10-01
Enhancing cardiovascular fitness can lead to substantial health benefits. High-intensity interval training (HIT) is an efficient way to develop cardiovascular fitness, yet comparisons between this type of training and traditional endurance training are equivocal. Our objective was to meta-analyse the effects of endurance training and HIT on the maximal oxygen consumption (VO2max) of healthy, young to middle-aged adults. Six electronic databases were searched (MEDLINE, PubMed, SPORTDiscus, Web of Science, CINAHL and Google Scholar) for original research articles. A search was conducted and search terms included 'high intensity', 'HIT', 'sprint interval training', 'endurance training', 'peak oxygen uptake', and 'VO2max'. Inclusion criteria were controlled trials, healthy adults aged 18-45 years, training duration ≥2 weeks, VO2max assessed pre- and post-training. Twenty-eight studies met the inclusion criteria and were included in the meta-analysis. This resulted in 723 participants with a mean ± standard deviation (SD) age and initial fitness of 25.1 ± 5 years and 40.8 ± 7.9 mL·kg(-1)·min(-1), respectively. We made probabilistic magnitude-based inferences for meta-analysed effects based on standardised thresholds for small, moderate and large changes (0.2, 0.6 and 1.2, respectively) derived from between-subject SDs for baseline VO2max. The meta-analysed effect of endurance training on VO2max was a possibly large beneficial effect (4.9 mL·kg(-1)·min(-1); 95 % confidence limits ±1.4 mL·kg(-1)·min(-1)), when compared with no-exercise controls. A possibly moderate additional increase was observed for typically younger subjects (2.4 mL·kg(-1)·min(-1); ±2.1 mL·kg(-1)·min(-1)) and interventions of longer duration (2.2 mL·kg(-1)·min(-1); ±3.0 mL·kg(-1)·min(-1)), and a small additional improvement for subjects with lower baseline fitness (1.4 mL·kg(-1)·min(-1); ±2.0 mL·kg(-1)·min(-1)). When compared with no-exercise controls, there was likely a large beneficial effect of HIT (5.5 mL·kg(-1)·min(-1); ±1.2 mL·kg(-1)·min(-1)), with a likely moderate greater additional increase for subjects with lower baseline fitness (3.2 mL·kg(-1)·min(-1); ±1.9 mL·kg(-1)·min(-1)) and interventions of longer duration (3.0 mL·kg(-1)·min(-1); ±1.9 mL·kg(-1)·min(-1)), and a small lesser effect for typically longer HIT repetitions (-1.8 mL·kg(-1)·min(-1); ±2.7 mL·kg(-1)·min(-1)). The modifying effects of age (0.8 mL·kg(-1)·min(-1); ±2.1 mL·kg(-1)·min(-1)) and work/rest ratio (0.5 mL·kg(-1)·min(-1); ±1.6 mL·kg(-1)·min(-1)) were unclear. When compared with endurance training, there was a possibly small beneficial effect for HIT (1.2 mL·kg(-1)·min(-1); ±0.9 mL·kg(-1)·min(-1)) with small additional improvements for typically longer HIT repetitions (2.2 mL·kg(-1)·min(-1); ±2.1 mL·kg(-1)·min(-1)), older subjects (1.8 mL·kg(-1)·min(-1); ±1.7 mL·kg(-1)·min(-1)), interventions of longer duration (1.7 mL·kg(-1)·min(-1); ±1.7 mL·kg(-1)·min(-1)), greater work/rest ratio (1.6 mL·kg(-1)·min(-1); ±1.5 mL·kg(-1)·min(-1)) and lower baseline fitness (0.8 mL·kg(-1)·min(-1); ±1.3 mL·kg(-1)·min(-1)). Endurance training and HIT both elicit large improvements in the VO2max of healthy, young to middle-aged adults, with the gains in VO2max being greater following HIT when compared with endurance training.
Hanssen, H; Minghetti, A; Magon, S; Rossmeissl, A; Rasenack, M; Papadopoulou, A; Klenk, C; Faude, O; Zahner, L; Sprenger, T; Donath, L
2018-03-01
Aerobic exercise training is a promising complementary treatment option in migraine and can reduce migraine days and improve retinal microvascular function. Our aim was to elucidate whether different aerobic exercise programs at high vs moderate intensities distinctly affect migraine days as primary outcome and retinal vessel parameters as a secondary. In this randomized controlled trial, migraine days were recorded by a validated migraine diary in 45 migraineurs of which 36 (female: 28; age: 36 (SD:10)/BMI: 23.1 (5.3) completed the training period (dropout: 20%). Participants were assigned (Strata: age, gender, fitness and migraine symptomatology) to either high intensity interval training (HIT), moderate continuous training (MCT), or a control group (CON). Intervention groups trained twice a week over a 12-week intervention period. Static retinal vessel analysis, central retinal arteriolar (CRAE) and venular (CRVE) diameters, as well as the arteriolar-to-venular diameter ratio (AVR) were obtained for cerebrovascular health assessment. Incremental treadmill testing yielded maximal and submaximal fitness parameters. Overall, moderate migraine day reductions were observed (ηP2 = .12): HIT revealed 89% likely beneficial effects (SMD = 1.05) compared to MCT (SMD = 0.50) and CON (SMD = 0.59). Very large intervention effects on AVR improvement (ηP2 = 0.27), slightly favoring HIT (SMD=-0.43) over CON (SMD=0), were observed. HIT seems more effective for migraine day reduction and improvement of cerebrovascular health compared to MCT. Intermittent exercise programs of higher intensities may need to be considered as an additional treatment option in migraine patients. © 2017 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.
Zhu, Tian; Cao, Shuyi; Su, Pin-Chih; Patel, Ram; Shah, Darshan; Chokshi, Heta B.; Szukala, Richard; Johnson, Michael E.; Hevener, Kirk E.
2013-01-01
A critical analysis of virtual screening results published between 2007 and 2011 was performed. The activity of reported hit compounds from over 400 studies was compared to their hit identification criteria. Hit rates and ligand efficiencies were calculated to assist in these analyses and the results were compared with factors such as the size of the virtual library and the number of compounds tested. A series of promiscuity, drug-like, and ADMET filters were applied to the reported hits to assess the quality of compounds reported and a careful analysis of a subset of the studies which presented hit optimization was performed. This data allowed us to make several practical recommendations with respect to selection of compounds for experimental testing, defining hit identification criteria, and general virtual screening hit criteria to allow for realistic hit optimization. A key recommendation is the use of size-targeted ligand efficiency values as hit identification criteria. PMID:23688234
Minet, V; Baudar, J; Bailly, N; Douxfils, J; Laloy, J; Lessire, S; Gourdin, M; Devalet, B; Chatelain, B; Dogné, J M; Mullier, F
2014-06-01
Accurate diagnosis of heparin-induced thrombocytopenia (HIT) is essential but remains challenging. We have previously demonstrated, in a retrospective study, the usefulness of the combination of the 4Ts score, AcuStar HIT and heparin-induced multiple electrode aggregometry (HIMEA) with optimized thresholds. We aimed at exploring prospectively the performances of our optimized diagnostic algorithm on suspected HIT patients. The secondary objective is to evaluate performances of AcuStar HIT-Ab (PF4-H) in comparison with the clinical outcome. 116 inpatients with clinically suspected immune HIT were included. Our optimized diagnostic algorithm was applied to each patient. Sensitivity, specificity, negative predictive value (NPV), positive predictive value (PPV) of the overall diagnostic strategy as well as AcuStar HIT-Ab (at manufacturer's thresholds and at our thresholds) were calculated using clinical diagnosis as the reference. Among 116 patients, 2 patients had clinically-diagnosed HIT. These 2 patients were positive on AcuStar HIT-Ab, AcuStar HIT-IgG and HIMEA. Using our optimized algorithm, all patients were correctly diagnosed. AcuStar HIT-Ab at our cut-off (>9.41 U/mL) and at manufacturer's cut-off (>1.00 U/mL) showed both a sensitivity of 100.0% and a specificity of 99.1% and 90.4%, respectively. The combination of the 4Ts score, the HemosIL® AcuStar HIT and HIMEA with optimized thresholds may be useful for the rapid and accurate exclusion of the diagnosis of immune HIT. Copyright © 2014 Elsevier Ltd. All rights reserved.
McMahon, C M; Tanhehco, Y C; Cuker, A
2017-02-01
Essentials Misdiagnosis of heparin-induced thrombocytopenia (HIT) may be associated with adverse outcomes. We conducted a study of patients with a heparin allergy in the chart due to misdiagnosis of HIT. 42% of patients with a heparin allergy due to suspected HIT were clearly HIT-negative. 68% were unnecessarily treated with an alternative anticoagulant, 66% of whom had major bleeding. Background It is recommended that heparin be added to the allergy list of patients with heparin-induced thrombocytopenia (HIT). Misdiagnosis of HIT could lead to inappropriate documentation of a heparin allergy and adverse outcomes. Objectives To determine the frequency and consequences of inappropriate documentation of a heparin allergy because of misdiagnosis of HIT. Methods We conducted a cohort study of patients with an inappropriate heparin allergy listed in the electronic medical record (EMR) because of misdiagnosis of HIT. We searched the EMR for patients with a new heparin allergy. Patients were eligible if the reason for allergy listing was suspected acute HIT and laboratory testing for HIT was performed within 60 days. Subjects were defined as 'HIT-negative' if they had a 4Ts score of ≤ 3 or negative laboratory test results. Results Of 239 subjects with a new heparin allergy documented because of concern regarding HIT, 100 (42%) met the prespecified definition of HIT-negative. Sixty-eight (68%) HIT-negative subjects unnecessarily received an alternative parenteral anticoagulant for a median duration of 10.5 days. Among these 68 patients, 45 (66%) met criteria for major bleeding. Sixty-eight (68%) of the 100 HIT-negative subjects had an inappropriate allergy to heparin documented that persisted in the EMR for > 3 years beyond the index hospitalization. Conclusions Inappropriate listing of heparin as an allergy in the EMR because of misdiagnosis of HIT is common, is associated with substantial rates of unnecessary alternative anticoagulant use and major bleeding, and tends to persist beyond the index admission. © 2016 International Society on Thrombosis and Haemostasis.
Historical Channel Changes in Cache Creek, Capay Valley, California
NASA Astrophysics Data System (ADS)
Higgins, S. A.; Kamman, G. R.
2009-12-01
Historical channel changes were assessed for the 21-mile segment of Cache Creek through Capay Valley in order to evaluate temporal changes in stream channel morphology. The Capay Valley segment of Cache Creek is primarily a low-gradient channel with a gravel/cobble substrate. Hydrologic conditions have been affected by dam operations that store runoff during the wet season and deliver water during the dry season for downstream irrigation uses. Widespread distribution of invasive plant species has altered the condition of the riparian corridor. The assessment evaluated a hypothesis that historical changes in hydrology and vegetation cover have triggered changes in geomorphic conditions. Historic channel alignments were digitized to assess planform channel adjustments. Results illustrate a dynamic system with frequent channel movements throughout the historic period. Evaluation of longitudinal channel adjustments revealed a relatively stable bed surface elevation since the 1930’s. Comparisons of cross-sectional channel geometry for topographic profiles surveyed in 1984 were compared to equivalent features in a LiDAR survey from 2008. The comparisons show a relatively consistent channel geometry that has maintained a similar form despite rather large planform adjustments with areas of bank retreat in excess of 500 feet. Results suggest that the study reach has maintained a relatively stable morphology through a series of dynamic planform adjustments during the historic period.
Erlandson, Jon M.; Thomas-Barnett, Lisa; Vellanoweth, René L.; Schwartz, Steven J.; Muhs, Daniel R.
2013-01-01
A cache feature salvaged from an eroding sea cliff on San Nicolas Island produced two redwood boxes containing more than 200 artifacts of Nicoleño, Native Alaskan, and Euro-American origin. Outside the boxes were four asphaltum-coated baskets, abalone shells, a sandstone dish, and a hafted stone knife. The boxes, made from split redwood planks, contained a variety of artifacts and numerous unmodified bones and teeth from marine mammals, fish, birds, and large land mammals. Nicoleño-style artifacts include 11 knives with redwood handles and stone blades, stone projectile points, steatite ornaments and effigies, a carved stone pipe, abraders and burnishing stones, bird bone whistles, bone and shell pendants, abalone shell dishes, and two unusual barbed shell fishhooks. Artifacts of Native Alaskan style include four bone toggling harpoons, two unilaterally barbed bone harpoon heads, bone harpoon fore-shafts, a ground slate blade, and an adze blade. Objects of Euro-American origin or materials include a brass button, metal harpoon blades, and ten flaked glass bifaces. The contents of the cache feature, dating to the early-to-mid nineteenth century, provide an extraordinary window on a time of European expansion and global economic development that created unique cultural interactions and social transformations.
Brodin, Anders; Urhan, A Utku
2013-07-01
Laboratory studies of scatter hoarding birds have become a model system for spatial memory studies. Considering that such birds are known to have a good spatial memory, recovery success in lab studies seems low. In parids (titmice and chickadees) typically ranging between 25 and 60% if five seeds are cached in 50-128 available caching sites. Since these birds store many thousands of food items in nature in one autumn one might expect that they should easily retrieve five seeds in a laboratory where they know the environment with its caching sites in detail. We designed a laboratory set up to be as similar as possible with previous studies and trained wild caught marsh tits Poecile palustris to store and retrieve in this set up. Our results agree closely with earlier studies, of the first ten looks around 40% were correct when the birds had stored five seeds in 100 available sites both 5 and 24h after storing. The cumulative success curve suggests high success during the first 15 looks where after it declines. Humans performed much better, in the first five looks most subjects were 100% correct. We discuss possible reasons for why the birds were not doing better. Copyright © 2013 Elsevier B.V. All rights reserved.
The effect of environmental harshness on neurogenesis: a large-scale comparison.
Chancellor, Leia V; Roth, Timothy C; LaDage, Lara D; Pravosudov, Vladimir V
2011-03-01
Harsh environmental conditions may produce strong selection pressure on traits, such as memory, that may enhance fitness. Enhanced memory may be crucial for survival in animals that use memory to find food and, thus, particularly important in environments where food sources may be unpredictable. For example, animals that cache and later retrieve their food may exhibit enhanced spatial memory in harsh environments compared with those in mild environments. One way that selection may enhance memory is via the hippocampus, a brain region involved in spatial memory. In a previous study, we established a positive relationship between environmental severity and hippocampal morphology in food-caching black-capped chickadees (Poecile atricapillus). Here, we expanded upon this previous work to investigate the relationship between environmental harshness and neurogenesis, a process that may support hippocampal cytoarchitecture. We report a significant and positive relationship between the degree of environmental harshness across several populations over a large geographic area and (1) the total number of immature hippocampal neurons, (2) the number of immature neurons relative to the hippocampal volume, and (3) the number of immature neurons relative to the total number of hippocampal neurons. Our results suggest that hippocampal neurogenesis may play an important role in environments where increased reliance on memory for cache recovery is critical. Copyright © 2010 Wiley Periodicals, Inc.
Distributed shared memory for roaming large volumes.
Castanié, Laurent; Mion, Christophe; Cavin, Xavier; Lévy, Bruno
2006-01-01
We present a cluster-based volume rendering system for roaming very large volumes. This system allows to move a gigabyte-sized probe inside a total volume of several tens or hundreds of gigabytes in real-time. While the size of the probe is limited by the total amount of texture memory on the cluster, the size of the total data set has no theoretical limit. The cluster is used as a distributed graphics processing unit that both aggregates graphics power and graphics memory. A hardware-accelerated volume renderer runs in parallel on the cluster nodes and the final image compositing is implemented using a pipelined sort-last rendering algorithm. Meanwhile, volume bricking and volume paging allow efficient data caching. On each rendering node, a distributed hierarchical cache system implements a global software-based distributed shared memory on the cluster. In case of a cache miss, this system first checks page residency on the other cluster nodes instead of directly accessing local disks. Using two Gigabit Ethernet network interfaces per node, we accelerate data fetching by a factor of 4 compared to directly accessing local disks. The system also implements asynchronous disk access and texture loading, which makes it possible to overlap data loading, volume slicing and rendering for optimal volume roaming.
77 FR 32639 - HIT Standards Committee and HIT Policy Committee; Call for Nominations
Federal Register 2010, 2011, 2012, 2013, 2014
2012-06-01
... with the implementation of the Federal Health IT Strategic Plan, and in accordance with policies... DEPARTMENT OF HEALTH AND HUMAN SERVICES HIT Standards Committee and HIT Policy Committee; Call for... Health Information Technology Policy Committee (HITPC). Name of Committees: HIT Standards Committee and...
2007-01-01
parameter dimension between the two models). 93 were tested.3 Model 1 log( pHits 1− pHits ) = α + β1 ∗ MetricScore (6.6) The results for each of the...505.67 oTERavg .357 .13 .007 log( pHits 1− pHits ), that is, log-odds of correct task performance, of 2.79 over the intercept only model. All... pHits 1− pHits ) = −1.15− .418× I[MT=2] − .527× I[MT=3] + 1.78×METEOR+ 1.28×METEOR × I[MT=2] + 1.86×METEOR × I[MT=3] (6.7) Model 3 log( pHits 1− pHits
Review of Social and Organizational Issues in Health Information Technology.
Kuziemsky, Craig E
2015-07-01
This paper reviews organizational and social issues (OSIs) in health information technology (HIT). A review and synthesis of the literature on OSIs in HIT was conducted. Five overarching themes with respect to OSIs in HIT were identified and discussed: scope and frameworks for defining OSIs in HIT, context matters, process immaturity and complexity, trade-offs will happen and need to be discussed openly, and means of studying OSIs in HIT. There is a wide body of literature that provides insight into OSIs in HIT, even if many of the studies are not explicitly labelled as such. The two biggest research needs are more explicit and theoretical studies of OSI in HITs and more research on integrating micro and macro perspectives of HIT use in organizations.
Kang, Hong; Wang, Frank; Zhou, Sicheng; Miao, Qi; Gong, Yang
2017-01-01
Health information technology (HIT) events, a subtype of patient safety events, pose a major threat and barrier toward a safer healthcare system. It is crucial to gain a better understanding of the nature of the errors and adverse events caused by current HIT systems. The scarcity of HIT event-exclusive databases and event reporting systems indicates the challenge of identifying the HIT events from existing resources. FDA Manufacturer and User Facility Device Experience (MAUDE) database is a potential resource for HIT events. However, the low proportion and the rapid evolvement of HIT-related events present challenges for distinguishing them from other equipment failures and hazards. We proposed a strategy to identify and synchronize HIT events from MAUDE by using a filter based on structured features and classifiers based on unstructured features. The strategy will help us develop and grow an HIT event-exclusive database, keeping pace with updates to MAUDE toward shared learning.
Multiple-hit parameter estimation in monolithic detectors.
Hunter, William C J; Barrett, Harrison H; Lewellen, Tom K; Miyaoka, Robert S
2013-02-01
We examine a maximum-a-posteriori method for estimating the primary interaction position of gamma rays with multiple interaction sites (hits) in a monolithic detector. In assessing the performance of a multiple-hit estimator over that of a conventional one-hit estimator, we consider a few different detector and readout configurations of a 50-mm-wide square cerium-doped lutetium oxyorthosilicate block. For this study, we use simulated data from SCOUT, a Monte-Carlo tool for photon tracking and modeling scintillation- camera output. With this tool, we determine estimate bias and variance for a multiple-hit estimator and compare these with similar metrics for a one-hit maximum-likelihood estimator, which assumes full energy deposition in one hit. We also examine the effect of event filtering on these metrics; for this purpose, we use a likelihood threshold to reject signals that are not likely to have been produced under the assumed likelihood model. Depending on detector design, we observe a 1%-12% improvement of intrinsic resolution for a 1-or-2-hit estimator as compared with a 1-hit estimator. We also observe improved differentiation of photopeak events using a 1-or-2-hit estimator as compared with the 1-hit estimator; more than 6% of photopeak events that were rejected by likelihood filtering for the 1-hit estimator were accurately identified as photopeak events and positioned without loss of resolution by a 1-or-2-hit estimator; for PET, this equates to at least a 12% improvement in coincidence-detection efficiency with likelihood filtering applied.
Health information technology and dynamic capabilities.
Leung, Ricky C
2012-01-01
Health information technology (HIT) purports to increase quality and efficiency in health care organizations. However, health care organizations are situated in constantly changing environments. They need dynamic capabilities to implement HIT effectively. This article builds on the dynamic capabilities perspective and generates propositions about implementing HIT in dynamic environments. Specifically, I identify the (1) the necessary resources and capabilities for organizations to implement HIT; (2) the organizational capabilities and benefits that can be enhanced by HIT; and (3) the similarities and differences between three distinct forms of HIT. I synthesized the literature on dynamic capabilities and HIT to identify dynamic capabilities that are associated with (1) electronic medical records, (2) telemedicine, and (3) social media. In addition, I discuss the benefits of these HITs for improving the dynamic capabilities of health care organizations. PROPOSITIONS/FINDINGS: This article generates three sets of propositions that can be tested empirically. First, I am concerned with how organizational size and human resources affect successful implementation of HIT. In addition, I argue that three technology-specific factors--hospital type, medical specialty, and socially desirable technical features--may affect the implementation of HIT. To cope with constantly changing environmental pressures, health administrators need to deploy, modify, and/or acquire organizational resources skillfully. Practitioners need to identify dynamic capabilities to support specific forms of HIT and understand how HIT enables health care organizations in turn. The concept of evolutionary fitness in the dynamic capabilities perspective may be developed to measure HIT implementation.
Lyon, Aaron R; Lewis, Cara C; Melvin, Abigail; Boyd, Meredith; Nicodimos, Semret; Liu, Freda F; Jungbluth, Nathaniel
2016-09-22
Health information technologies (HIT) have become nearly ubiquitous in the contemporary healthcare landscape, but information about HIT development, functionality, and implementation readiness is frequently siloed. Theory-driven methods of compiling, evaluating, and integrating information from the academic and commercial sectors are necessary to guide stakeholder decision-making surrounding HIT adoption and to develop pragmatic HIT research agendas. This article presents the Health Information Technologies-Academic and Commercial Evaluation (HIT-ACE) methodology, a structured, theory-driven method for compiling and evaluating information from multiple sectors. As an example demonstration of the methodology, we apply HIT-ACE to mental and behavioral health measurement feedback systems (MFS). MFS are a specific class of HIT that support the implementation of routine outcome monitoring, an evidence-based practice. HIT-ACE is guided by theories and frameworks related to user-centered design and implementation science. The methodology involves four phases: (1) coding academic and commercial materials, (2) developer/purveyor interviews, (3) linking putative implementation mechanisms to hit capabilities, and (4) experimental testing of capabilities and mechanisms. In the current demonstration, phase 1 included a systematic process to identify MFS in mental and behavioral health using academic literature and commercial websites. Using user-centered design, implementation science, and feedback frameworks, the HIT-ACE coding system was developed, piloted, and used to review each identified system for the presence of 38 capabilities and 18 additional characteristics via a consensus coding process. Bibliometic data were also collected to examine the representation of the systems in the scientific literature. As an example, results are presented for the application of HIT-ACE phase 1 to MFS wherein 49 separate MFS were identified, reflecting a diverse array of characteristics and capabilities. Preliminary findings demonstrate the utility of HIT-ACE to represent the scope and diversity of a given class of HIT beyond what can be identified in the academic literature. Phase 2 data collection is expected to confirm and expand the information presented and phases 3 and 4 will provide more nuanced information about the impact of specific HIT capabilities. In all, HIT-ACE is expected to support adoption decisions and additional HIT development and implementation research.
On the relative contributions of wind vs. animals to seed dispersal of four Sierra Nevada pines.
Vander Wall, Stephen B
2008-07-01
Selective pressures that influence the form of seed dispersal syndromes are poorly understood. Morphology of plant propagules is often used to infer the means of dispersal, but morphology can be misleading. Several species of pines, for example, have winged seeds adapted for wind dispersal but owe much of their establishment to scatter-hoarding animals. Here the relative importance of wind vs. animal dispersal is assessed for four species of pines of the eastern Sierra Nevada that have winged seeds but differed in seed size: lodgepole pine (Pinus contorta murrayana, 8 mg); ponderosa pine (Pinus ponderosa ponderosa, 56 mg); Jeffrey pine (Pinus jeffreyi, 160 mg); and sugar pine (Pinus lambertiana, 231 mg). Pre-dispersal seed mortality eliminated much of the ponderosa pine seed crop (66%), but had much less effect on Jeffrey pine (32% of seeds destroyed), lodgepole pine (29%), and sugar pine (7%). When cones opened most filled seeds were dispersed by wind. Animals removed > 99% of wind-dispersed Jeffrey and sugar pine seeds from the ground within 60 days, but animals gathered only 93% of lodgepole pine seeds and 38% of ponderosa pine seeds during the same period. Animals gathered and scatter hoarded radioactively labeled ponderosa, Jeffrey, and sugar pine seeds, making a total of 2103 caches over three years of study. Only three lodgepole pine caches were found. Caches typically contained 1-4 seeds buried 5-20 mm deep, depths suitable for seedling emergence. Although Jeffrey and sugar pine seeds are initially wind dispersed, nearly all seedlings arise from animal caches. Lodgepole pine is almost exclusively wind dispersed, with animals acting as seed predators. Animals treated ponderosa pine in an intermediate fashion. Two-phased dispersal of large, winged pine seeds appears adaptive; initial wind dispersal helps to minimize pre-dispersal seed mortality whereas scatter hoarding by animals places seeds in sites with a higher probability of seedling establishment.
NASA Astrophysics Data System (ADS)
Johnsson, L.; Netzer, G.
2016-10-01
Moore's law, the doubling of transistors per unit area for each CMOS technology generation, is expected to continue throughout the decade, while Dennard voltage scaling resulting in constant power per unit area stopped about a decade ago. The semiconductor industry's response to the loss of Dennard scaling and the consequent challenges in managing power distribution and dissipation has been leveled off clock rates, a die performance gain reduced from about a factor of 2.8 to 1.4 per technology generation, and multi-core processor dies with increased cache sizes. Increased caches sizes offers performance benefits for many applications as well as energy savings. Accessing data in cache is considerably more energy efficient than main memory accesses. Further, caches consume less power than a corresponding amount of functional logic. As feature sizes continue to be scaled down an increasing fraction of the die must be “underutilized” or “dark” due to power constraints. With power being a prime design constraint there is a concerted effort to find significantly more energy efficient chip architectures than dominant in servers today, with chips potentially incorporating several types of cores to cover a range of applications, or different functions in an application, as is already common for the mobile processor market. Digital Signal Processors (DSPs), largely targeting the embedded and mobile processor markets, typically have been designed for a power consumption of 10% or less of a typical x86 CPU, yet with much more than 10% of the floating-point capability of the same technology generation x86 CPUs. Thus, DSPs could potentially offer an energy efficient alternative to x86 CPUs. Here we report an assessment of the Texas Instruments TMS320C6678 DSP in regards to its energy efficiency for two common HPC benchmarks: STREAM (memory system benchmark) and HPL (CPU benchmark)
Pansing, Elizabeth R; Tomback, Diana F; Wunder, Michael B; French, Joshua P; Wagner, Aaron C
2017-11-01
Tree recruitment is a spatially structured process that may undergo change over time because of variation in postdispersal processes. We examined seed pilferage, seed germination, and seedling survival in whitebark pine to determine whether 1) microsite type alters the initial spatial pattern of seed caches, 2) higher abiotic stress (i.e. higher elevations) exacerbates spatial distribution changes, and 3) these postdispersal processes are spatially clustered. At two study areas, we created a seed distribution pattern by burying seed caches in microsite types frequently used by whitebark pine's avian seed disperser (Clark's nutcracker) in upper subalpine forest and at treeline, the latter characterized by high abiotic environmental stress. We monitored caches for two years for pilferage, germination, and seedling survival. Odds of pilferage (both study areas), germination (northern study area), and survival (southern study area) were higher at treeline relative to subalpine forest. At the southern study area, we found higher odds of 1) pilferage near rocks and trees relative to no object in subalpine forest, 2) germination near rocks relative to trees within both elevation zones, and 3) seedling survival near rocks and trees relative to no object at treeline. No microsite effects were detected at the northern study area. Findings indicated that the microsite distribution of seed caches changes with seed/seedling stage. Higher odds of seedling survival near rocks and trees were observed at treeline, suggesting abiotic stress may limit safe site availability, thereby shifting the spatial distribution toward protective microsites. Higher odds of pilferage at treeline, however, suggest rodents may limit treeline recruitment. Further, odds of pilferage were higher near rocks and trees relative to no object in subalpine forest but did not differ among microsites at treeline, suggesting pilferage can modulate the spatial structure of regeneration, a finding supported by limited clustering of postdispersal processes.
Going, going, still there: using the WebCite service to permanently archive cited web pages.
Eysenbach, Gunther; Trudel, Mathieu
2005-12-30
Scholars are increasingly citing electronic "web references" which are not preserved in libraries or full text archives. WebCite is a new standard for citing web references. To "webcite" a document involves archiving the cited Web page through www.webcitation.org and citing the WebCite permalink instead of (or in addition to) the unstable live Web page. This journal has amended its "instructions for authors" accordingly, asking authors to archive cited Web pages before submitting a manuscript. Almost 200 other journals are already using the system. We discuss the rationale for WebCite, its technology, and how scholars, editors, and publishers can benefit from the service. Citing scholars initiate an archiving process of all cited Web references, ideally before they submit a manuscript. Authors of online documents and websites which are expected to be cited by others can ensure that their work is permanently available by creating an archived copy using WebCite and providing the citation information including the WebCite link on their Web document(s). Editors should ask their authors to cache all cited Web addresses (Uniform Resource Locators, or URLs) "prospectively" before submitting their manuscripts to their journal. Editors and publishers should also instruct their copyeditors to cache cited Web material if the author has not done so already. Finally, WebCite can process publisher submitted "citing articles" (submitted for example as eXtensible Markup Language [XML] documents) to automatically archive all cited Web pages shortly before or on publication. Finally, WebCite can act as a focussed crawler, caching retrospectively references of already published articles. Copyright issues are addressed by honouring respective Internet standards (robot exclusion files, no-cache and no-archive tags). Long-term preservation is ensured by agreements with libraries and digital preservation organizations. The resulting WebCite Index may also have applications for research assessment exercises, being able to measure the impact of Web services and published Web documents through access and Web citation metrics.