Chen, Wei; Deng, Da
2014-11-11
We report a new, low-cost and simple top-down approach, "sodium-cutting", to cut and open nanostructures deposited on a nonplanar surface on a large scale. The feasibility of sodium-cutting was demonstrated with the successfully cutting open of ∼100% carbon nanospheres into nanobowls on a large scale from Sn@C nanospheres for the first time.
Outbreaks associated to large open air festivals, including music festivals, 1980 to 2012.
Botelho-Nevers, E; Gautret, P
2013-03-14
In the minds of many, large scale open air festivals have become associated with spring and summer, attracting many people, and in the case of music festivals, thousands of music fans. These festivals share the usual health risks associated with large mass gatherings, including transmission of communicable diseases and risk of outbreaks. Large scale open air festivals have however specific characteristics, including outdoor settings, on-site housing and food supply and the generally young age of the participants. Outbreaks at large scale open air festivals have been caused by Cryptosporium parvum, Campylobacter spp., Escherichia coli, Salmonella enterica, Shigella sonnei, Staphylococcus aureus, hepatitis A virus, influenza virus, measles virus, mumps virus and norovirus. Faecal-oral and respiratory transmissions of pathogens result from non-compliance with hygiene rules, inadequate sanitation and insufficient vaccination coverage. Sexual transmission of infectious diseases may also occur and is likely to be underestimated and underreported. Enhanced surveillance during and after festivals is essential. Preventive measures such as immunisations of participants and advice on-site and via social networks should be considered to reduce outbreaks at these large scale open air festivals.
Cosmic microwave background anomalies in an open universe.
Liddle, Andrew R; Cortês, Marina
2013-09-13
We argue that the observed large-scale cosmic microwave anomalies, discovered by WMAP and confirmed by the Planck satellite, are most naturally explained in the context of a marginally open universe. Particular focus is placed on the dipole power asymmetry, via an open universe implementation of the large-scale gradient mechanism of Erickcek et al. Open inflation models, which are motivated by the string landscape and which can excite "supercurvature" perturbation modes, can explain the presence of a very-large-scale perturbation that leads to a dipole modulation of the power spectrum measured by a typical observer. We provide a specific implementation of the scenario which appears compatible with all existing constraints.
Emerging Roles: Key Insights from Librarians in a Massive Open Online Course
ERIC Educational Resources Information Center
Stephens, Michael; Jones, Kyle M. L.
2015-01-01
From the cutting edge of innovations in online education comes the MOOC (Massive Open Online Course), a potentially disruptive and transformational mechanism for large-scale learning. What's the role of librarians in a MOOC? What can librarians learn from participating in a large-scale professional development opportunity delivered in an open…
On the large-scale structures formed by wakes of open cosmic strings
NASA Technical Reports Server (NTRS)
Hara, Tetsuya; Morioka, Shoji; Miyoshi, Shigeru
1990-01-01
Large-scale structures of the universe have been variously described as sheetlike, filamentary, cellular, bubbles or spongelike. Recently cosmic strings became one of viable candidates for a galaxy formation scenario, and some of the large-scale structures seem to be simply explained by the open cosmic strings. According to this scenario, sheets are wakes which are traces of moving open cosmic strings where dark matter and baryonic matter have accumulated. Filaments are intersections of such wakes and high density regions are places where three wakes intersect almost orthogonally. The wakes formed at t sub eq become the largest surface density among all wakes, where t sub eq is the epoch when matter density equals to radiation density. If we assume that there is one open cosmic string per each horizon, then it can be explained that the typical distances among wakes, filaments and clusters are also approx. 10(exp 2) Mpc. This model does not exclude a much more large scale structure. Open cosmic string may move even now and accumulate cold dark matter after its traces. However, the surface density is much smaller than the ones formed at t sub eq. From this model, it is expected that the typical high density region will have extended features such as six filaments and three sheets and be surrounded by eight empty regions (voids). Here, the authors are mainly concerned with such structures and have made numerical simulations for the formation of such large scale structures.
2012-04-01
Both flame lengths shrink and large scale disruptions occur downstream with vortex shedding carrying reaction zones. Flames in both flameholders...9) the flame structure changes dramatically for both regular and open-slit V-gutter. Both flame lengths shrink and large scale disruptions occur...reduces the flame length . However, qualitatively the open-slit V-gutter appears to be more sensitive than the regular V-gutter. Both flames remain
Do large-scale assessments measure students' ability to integrate scientific knowledge?
NASA Astrophysics Data System (ADS)
Lee, Hee-Sun
2010-03-01
Large-scale assessments are used as means to diagnose the current status of student achievement in science and compare students across schools, states, and countries. For efficiency, multiple-choice items and dichotomously-scored open-ended items are pervasively used in large-scale assessments such as Trends in International Math and Science Study (TIMSS). This study investigated how well these items measure secondary school students' ability to integrate scientific knowledge. This study collected responses of 8400 students to 116 multiple-choice and 84 open-ended items and applied an Item Response Theory analysis based on the Rasch Partial Credit Model. Results indicate that most multiple-choice items and dichotomously-scored open-ended items can be used to determine whether students have normative ideas about science topics, but cannot measure whether students integrate multiple pieces of relevant science ideas. Only when the scoring rubric is redesigned to capture subtle nuances of student open-ended responses, open-ended items become a valid and reliable tool to assess students' knowledge integration ability.
Large-Scale 1:1 Computing Initiatives: An Open Access Database
ERIC Educational Resources Information Center
Richardson, Jayson W.; McLeod, Scott; Flora, Kevin; Sauers, Nick J.; Kannan, Sathiamoorthy; Sincar, Mehmet
2013-01-01
This article details the spread and scope of large-scale 1:1 computing initiatives around the world. What follows is a review of the existing literature around 1:1 programs followed by a description of the large-scale 1:1 database. Main findings include: 1) the XO and the Classmate PC dominate large-scale 1:1 initiatives; 2) if professional…
Effects of trade openness and market scale on different regions
NASA Astrophysics Data System (ADS)
Tian, Renqu; Yang, Zisheng
2017-04-01
This paper revisits the relationship between growth, trade openness and market scale. Empirical studies have provided that area develops lopsided problem in China is increasingly serious, while large trade openness and market scale bring about more economic growth. We use a number of data set from province-level’s gross domestic product and socio-economic, as well as statistical methods panel ordinary least squares and instrumental variables estimation techniques to explore the effects of trade openness and regional market scale on the three major economic regions. The results indicate: Firstly, the impact of market scale and trade openness on economic growth is found to be positive. Secondly, the overall regional disparity is owing to the trade openness, market scale and macroeconomic policies. Thirdly, midland and western region should take advantage of regional geographical location and resource to expand exports and narrow the regional difference.
Giroux, Marie-Andrée; Valiquette, Éliane; Tremblay, Jean-Pierre; Côté, Steeve D
2015-01-01
Documenting habitat-related patterns in foraging behaviour at the individual level and over large temporal scales remains challenging for large herbivores. Stable isotope analysis could represent a valuable tool to quantify habitat-related foraging behaviour at the scale of individuals and over large temporal scales in forest dwelling large herbivores living in coastal environments, because the carbon (δ13C) or nitrogen (δ15N) isotopic signatures of forage can differ between open and closed habitats or between terrestrial and littoral forage, respectively. Here, we examined if we could detect isotopic differences between the different assemblages of forage taxa consumed by white-tailed deer that can be found in open, closed, supralittoral, and littoral habitats. We showed that δ13C of assemblages of forage taxa were 3.0 ‰ lower in closed than in open habitats, while δ15N were 2.0 ‰ and 7.4 ‰ higher in supralittoral and littoral habitats, respectively, than in terrestrial habitats. Stable isotope analysis may represent an additional technique for ecologists interested in quantifiying the consumption of terrestrial vs. marine autotrophs. Yet, given the relative isotopic proximity and the overlap between forage from open, closed, and supralittoral habitats, the next step would be to determine the potential to estimate their contribution to herbivore diet.
Implementation of highly parallel and large scale GW calculations within the OpenAtom software
NASA Astrophysics Data System (ADS)
Ismail-Beigi, Sohrab
The need to describe electronic excitations with better accuracy than provided by band structures produced by Density Functional Theory (DFT) has been a long-term enterprise for the computational condensed matter and materials theory communities. In some cases, appropriate theoretical frameworks have existed for some time but have been difficult to apply widely due to computational cost. For example, the GW approximation incorporates a great deal of important non-local and dynamical electronic interaction effects but has been too computationally expensive for routine use in large materials simulations. OpenAtom is an open source massively parallel ab initiodensity functional software package based on plane waves and pseudopotentials (http://charm.cs.uiuc.edu/OpenAtom/) that takes advantage of the Charm + + parallel framework. At present, it is developed via a three-way collaboration, funded by an NSF SI2-SSI grant (ACI-1339804), between Yale (Ismail-Beigi), IBM T. J. Watson (Glenn Martyna) and the University of Illinois at Urbana Champaign (Laxmikant Kale). We will describe the project and our current approach towards implementing large scale GW calculations with OpenAtom. Potential applications of large scale parallel GW software for problems involving electronic excitations in semiconductor and/or metal oxide systems will be also be pointed out.
van Scheppingen, Arjella R; de Vroome, Ernest M M; Ten Have, Kristin C J M; Bos, Ellen H; Zwetsloot, Gerard I J M; van Mechelen, W
2014-11-01
To examine the effectiveness of an organizational large-scale intervention applied to induce a health-promoting organizational change process. A quasi-experimental, "as-treated" design was used. Regression analyses on data of employees of a Dutch dairy company (n = 324) were used to examine the effects on bonding social capital, openness, and autonomous motivation toward health and on employees' lifestyle, health, vitality, and sustainable employability. Also, the sensitivity of the intervention components was examined. Intervention effects were found for bonding social capital, openness toward health, smoking, healthy eating, and sustainable employability. The effects were primarily attributable to the intervention's dialogue component. The change process initiated by the large-scale intervention contributed to a social climate in the workplace that promoted health and ownership toward health. The study confirms the relevance of collective change processes for health promotion.
Data management strategies for multinational large-scale systems biology projects.
Wruck, Wasco; Peuker, Martin; Regenbrecht, Christian R A
2014-01-01
Good accessibility of publicly funded research data is essential to secure an open scientific system and eventually becomes mandatory [Wellcome Trust will Penalise Scientists Who Don't Embrace Open Access. The Guardian 2012]. By the use of high-throughput methods in many research areas from physics to systems biology, large data collections are increasingly important as raw material for research. Here, we present strategies worked out by international and national institutions targeting open access to publicly funded research data via incentives or obligations to share data. Funding organizations such as the British Wellcome Trust therefore have developed data sharing policies and request commitment to data management and sharing in grant applications. Increased citation rates are a profound argument for sharing publication data. Pre-publication sharing might be rewarded by a data citation credit system via digital object identifiers (DOIs) which have initially been in use for data objects. Besides policies and incentives, good practice in data management is indispensable. However, appropriate systems for data management of large-scale projects for example in systems biology are hard to find. Here, we give an overview of a selection of open-source data management systems proved to be employed successfully in large-scale projects.
Data management strategies for multinational large-scale systems biology projects
Peuker, Martin; Regenbrecht, Christian R.A.
2014-01-01
Good accessibility of publicly funded research data is essential to secure an open scientific system and eventually becomes mandatory [Wellcome Trust will Penalise Scientists Who Don’t Embrace Open Access. The Guardian 2012]. By the use of high-throughput methods in many research areas from physics to systems biology, large data collections are increasingly important as raw material for research. Here, we present strategies worked out by international and national institutions targeting open access to publicly funded research data via incentives or obligations to share data. Funding organizations such as the British Wellcome Trust therefore have developed data sharing policies and request commitment to data management and sharing in grant applications. Increased citation rates are a profound argument for sharing publication data. Pre-publication sharing might be rewarded by a data citation credit system via digital object identifiers (DOIs) which have initially been in use for data objects. Besides policies and incentives, good practice in data management is indispensable. However, appropriate systems for data management of large-scale projects for example in systems biology are hard to find. Here, we give an overview of a selection of open-source data management systems proved to be employed successfully in large-scale projects. PMID:23047157
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, Junghyun; Gangwon, Jo; Jaehoon, Jung
Applications written solely in OpenCL or CUDA cannot execute on a cluster as a whole. Most previous approaches that extend these programming models to clusters are based on a common idea: designating a centralized host node and coordinating the other nodes with the host for computation. However, the centralized host node is a serious performance bottleneck when the number of nodes is large. In this paper, we propose a scalable and distributed OpenCL framework called SnuCL-D for large-scale clusters. SnuCL-D's remote device virtualization provides an OpenCL application with an illusion that all compute devices in a cluster are confined inmore » a single node. To reduce the amount of control-message and data communication between nodes, SnuCL-D replicates the OpenCL host program execution and data in each node. We also propose a new OpenCL host API function and a queueing optimization technique that significantly reduce the overhead incurred by the previous centralized approaches. To show the effectiveness of SnuCL-D, we evaluate SnuCL-D with a microbenchmark and eleven benchmark applications on a large-scale CPU cluster and a medium-scale GPU cluster.« less
OpenMP parallelization of a gridded SWAT (SWATG)
NASA Astrophysics Data System (ADS)
Zhang, Ying; Hou, Jinliang; Cao, Yongpan; Gu, Juan; Huang, Chunlin
2017-12-01
Large-scale, long-term and high spatial resolution simulation is a common issue in environmental modeling. A Gridded Hydrologic Response Unit (HRU)-based Soil and Water Assessment Tool (SWATG) that integrates grid modeling scheme with different spatial representations also presents such problems. The time-consuming problem affects applications of very high resolution large-scale watershed modeling. The OpenMP (Open Multi-Processing) parallel application interface is integrated with SWATG (called SWATGP) to accelerate grid modeling based on the HRU level. Such parallel implementation takes better advantage of the computational power of a shared memory computer system. We conducted two experiments at multiple temporal and spatial scales of hydrological modeling using SWATG and SWATGP on a high-end server. At 500-m resolution, SWATGP was found to be up to nine times faster than SWATG in modeling over a roughly 2000 km2 watershed with 1 CPU and a 15 thread configuration. The study results demonstrate that parallel models save considerable time relative to traditional sequential simulation runs. Parallel computations of environmental models are beneficial for model applications, especially at large spatial and temporal scales and at high resolutions. The proposed SWATGP model is thus a promising tool for large-scale and high-resolution water resources research and management in addition to offering data fusion and model coupling ability.
Energy Spectral Behaviors of Communication Networks of Open-Source Communities
Yang, Jianmei; Yang, Huijie; Liao, Hao; Wang, Jiangtao; Zeng, Jinqun
2015-01-01
Large-scale online collaborative production activities in open-source communities must be accompanied by large-scale communication activities. Nowadays, the production activities of open-source communities, especially their communication activities, have been more and more concerned. Take CodePlex C # community for example, this paper constructs the complex network models of 12 periods of communication structures of the community based on real data; then discusses the basic concepts of quantum mapping of complex networks, and points out that the purpose of the mapping is to study the structures of complex networks according to the idea of quantum mechanism in studying the structures of large molecules; finally, according to this idea, analyzes and compares the fractal features of the spectra in different quantum mappings of the networks, and concludes that there are multiple self-similarity and criticality in the communication structures of the community. In addition, this paper discusses the insights and application conditions of different quantum mappings in revealing the characteristics of the structures. The proposed quantum mapping method can also be applied to the structural studies of other large-scale organizations. PMID:26047331
How OpenLearn Supports a Business Model for OER
ERIC Educational Resources Information Center
Law, Patrina; Perryman, Leigh-Anne
2017-01-01
In 2013, the Open University (OU) in the UK launched a large-scale survey of users of its OpenLearn platform for open educational resources. The survey results revealed that OpenLearn is functioning as a showcase and a taster for the OU, thereby offering informal learners a bridge to formal education. In 2014 and 2015, the OpenLearn survey was…
High-uniformity centimeter-wide Si etching method for MEMS devices with large opening elements
NASA Astrophysics Data System (ADS)
Okamoto, Yuki; Tohyama, Yukiya; Inagaki, Shunsuke; Takiguchi, Mikio; Ono, Tomoki; Lebrasseur, Eric; Mita, Yoshio
2018-04-01
We propose a compensated mesh pattern filling method to achieve highly uniform wafer depth etching (over hundreds of microns) with a large-area opening (over centimeter). The mesh opening diameter is gradually changed between the center and the edge of a large etching area. Using such a design, the etching depth distribution depending on sidewall distance (known as the local loading effect) inversely compensates for the over-centimeter-scale etching depth distribution, known as the global or within-die(chip)-scale loading effect. Only a single DRIE with test structure patterns provides a micro-electromechanical systems (MEMS) designer with the etched depth dependence on the mesh opening size as well as on the distance from the chip edge, and the designer only has to set the opening size so as to obtain a uniform etching depth over the entire chip. This method is useful when process optimization cannot be performed, such as in the cases of using standard conditions for a foundry service and of short turn-around-time prototyping. To demonstrate, a large MEMS mirror that needed over 1 cm2 of backside etching was successfully fabricated using as-is-provided DRIE conditions.
ERIC Educational Resources Information Center
Alexopoulou, Theodora; Michel, Marije; Murakami, Akira; Meurers, Detmar
2017-01-01
Large-scale learner corpora collected from online language learning platforms, such as the EF-Cambridge Open Language Database (EFCAMDAT), provide opportunities to analyze learner data at an unprecedented scale. However, interpreting the learner language in such corpora requires a precise understanding of tasks: How does the prompt and input of a…
Open Education as a "Heterotopia of Desire"
ERIC Educational Resources Information Center
Gourlay, Lesley
2015-01-01
The movement towards "openness" in education has tended to position itself as inherently democratising, radical, egalitarian and critical of powerful gatekeepers to learning. While "openness" is often positioned as a critique, I will argue that its mainstream discourses--while appearing to oppose large-scale operations of…
RE-Europe, a large-scale dataset for modeling a highly renewable European electricity system
Jensen, Tue V.; Pinson, Pierre
2017-01-01
Future highly renewable energy systems will couple to complex weather and climate dynamics. This coupling is generally not captured in detail by the open models developed in the power and energy system communities, where such open models exist. To enable modeling such a future energy system, we describe a dedicated large-scale dataset for a renewable electric power system. The dataset combines a transmission network model, as well as information for generation and demand. Generation includes conventional generators with their technical and economic characteristics, as well as weather-driven forecasts and corresponding realizations for renewable energy generation for a period of 3 years. These may be scaled according to the envisioned degrees of renewable penetration in a future European energy system. The spatial coverage, completeness and resolution of this dataset, open the door to the evaluation, scaling analysis and replicability check of a wealth of proposals in, e.g., market design, network actor coordination and forecasting of renewable power generation. PMID:29182600
RE-Europe, a large-scale dataset for modeling a highly renewable European electricity system.
Jensen, Tue V; Pinson, Pierre
2017-11-28
Future highly renewable energy systems will couple to complex weather and climate dynamics. This coupling is generally not captured in detail by the open models developed in the power and energy system communities, where such open models exist. To enable modeling such a future energy system, we describe a dedicated large-scale dataset for a renewable electric power system. The dataset combines a transmission network model, as well as information for generation and demand. Generation includes conventional generators with their technical and economic characteristics, as well as weather-driven forecasts and corresponding realizations for renewable energy generation for a period of 3 years. These may be scaled according to the envisioned degrees of renewable penetration in a future European energy system. The spatial coverage, completeness and resolution of this dataset, open the door to the evaluation, scaling analysis and replicability check of a wealth of proposals in, e.g., market design, network actor coordination and forecasting of renewable power generation.
RE-Europe, a large-scale dataset for modeling a highly renewable European electricity system
NASA Astrophysics Data System (ADS)
Jensen, Tue V.; Pinson, Pierre
2017-11-01
Future highly renewable energy systems will couple to complex weather and climate dynamics. This coupling is generally not captured in detail by the open models developed in the power and energy system communities, where such open models exist. To enable modeling such a future energy system, we describe a dedicated large-scale dataset for a renewable electric power system. The dataset combines a transmission network model, as well as information for generation and demand. Generation includes conventional generators with their technical and economic characteristics, as well as weather-driven forecasts and corresponding realizations for renewable energy generation for a period of 3 years. These may be scaled according to the envisioned degrees of renewable penetration in a future European energy system. The spatial coverage, completeness and resolution of this dataset, open the door to the evaluation, scaling analysis and replicability check of a wealth of proposals in, e.g., market design, network actor coordination and forecasting of renewable power generation.
Desland, Fiona A; Afzal, Aqeela; Warraich, Zuha; Mocco, J
2014-01-01
Animal models of stroke have been crucial in advancing our understanding of the pathophysiology of cerebral ischemia. Currently, the standards for determining neurological deficit in rodents are the Bederson and Garcia scales, manual assessments scoring animals based on parameters ranked on a narrow scale of severity. Automated open field analysis of a live-video tracking system that analyzes animal behavior may provide a more sensitive test. Results obtained from the manual Bederson and Garcia scales did not show significant differences between pre- and post-stroke animals in a small cohort. When using the same cohort, however, post-stroke data obtained from automated open field analysis showed significant differences in several parameters. Furthermore, large cohort analysis also demonstrated increased sensitivity with automated open field analysis versus the Bederson and Garcia scales. These early data indicate use of automated open field analysis software may provide a more sensitive assessment when compared to traditional Bederson and Garcia scales.
Novel approach for extinguishing large-scale coal fires using gas-liquid foams in open pit mines.
Lu, Xinxiao; Wang, Deming; Qin, Botao; Tian, Fuchao; Shi, Guangyi; Dong, Shuaijun
2015-12-01
Coal fires are a serious threat to the workers' security and safe production in open pit mines. The coal fire source is hidden and innumerable, and the large-area cavity is prevalent in the coal seam after the coal burned, causing the conventional extinguishment technology difficult to work. Foams are considered as an efficient means of fire extinguishment in these large-scale workplaces. A noble foam preparation method is introduced, and an original design of cavitation jet device is proposed to add foaming agent stably. The jet cavitation occurs when the water flow rate and pressure ratio reach specified values. Through self-building foaming system, the high performance foams are produced and then infused into the blast drilling holes at a large flow. Without complicated operation, this system is found to be very suitable for extinguishing large-scale coal fires. Field application shows that foam generation adopting the proposed key technology makes a good fire extinguishment effect. The temperature reduction using foams is 6-7 times higher than water, and CO concentration is reduced from 9.43 to 0.092‰ in the drilling hole. The coal fires are controlled successfully in open pit mines, ensuring the normal production as well as the security of personnel and equipment.
An Open, Large-Scale, Collaborative Effort to Estimate the Reproducibility of Psychological Science.
2012-11-01
Reproducibility is a defining feature of science. However, because of strong incentives for innovation and weak incentives for confirmation, direct replication is rarely practiced or published. The Reproducibility Project is an open, large-scale, collaborative effort to systematically examine the rate and predictors of reproducibility in psychological science. So far, 72 volunteer researchers from 41 institutions have organized to openly and transparently replicate studies published in three prominent psychological journals in 2008. Multiple methods will be used to evaluate the findings, calculate an empirical rate of replication, and investigate factors that predict reproducibility. Whatever the result, a better understanding of reproducibility will ultimately improve confidence in scientific methodology and findings. © The Author(s) 2012.
Coordinating the Commons: Diversity & Dynamics in Open Collaborations
ERIC Educational Resources Information Center
Morgan, Jonathan T.
2013-01-01
The success of Wikipedia demonstrates that open collaboration can be an effective model for organizing geographically-distributed volunteers to perform complex, sustained work at a massive scale. However, Wikipedia's history also demonstrates some of the challenges that large, long-term open collaborations face: the core community of Wikipedia…
Large-angle cosmic microwave background anisotropies in an open universe
NASA Technical Reports Server (NTRS)
Kamionkowski, Marc; Spergel, David N.
1994-01-01
If the universe is open, scales larger than the curvature scale may be probed by observation of large-angle fluctuations in the cosmic microwave background (CMB). We consider primordial adiabatic perturbations and discuss power spectra that are power laws in volume, wavelength, and eigenvalue of the Laplace operator. Such spectra may have arisen if, for example, the universe underwent a period of `frustated' inflation. The resulting large-angle anisotropies of the CMB are computed. The amplitude generally increases as Omega is decreased but decreases as h is increased. Interestingly enough, for all three Ansaetze, anisotropies on angular scales larger than the curvature scale are suppressed relative to the anisotropies on scales smaller than the curvature scale, but cosmic variance makes discrimination between various models difficult. Models with 0.2 approximately less than Omega h approximately less than 0.3 appear compatible with CMB fluctuations detected by Cosmic Background Explorer Satellite (COBE) and the Tenerife experiment and with the amplitude and spectrum of fluctuations of galaxy counts in the APM, CfA, and 1.2 Jy IRAS surveys. COBE normalization for these models yields sigma(sub 8) approximately = 0.5 - 0.7. Models with smaller values of Omega h when normalized to COBE require bias factors in excess of 2 to be compatible with the observed galaxy counts on the 8/h Mpc scale. Requiring that the age of the universe exceed 10 Gyr implies that Omega approximately greater than 0.25, while requiring that from the last-scattering term in the Sachs-Wolfe formula, large-angle anisotropies come primarily from the decay of potential fluctuations at z approximately less than 1/Omega. Thus, if the universe is open, COBE has been detecting temperature fluctuations produced at moderate redshift rather than at z approximately 1300.
Composite turbine blade design options for Claude (open) cycle OTEC power systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Penney, T R
1985-11-01
Small-scale turbine rotors made from composites offer several technical advantages for a Claude (open) cycle ocean thermal energy conversion (OTEC) power system. Westinghouse Electric Corporation has designed a composite turbine rotor/disk using state-of-the-art analysis methods for large-scale (100-MW/sub e/) open cycle OTEC applications. Near-term demonstrations using conventional low-pressure turbine blade shapes with composite material would achieve feasibility and modern credibility of the open cycle OTEC power system. Application of composite blades for low-pressure turbo-machinery potentially improves the reliability of conventional metal blades affected by stress corrosion.
Using Large-Scale Databases in Evaluation: Advances, Opportunities, and Challenges
ERIC Educational Resources Information Center
Penuel, William R.; Means, Barbara
2011-01-01
Major advances in the number, capabilities, and quality of state, national, and transnational databases have opened up new opportunities for evaluators. Both large-scale data sets collected for administrative purposes and those collected by other researchers can provide data for a variety of evaluation-related activities. These include (a)…
The observation of possible reconnection events in the boundary changes of solar coronal holes
NASA Technical Reports Server (NTRS)
Kahler, S. W.; Moses, J. Daniel
1989-01-01
Coronal holes are large scale regions of magnetically open fields which are easily observed in solar soft X-ray images. The boundaries of coronal holes are separatrices between large scale regions of open and closed magnetic fields where one might expect to observe evidence of solar magnetic reconnection. Previous studies by Nolte and colleagues using Skylab X-ray images established that large scale (greater than or equal to 9 x 10(4) km) changes in coronal hole boundaries were due to coronal processes, i.e., magnetic reconnection, rather than to photospheric motions. Those studies were limited to time scales of about one day, and no conclusion could be drawn about the size and time scales of the reconnection process at hole boundaries. Sequences of appropriate Skylab X-ray images were used with a time resolution of about 90 min during times of the central meridian passages of the coronal hole labelled Coronal Hole 1 to search for hole boundary changes which can yield the spatial and temporal scales of coronal magnetic reconnection. It was found that 29 of 32 observed boundary changes could be associated with bright points. The appearance of the bright point may be the signature of reconnection between small scale and large scale magnetic fields. The observed boundary changes contributed to the quasi-rigid rotation of Coronal Hole 1.
Agent-based large-scale emergency evacuation using real-time open government data.
DOT National Transportation Integrated Search
2014-01-01
The open government initiatives have provided tremendous data resources for the : transportation system and emergency services in urban areas. This paper proposes : a traffic simulation framework using high temporal resolution demographic data : and ...
Analogue scale modelling of extensional tectonic processes using a large state-of-the-art centrifuge
NASA Astrophysics Data System (ADS)
Park, Heon-Joon; Lee, Changyeol
2017-04-01
Analogue scale modelling of extensional tectonic processes such as rifting and basin opening has been numerously conducted. Among the controlling factors, gravitational acceleration (g) on the scale models was regarded as a constant (Earth's gravity) in the most of the analogue model studies, and only a few model studies considered larger gravitational acceleration by using a centrifuge (an apparatus generating large centrifugal force by rotating the model at a high speed). Although analogue models using a centrifuge allow large scale-down and accelerated deformation that is derived by density differences such as salt diapir, the possible model size is mostly limited up to 10 cm. A state-of-the-art centrifuge installed at the KOCED Geotechnical Centrifuge Testing Center, Korea Advanced Institute of Science and Technology (KAIST) allows a large surface area of the scale-models up to 70 by 70 cm under the maximum capacity of 240 g-tons. Using the centrifuge, we will conduct analogue scale modelling of the extensional tectonic processes such as opening of the back-arc basin. Acknowledgement This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (grant number 2014R1A6A3A04056405).
NASA Astrophysics Data System (ADS)
Kane, V. R.; McGaughey, R. J.; Asner, G. P.; Kane, J. T.; Churchill, D.; Vaughn, N.
2016-12-01
Most natural forests are structured as mosaics of tree clumps and openings. These mosaics reflect both the underlying patterns of the biophysical environment and the finer scale patterns of disturbance and regrowth. We have developed methods to quantify and map patterns of tree clumps and openings at scales from within stands to landscapes using airborne LiDAR. While many studies have used LiDAR data to identify individual trees, we also identify clumps as adjacent trees with similar heights within a stand that likely established at a similar time following a disturbance. We characterize openings by both size class and shape complexity. Spatial statistics are used to identify patterns of tree clumps and openings at the local (0.81 ha) scale, and these patterns are then mapped across entire landscapes. We use LiDAR data acquired over Sequoia National Park, California, USA, to show how forest structure varies with patterns of productivity driven by the biophysical environment. We then show how clump and opening patterns vary with different fire histories and how recent drought mortality correlates with different tree clump and opening structural mosaics. We also demonstrate that nesting sites for the California spotted owl, a species of concern, are associated with clumps of large (>32 and especially >48 m) trees but that the surrounding foraging areas consist of a heterogeneous pattern of forest structure. These methods are especially useful for studying clumps of large trees, which dominate above ground forest biomass, and the effects of disturbance on the abundance and pattern of large trees as key forest structures.
Evolution of Large-Scale Magnetic Fields and State Transitions in Black Hole X-Ray Binaries
NASA Astrophysics Data System (ADS)
Wang, Ding-Xiong; Huang, Chang-Yin; Wang, Jiu-Zhou
2010-04-01
The state transitions of black hole (BH) X-ray binaries are discussed based on the evolution of large-scale magnetic fields, in which the combination of three energy mechanisms are involved: (1) the Blandford-Znajek (BZ) process related to the open field lines connecting a rotating BH with remote astrophysical loads, (2) the magnetic coupling (MC) process related to the closed field lines connecting the BH with its surrounding accretion disk, and (3) the Blandford-Payne (BP) process related to the open field lines connecting the disk with remote astrophysical loads. It turns out that each spectral state of the BH binaries corresponds to each configuration of magnetic field in BH magnetosphere, and the main characteristics of low/hard (LH) state, hard intermediate (HIM) state and steep power law (SPL) state are roughly fitted based on the evolution of large-scale magnetic fields associated with disk accretion.
ERIC Educational Resources Information Center
Xiong, Yao; Suen, Hoi K.
2018-01-01
The development of massive open online courses (MOOCs) has launched an era of large-scale interactive participation in education. While massive open enrolment and the advances of learning technology are creating exciting potentials for lifelong learning in formal and informal ways, the implementation of efficient and effective assessment is still…
NASA Astrophysics Data System (ADS)
Yang, D.; Fu, C. S.; Binford, M. W.
2017-12-01
The southeastern United States has high landscape heterogeneity, withheavily managed forestlands, highly developed agriculture lands, and multiple metropolitan areas. Human activities are transforming and altering land patterns and structures in both negative and positive manners. A land-use map for at the greater scale is a heavy computation task but is critical to most landowners, researchers, and decision makers, enabling them to make informed decisions for varying objectives. There are two major difficulties in generating the classification maps at the regional scale: the necessity of large training point sets and the expensive computation cost-in terms of both money and time-in classifier modeling. Volunteered Geographic Information (VGI) opens a new era in mapping and visualizing our world, where the platform is open for collecting valuable georeferenced information by volunteer citizens, and the data is freely available to the public. As one of the most well-known VGI initiatives, OpenStreetMap (OSM) contributes not only road network distribution, but also the potential for using this data to justify land cover and land use classifications. Google Earth Engine (GEE) is a platform designed for cloud-based mapping with a robust and fast computing power. Most large scale and national mapping approaches confuse "land cover" and "land-use", or build up the land-use database based on modeled land cover datasets. Unlike most other large-scale approaches, we distinguish and differentiate land-use from land cover. By focusing our prime objective of mapping land-use and management practices, a robust regional land-use mapping approach is developed by incorporating the OpenstreepMap dataset into Earth observation remote sensing imageries instead of the often-used land cover base maps.
Spectral fingerprints of large-scale neuronal interactions.
Siegel, Markus; Donner, Tobias H; Engel, Andreas K
2012-01-11
Cognition results from interactions among functionally specialized but widely distributed brain regions; however, neuroscience has so far largely focused on characterizing the function of individual brain regions and neurons therein. Here we discuss recent studies that have instead investigated the interactions between brain regions during cognitive processes by assessing correlations between neuronal oscillations in different regions of the primate cerebral cortex. These studies have opened a new window onto the large-scale circuit mechanisms underlying sensorimotor decision-making and top-down attention. We propose that frequency-specific neuronal correlations in large-scale cortical networks may be 'fingerprints' of canonical neuronal computations underlying cognitive processes.
A relativistic signature in large-scale structure
NASA Astrophysics Data System (ADS)
Bartolo, Nicola; Bertacca, Daniele; Bruni, Marco; Koyama, Kazuya; Maartens, Roy; Matarrese, Sabino; Sasaki, Misao; Verde, Licia; Wands, David
2016-09-01
In General Relativity, the constraint equation relating metric and density perturbations is inherently nonlinear, leading to an effective non-Gaussianity in the dark matter density field on large scales-even if the primordial metric perturbation is Gaussian. Intrinsic non-Gaussianity in the large-scale dark matter overdensity in GR is real and physical. However, the variance smoothed on a local physical scale is not correlated with the large-scale curvature perturbation, so that there is no relativistic signature in the galaxy bias when using the simplest model of bias. It is an open question whether the observable mass proxies such as luminosity or weak lensing correspond directly to the physical mass in the simple halo bias model. If not, there may be observables that encode this relativistic signature.
NASA Astrophysics Data System (ADS)
McCoy, Isabel L.; Wood, Robert; Fletcher, Jennifer K.
2017-11-01
Mesoscale cellular convective (MCC) clouds occur in large-scale patterns over the ocean and have important radiative effects on the climate system. An examination of time-varying meteorological conditions associated with satellite-observed open and closed MCC clouds is conducted to illustrate the influence of large-scale meteorological conditions. Marine cold air outbreaks (MCAO) influence the development of open MCC clouds and the transition from closed to open MCC clouds. MCC neural network classifications on Moderate Resolution Imaging Spectroradiometer (MODIS) data for 2008 are collocated with Clouds and the Earth's Radiant Energy System (CERES) data and ERA-Interim reanalysis to determine the radiative effects of MCC clouds and their thermodynamic environments. Closed MCC clouds are found to have much higher albedo on average than open MCC clouds for the same cloud fraction. Three meteorological control metrics are tested: sea-air temperature difference (ΔT), estimated inversion strength (EIS), and a MCAO index (M). These predictive metrics illustrate the importance of atmospheric surface forcing and static stability for open and closed MCC cloud formation. Predictive sigmoidal relations are found between M and MCC cloud frequency globally and regionally: negative for closed MCC cloud and positive for open MCC cloud. The open MCC cloud seasonal cycle is well correlated with M, while the seasonality of closed MCC clouds is well correlated with M in the midlatitudes and EIS in the tropics and subtropics. M is found to best distinguish open and closed MCC clouds on average over shorter time scales. The possibility of a MCC cloud feedback is discussed.
Research on OpenStack of open source cloud computing in colleges and universities’ computer room
NASA Astrophysics Data System (ADS)
Wang, Lei; Zhang, Dandan
2017-06-01
In recent years, the cloud computing technology has a rapid development, especially open source cloud computing. Open source cloud computing has attracted a large number of user groups by the advantages of open source and low cost, have now become a large-scale promotion and application. In this paper, firstly we briefly introduced the main functions and architecture of the open source cloud computing OpenStack tools, and then discussed deeply the core problems of computer labs in colleges and universities. Combining with this research, it is not that the specific application and deployment of university computer rooms with OpenStack tool. The experimental results show that the application of OpenStack tool can efficiently and conveniently deploy cloud of university computer room, and its performance is stable and the functional value is good.
Gray, B.R.; Shi, W.; Houser, J.N.; Rogala, J.T.; Guan, Z.; Cochran-Biederman, J. L.
2011-01-01
Ecological restoration efforts in large rivers generally aim to ameliorate ecological effects associated with large-scale modification of those rivers. This study examined whether the effects of restoration efforts-specifically those of island construction-within a largely open water restoration area of the Upper Mississippi River (UMR) might be seen at the spatial scale of that 3476ha area. The cumulative effects of island construction, when observed over multiple years, were postulated to have made the restoration area increasingly similar to a positive reference area (a proximate area comprising contiguous backwater areas) and increasingly different from two negative reference areas. The negative reference areas represented the Mississippi River main channel in an area proximate to the restoration area and an open water area in a related Mississippi River reach that has seen relatively little restoration effort. Inferences on the effects of restoration were made by comparing constrained and unconstrained models of summer chlorophyll a (CHL), summer inorganic suspended solids (ISS) and counts of benthic mayfly larvae. Constrained models forced trends in means or in both means and sampling variances to become, over time, increasingly similar to those in the positive reference area and increasingly dissimilar to those in the negative reference areas. Trends were estimated over 12- (mayflies) or 14-year sampling periods, and were evaluated using model information criteria. Based on these methods, restoration effects were observed for CHL and mayflies while evidence in favour of restoration effects on ISS was equivocal. These findings suggest that the cumulative effects of island building at relatively large spatial scales within large rivers may be estimated using data from large-scale surveillance monitoring programs. Published in 2010 by John Wiley & Sons, Ltd.
Explorative Function in Williams Syndrome Analyzed through a Large-Scale Task with Multiple Rewards
ERIC Educational Resources Information Center
Foti, F.; Petrosini, L.; Cutuli, D.; Menghini, D.; Chiarotti, F.; Vicari, S.; Mandolesi, L.
2011-01-01
This study aimed to evaluate spatial function in subjects with Williams syndrome (WS) by using a large-scale task with multiple rewards and comparing the spatial abilities of WS subjects with those of mental age-matched control children. In the present spatial task, WS participants had to explore an open space to search nine rewards placed in…
Neural ensemble communities: open-source approaches to hardware for large-scale electrophysiology.
Siegle, Joshua H; Hale, Gregory J; Newman, Jonathan P; Voigts, Jakob
2015-06-01
One often-overlooked factor when selecting a platform for large-scale electrophysiology is whether or not a particular data acquisition system is 'open' or 'closed': that is, whether or not the system's schematics and source code are available to end users. Open systems have a reputation for being difficult to acquire, poorly documented, and hard to maintain. With the arrival of more powerful and compact integrated circuits, rapid prototyping services, and web-based tools for collaborative development, these stereotypes must be reconsidered. We discuss some of the reasons why multichannel extracellular electrophysiology could benefit from open-source approaches and describe examples of successful community-driven tool development within this field. In order to promote the adoption of open-source hardware and to reduce the need for redundant development efforts, we advocate a move toward standardized interfaces that connect each element of the data processing pipeline. This will give researchers the flexibility to modify their tools when necessary, while allowing them to continue to benefit from the high-quality products and expertise provided by commercial vendors. Copyright © 2014 Elsevier Ltd. All rights reserved.
Large Scale Spectral Line Mapping of Galactic Regions with CCAT-Prime
NASA Astrophysics Data System (ADS)
Simon, Robert
2018-01-01
CCAT-prime is a 6-m submillimeter telescope that is being built on the top of Cerro Chajnantor (5600 m altitude) overlooking the ALMA plateau in the Atacama Desert. Its novel Crossed-Dragone design enables a large field of view without blockage and is thus particularly well suited for large scale surveys in the continuum and spectral lines targeting important questions ranging from star formation in the Milky Way to cosmology. On this poster, we focus on the large scale mapping opportunities in important spectral cooling lines of the interstellar medium opened up by CCAT-prime and the Cologne heterodyne instrument CHAI.
DOE Office of Scientific and Technical Information (OSTI.GOV)
2014-08-21
Recent advancements in technology scaling have shown a trend towards greater integration with large-scale chips containing thousands of processors connected to memories and other I/O devices using non-trivial network topologies. Software simulation proves insufficient to study the tradeoffs in such complex systems due to slow execution time, whereas hardware RTL development is too time-consuming. We present OpenSoC Fabric, an on-chip network generation infrastructure which aims to provide a parameterizable and powerful on-chip network generator for evaluating future high performance computing architectures based on SoC technology. OpenSoC Fabric leverages a new hardware DSL, Chisel, which contains powerful abstractions provided by itsmore » base language, Scala, and generates both software (C++) and hardware (Verilog) models from a single code base. The OpenSoC Fabric2 infrastructure is modeled after existing state-of-the-art simulators, offers large and powerful collections of configuration options, and follows object-oriented design and functional programming to make functionality extension as easy as possible.« less
Relationship between Birkeland current regions, particle precipitation, and electric fields
NASA Technical Reports Server (NTRS)
De La Beaujardiere, O.; Watermann, J.; Newell, P.; Rich, F.
1993-01-01
The relationship of the large-scale dayside Birkeland currents to large-scale particle precipitation patterns, currents, and convection is examined using DMSP and Sondrestrom radar observations. It is found that the local time of the mantle currents is not limited to the longitude of the cusp proper, but covers a larger local time extent. The mantle currents flow entirely on open field lines. About half of region 1 currents flow on open field lines, consistent with the assumption that the region 1 currents are generated by the solar wind dynamo and flow within the surface that separates open and closed field lines. More than 80 percent of the Birkeland current boundaries do not correspond to particle precipitation boundaries. Region 2 currents extend beyond the plasma sheet poleward boundary; region 1 currents flow in part on open field lines; mantle currents and mantle particles are not coincident. On most passes when a triple current sheet is observed, the convection reversal is located on closed field lines.
Aaron B. Shiels; Grizelle Gonzalez; D. Jean Lodge; Michael R Willig; Jess K. Zimmerman
2015-01-01
Intense hurricanes disturb many tropical forests, but the key mechanisms driving post-hurricane forest changes are not fully understood. In Puerto Rico, we used a replicated factorial experiment to determine the mechanisms of forest change associated with canopy openness and organic matter (debris) addition. Cascading effects from canopy openness accounted for...
NASA Astrophysics Data System (ADS)
Federico, Ivan; Pinardi, Nadia; Coppini, Giovanni; Oddo, Paolo; Lecci, Rita; Mossa, Michele
2017-01-01
SANIFS (Southern Adriatic Northern Ionian coastal Forecasting System) is a coastal-ocean operational system based on the unstructured grid finite-element three-dimensional hydrodynamic SHYFEM model, providing short-term forecasts. The operational chain is based on a downscaling approach starting from the large-scale system for the entire Mediterranean Basin (MFS, Mediterranean Forecasting System), which provides initial and boundary condition fields to the nested system. The model is configured to provide hydrodynamics and active tracer forecasts both in open ocean and coastal waters of southeastern Italy using a variable horizontal resolution from the open sea (3-4 km) to coastal areas (50-500 m). Given that the coastal fields are driven by a combination of both local (also known as coastal) and deep-ocean forcings propagating along the shelf, the performance of SANIFS was verified both in forecast and simulation mode, first (i) on the large and shelf-coastal scales by comparing with a large-scale survey CTD (conductivity-temperature-depth) in the Gulf of Taranto and then (ii) on the coastal-harbour scale (Mar Grande of Taranto) by comparison with CTD, ADCP (acoustic doppler current profiler) and tide gauge data. Sensitivity tests were performed on initialization conditions (mainly focused on spin-up procedures) and on surface boundary conditions by assessing the reliability of two alternative datasets at different horizontal resolution (12.5 and 6.5 km). The SANIFS forecasts at a lead time of 1 day were compared with the MFS forecasts, highlighting that SANIFS is able to retain the large-scale dynamics of MFS. The large-scale dynamics of MFS are correctly propagated to the shelf-coastal scale, improving the forecast accuracy (+17 % for temperature and +6 % for salinity compared to MFS). Moreover, the added value of SANIFS was assessed on the coastal-harbour scale, which is not covered by the coarse resolution of MFS, where the fields forecasted by SANIFS reproduced the observations well (temperature RMSE equal to 0.11 °C). Furthermore, SANIFS simulations were compared with hourly time series of temperature, sea level and velocity measured on the coastal-harbour scale, showing a good agreement. Simulations in the Gulf of Taranto described a circulation mainly characterized by an anticyclonic gyre with the presence of cyclonic vortexes in shelf-coastal areas. A surface water inflow from the open sea to Mar Grande characterizes the coastal-harbour scale.
CFD Script for Rapid TPS Damage Assessment
NASA Technical Reports Server (NTRS)
McCloud, Peter
2013-01-01
This grid generation script creates unstructured CFD grids for rapid thermal protection system (TPS) damage aeroheating assessments. The existing manual solution is cumbersome, open to errors, and slow. The invention takes a large-scale geometry grid and its large-scale CFD solution, and creates a unstructured patch grid that models the TPS damage. The flow field boundary condition for the patch grid is then interpolated from the large-scale CFD solution. It speeds up the generation of CFD grids and solutions in the modeling of TPS damages and their aeroheating assessment. This process was successfully utilized during STS-134.
Joanne M. Sharpe; Aaron B. Shiels
2014-01-01
Ferns are abundant in most rainforest understories yet their responses to hurricanes have not been well studied. Fern community structure, growth and spore production were monitored for two years before and five years after a large-scale experiment that simulated two key components of severe hurricane disturbance: canopy openness and debris deposition. The canopy was...
Peng, Tao; Lv, Haifeng; He, Daping; Pan, Mu; Mu, Shichun
2013-01-01
A large-scale availability of the graphene is critical to the successful application of graphene-based electronic devices. The growth of epitaxial graphene (EG) on insulating silicon carbide (SiC) surfaces has opened a new promising route for large-scale high-quality graphene production. However, two key obstacles to epitaxial growth are extremely high requirements for almost perfectly ordered crystal SiC and harsh process conditions. Here, we report that the amorphous SiC (a-Si1−xCx) nano-shell (nano-film) can be directly transformed into graphene by using chlorination method under very mild reaction conditions of relative low temperature (800°C) and the ambient pressure in chlorine (Cl2) atmosphere. Therefore, our finding, the direct transformation of a-Si1−xCx into graphene under much milder condition, will open a door to apply this new method to the large-scale production of graphene at low costs. PMID:23359349
Inexpensive Device for Demonstrating Rock Slope Failure and Other Collapse Phenomena.
ERIC Educational Resources Information Center
Stimpson, B.
1980-01-01
Describes an inexpensive modeling technique for demonstrating large-scale displacement phenomena in rock masses, such as slope collapse and failure of underground openings. Excavation of the model material occurs through openings made in the polyurethane foam in the correct excavation sequence. (Author/SA)
Contact information and guidances for each state and selected territories's environmental agencies and programs relevant to large-scale residential demolition including asbestos, lead, and open burning.
Coronal hole evolution by sudden large scale changes
NASA Technical Reports Server (NTRS)
Nolte, J. T.; Gerassimenko, M.; Krieger, A. S.; Solodyna, C. V.
1978-01-01
Sudden shifts in coronal-hole boundaries observed by the S-054 X-ray telescope on Skylab between May and November, 1973, within 1 day of CMP of the holes, at latitudes not exceeding 40 deg, are compared with the long-term evolution of coronal-hole area. It is found that large-scale shifts in boundary locations can account for most if not all of the evolution of coronal holes. The temporal and spatial scales of these large-scale changes imply that they are the results of a physical process occurring in the corona. It is concluded that coronal holes evolve by magnetic-field lines' opening when the holes are growing, and by fields' closing as the holes shrink.
Open star clusters and Galactic structure
NASA Astrophysics Data System (ADS)
Joshi, Yogesh C.
2018-04-01
In order to understand the Galactic structure, we perform a statistical analysis of the distribution of various cluster parameters based on an almost complete sample of Galactic open clusters yet available. The geometrical and physical characteristics of a large number of open clusters given in the MWSC catalogue are used to study the spatial distribution of clusters in the Galaxy and determine the scale height, solar offset, local mass density and distribution of reddening material in the solar neighbourhood. We also explored the mass-radius and mass-age relations in the Galactic open star clusters. We find that the estimated parameters of the Galactic disk are largely influenced by the choice of cluster sample.
2012-10-01
using the open-source code Large-scale Atomic/Molecular Massively Parallel Simulator ( LAMMPS ) (http://lammps.sandia.gov) (23). The commercial...parameters are proprietary and cannot be ported to the LAMMPS 4 simulation code. In our molecular dynamics simulations at the atomistic resolution, we...IBI iterative Boltzmann inversion LAMMPS Large-scale Atomic/Molecular Massively Parallel Simulator MAPS Materials Processes and Simulations MS
A Large-scale Plume in an X-class Solar Flare
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fleishman, Gregory D.; Nita, Gelu M.; Gary, Dale E.
Ever-increasing multi-frequency imaging of solar observations suggests that solar flares often involve more than one magnetic fluxtube. Some of the fluxtubes are closed, while others can contain open fields. The relative proportion of nonthermal electrons among those distinct loops is highly important for understanding energy release, particle acceleration, and transport. The access of nonthermal electrons to the open field is also important because the open field facilitates the solar energetic particle (SEP) escape from the flaring site, and thus controls the SEP fluxes in the solar system, both directly and as seed particles for further acceleration. The large-scale fluxtubes aremore » often filled with a tenuous plasma, which is difficult to detect in either EUV or X-ray wavelengths; however, they can dominate at low radio frequencies, where a modest component of nonthermal electrons can render the source optically thick and, thus, bright enough to be observed. Here we report the detection of a large-scale “plume” at the impulsive phase of an X-class solar flare, SOL2001-08-25T16:23, using multi-frequency radio data from Owens Valley Solar Array. To quantify the flare’s spatial structure, we employ 3D modeling utilizing force-free-field extrapolations from the line of sight SOHO /MDI magnetograms with our modeling tool GX-Simulator. We found that a significant fraction of the nonthermal electrons that accelerated at the flare site low in the corona escapes to the plume, which contains both closed and open fields. We propose that the proportion between the closed and open fields at the plume is what determines the SEP population escaping into interplanetary space.« less
Fast Open-World Person Re-Identification.
Zhu, Xiatian; Wu, Botong; Huang, Dongcheng; Zheng, Wei-Shi
2018-05-01
Existing person re-identification (re-id) methods typically assume that: 1) any probe person is guaranteed to appear in the gallery target population during deployment (i.e., closed-world) and 2) the probe set contains only a limited number of people (i.e., small search scale). Both assumptions are artificial and breached in real-world applications, since the probe population in target people search can be extremely vast in practice due to the ambiguity of probe search space boundary. Therefore, it is unrealistic that any probe person is assumed as one target people, and a large-scale search in person images is inherently demanded. In this paper, we introduce a new person re-id search setting, called large scale open-world (LSOW) re-id, characterized by huge size probe images and open person population in search thus more close to practical deployments. Under LSOW, the under-studied problem of person re-id efficiency is essential in addition to that of commonly studied re-id accuracy. We, therefore, develop a novel fast person re-id method, called Cross-view Identity Correlation and vErification (X-ICE) hashing, for joint learning of cross-view identity representation binarisation and discrimination in a unified manner. Extensive comparative experiments on three large-scale benchmarks have been conducted to validate the superiority and advantages of the proposed X-ICE method over a wide range of the state-of-the-art hashing models, person re-id methods, and their combinations.
NASA Astrophysics Data System (ADS)
Moritz, R. E.
2005-12-01
The properties, distribution and temporal variation of sea-ice are reviewed for application to problems of ice-atmosphere chemical processes. Typical vertical structure of sea-ice is presented for different ice types, including young ice, first-year ice and multi-year ice, emphasizing factors relevant to surface chemistry and gas exchange. Time average annual cycles of large scale variables are presented, including ice concentration, ice extent, ice thickness and ice age. Spatial and temporal variability of these large scale quantities is considered on time scales of 1-50 years, emphasizing recent and projected changes in the Arctic pack ice. The amount and time evolution of open water and thin ice are important factors that influence ocean-ice-atmosphere chemical processes. Observations and modeling of the sea-ice thickness distribution function are presented to characterize the range of variability in open water and thin ice.
ERIC Educational Resources Information Center
Godwin, Stephen; McAndrew, Patrick; Santos, Andreia
2008-01-01
Web-enabled technology is now being applied on a large scale. In this paper we look at open access provision of teaching and learning leading to many users with varying patterns and motivations for use. This has provided us with a research challenge to find methods that help us understand and explain such initiatives. We describe ways to model the…
OpenSim: A Flexible Distributed Neural Network Simulator with Automatic Interactive Graphics.
Jarosch, Andreas; Leber, Jean Francois
1997-06-01
An object-oriented simulator called OpenSim is presented that achieves a high degree of flexibility by relying on a small set of building blocks. The state variables and algorithms put in this framework can easily be accessed through a command shell. This allows one to distribute a large-scale simulation over several workstations and to generate the interactive graphics automatically. OpenSim opens new possibilities for cooperation among Neural Network researchers. Copyright 1997 Elsevier Science Ltd.
Usage Patterns of Open Genomic Data
ERIC Educational Resources Information Center
Xia, Jingfeng; Liu, Ying
2013-01-01
This paper uses Genome Expression Omnibus (GEO), a data repository in biomedical sciences, to examine the usage patterns of open data repositories. It attempts to identify the degree of recognition of data reuse value and understand how e-science has impacted a large-scale scholarship. By analyzing a list of 1,211 publications that cite GEO data…
Quality Assurance in Asian Open and Distance Learning: Policies and Implementation
ERIC Educational Resources Information Center
Darojat, Ojat; Nilson, Michelle; Kaufman, David
2015-01-01
Open universities have emerged as an innovative pillar in the expansion of access to higher education participation, with single-mode distance education providers broadening access in many countries through economies of scale supported by large enrolments. These models raise questions about the quality of education provided. This paper reports on…
Incentivizing the Production and Use of Open Educational Resources in Higher Education Institutions
ERIC Educational Resources Information Center
Annand, David; Jensen, Tilly
2017-01-01
Substituting open educational resources (OER) for commercially-produced textbooks results in demonstrable cost savings for students in most higher education institutions. Yet OER are still not widely used, and progress toward large-scale adoption in most colleges and universities has been slow. This article reviews the literature informing…
Comparing Learner Community Behavior in Multiple Presentations of a Massive Open Online Course
ERIC Educational Resources Information Center
Gallagher, Silvia Elena; Savage, Timothy
2015-01-01
Massive Online Open Courses (MOOCs) can create large scale communities of learners who collaborate, interact and discuss learning materials and activities. MOOCs are often delivered multiple times with similar content to different cohorts of learners. However, research into the differences of learner communication, behavior and expectation between…
Comparing Learner Community Behavior in Multiple Presentations of a Massive Open Online Course
ERIC Educational Resources Information Center
Gallagher, Silvia Elena; Savage, Timothy
2016-01-01
Massive Online Open Courses (MOOCs) can create large scale communities of learners who collaborate, interact and discuss learning materials and activities. MOOCs are often delivered multiple times with similar content to different cohorts of learners. However, research into the differences of learner communication, behavior and expectation between…
Bringing Open Educational Practice to a Research-Intensive University: Prospects and Challenges
ERIC Educational Resources Information Center
Masterman, Elizabeth
2016-01-01
This article describes a small-scale study that explored the relationship between the pedagogical practices characterised as "open" and the existing model of undergraduate teaching and learning at a large research-intensive university (RIU). The aim was to determine the factors that might enable (conversely impede) the greater uptake of…
Neural ensemble communities: Open-source approaches to hardware for large-scale electrophysiology
Siegle, Joshua H.; Hale, Gregory J.; Newman, Jonathan P.; Voigts, Jakob
2014-01-01
One often-overlooked factor when selecting a platform for large-scale electrophysiology is whether or not a particular data acquisition system is “open” or “closed”: that is, whether or not the system’s schematics and source code are available to end users. Open systems have a reputation for being difficult to acquire, poorly documented, and hard to maintain. With the arrival of more powerful and compact integrated circuits, rapid prototyping services, and web-based tools for collaborative development, these stereotypes must be reconsidered. We discuss some of the reasons why multichannel extracellular electrophysiology could benefit from open-source approaches and describe examples of successful community-driven tool development within this field. In order to promote the adoption of open-source hardware and to reduce the need for redundant development efforts, we advocate a move toward standardized interfaces that connect each element of the data processing pipeline. This will give researchers the flexibility to modify their tools when necessary, while allowing them to continue to benefit from the high-quality products and expertise provided by commercial vendors. PMID:25528614
NASA Technical Reports Server (NTRS)
Kashlinsky, A.
1992-01-01
It is shown here that, by using galaxy catalog correlation data as input, measurements of microwave background radiation (MBR) anisotropies should soon be able to test two of the inflationary scenario's most basic predictions: (1) that the primordial density fluctuations produced were scale-invariant and (2) that the universe is flat. They should also be able to detect anisotropies of large-scale structure formed by gravitational evolution of density fluctuations present at the last scattering epoch. Computations of MBR anisotropies corresponding to the minimum of the large-scale variance of the MBR anisotropy are presented which favor an open universe with P(k) significantly different from the Harrison-Zeldovich spectrum predicted by most inflationary models.
Hao, Shijie; Cui, Lishan; Wang, Hua; ...
2016-02-10
Crystals held at ultrahigh elastic strains and stresses may exhibit exceptional physical and chemical properties. Individual metallic nanowires can sustain ultra-large elastic strains of 4-7%. However, retaining elastic strains of such magnitude in kilogram-scale nanowires is challenging. Here, we find that under active load, ~5.6% elastic strain can be achieved in Nb nanowires in a composite material. Moreover, large tensile (2.8%) and compressive (-2.4%) elastic strains can be retained in kilogram-scale Nb nanowires when the composite is unloaded to a free-standing condition. It is then demonstrated that the retained tensile elastic strains of Nb nanowires significantly increase their superconducting transitionmore » temperature and critical magnetic fields, corroborating ab initio calculations based on BCS theory. This free-standing nanocomposite design paradigm opens new avenues for retaining ultra-large elastic strains in great quantities of nanowires and elastic-strain-engineering at industrial scale.« less
Large-Scale Conformational Dynamics Control H5N1 Influenza Polymerase PB2 Binding to Importin α.
Delaforge, Elise; Milles, Sigrid; Bouvignies, Guillaume; Bouvier, Denis; Boivin, Stephane; Salvi, Nicola; Maurin, Damien; Martel, Anne; Round, Adam; Lemke, Edward A; Jensen, Malene Ringkjøbing; Hart, Darren J; Blackledge, Martin
2015-12-09
Influenza A RNA polymerase complex is formed from three components, PA, PB1, and PB2. PB2 is independently imported into the nucleus prior to polymerase reconstitution. All crystallographic structures of the PB2 C-terminus (residues 536-759) reveal two globular domains, 627 and NLS, that form a tightly packed heterodimer. The molecular basis of the affinity of 627-NLS for importins remained unclear from these structures, apparently requiring large-scale conformational changes prior to importin binding. Using a combination of solution-state NMR, small-angle neutron scattering, small-angle X-ray scattering (SAXS), and Förster resonance energy transfer (FRET), we show that 627-NLS populates a temperature-dependent dynamic equilibrium between closed and open states. The closed state is stabilized by a tripartite salt bridge involving the 627-NLS interface and the linker, that becomes flexible in the open state, with 627 and NLS dislocating into a highly dynamic ensemble. Activation enthalpies and entropies associated with the rupture of this interface were derived from simultaneous analysis of temperature-dependent chemical exchange saturation transfer measurements, revealing a strong temperature dependence of both open-state population and exchange rate. Single-molecule FRET and SAXS demonstrate that only the open-form is capable of binding to importin α and that, upon binding, the 627 domain samples a dynamic conformational equilibrium in the vicinity of the C-terminus of importin α. This intrinsic large-scale conformational flexibility therefore enables 627-NLS to bind importin through conformational selection from a temperature-dependent equilibrium comprising both functional forms of the protein.
Parameterization and scaling of arctic ice conditions in the context of ice-atmospheric processes
NASA Technical Reports Server (NTRS)
Barry, R. G.; Steffen, K.; Heinrichs, J. F.; Key, J. R.; Maslanik, J. A.; Serreze, M. C.; Weaver, R. L.
1995-01-01
The goals of this project are to observe how the open water/thin ice fraction in a high-concentration ice pack responds to different short-period atmospheric forcings, and how this response is represented in different scales of observation. The objectives can be summarized as follows: determine the feasibility and accuracy of ice concentration and ice typing by ERS-1 SAR backscatter data, and whether SAR data might be used to calibrate concentration estimates from optical and massive-microwave sensors; investigate methods to integrate SAR data with other satellite data for turbulent heat flux parameterization at the ocean/atmosphere interface; determine how the development and evolution of open water/thin ice areas within the interior ice pack vary under different atmospheric synoptic regimes; compare how open-water/thin ice fractions estimated from large-area divergence measurements differ from fractions determined by summing localized openings in the pack; relate these questions of scale and process to methods of observation, modeling, and averaging over time and space.
Modelling the large-scale redshift-space 3-point correlation function of galaxies
NASA Astrophysics Data System (ADS)
Slepian, Zachary; Eisenstein, Daniel J.
2017-08-01
We present a configuration-space model of the large-scale galaxy 3-point correlation function (3PCF) based on leading-order perturbation theory and including redshift-space distortions (RSD). This model should be useful in extracting distance-scale information from the 3PCF via the baryon acoustic oscillation method. We include the first redshift-space treatment of biasing by the baryon-dark matter relative velocity. Overall, on large scales the effect of RSD is primarily a renormalization of the 3PCF that is roughly independent of both physical scale and triangle opening angle; for our adopted Ωm and bias values, the rescaling is a factor of ˜1.8. We also present an efficient scheme for computing 3PCF predictions from our model, important for allowing fast exploration of the space of cosmological parameters in future analyses.
Application of Open-Source Enterprise Information System Modules: An Empirical Study
ERIC Educational Resources Information Center
Lee, Sang-Heui
2010-01-01
Although there have been a number of studies on large scale implementation of proprietary enterprise information systems (EIS), open-source software (OSS) for EIS has received limited attention in spite of its potential as a disruptive innovation. Cost saving is the main driver for adopting OSS among the other possible benefits including security…
Recommendations for open data science.
Gymrek, Melissa; Farjoun, Yossi
2016-01-01
Life science research increasingly relies on large-scale computational analyses. However, the code and data used for these analyses are often lacking in publications. To maximize scientific impact, reproducibility, and reuse, it is crucial that these resources are made publicly available and are fully transparent. We provide recommendations for improving the openness of data-driven studies in life sciences.
Historical open forest ecosystems in the Missouri Ozarks: reconstruction and restoration targets
Brice B. Hanberry; D. Todd Jones-Farrand; John M. Kabrick
2014-01-01
Current forests no longer resemble historical open forest ecosystems in the eastern United States. In the absence of representative forest ecosystems under a continuous surface fire regime at a large scale, reconstruction of historical landscapes can provide a reference for restoration efforts. For initial expert-assigned vegetation phases ranging from prairie to...
Improving Student Retention through Evidence Based Proactive Systems at the Open University (UK)
ERIC Educational Resources Information Center
Gibbs, Graham; Regan, Peter; Simpson, Ormond
2007-01-01
The Open University has been undertaking an extended initiative to improve student retention through enhanced support for at-risk students. This initiative has evolved through a series of stages from ad hoc small scale local interventions relying largely on tutors and student self-referral, to an institution-wide pro-active system implemented by…
NOAA's world-class weather and climate prediction center opens at
StumbleUpon Digg More Destinations NOAA's world-class weather and climate prediction center opens at currents and large-scale rain and snow storms. Billions of earth observations from around the world flow operations. Investing in this center is an investment in our human capital, serving as a world class facility
Visualization, documentation, analysis, and communication of large scale gene regulatory networks
Longabaugh, William J.R.; Davidson, Eric H.; Bolouri, Hamid
2009-01-01
Summary Genetic regulatory networks (GRNs) are complex, large-scale, and spatially and temporally distributed. These characteristics impose challenging demands on computational GRN modeling tools, and there is a need for custom modeling tools. In this paper, we report on our ongoing development of BioTapestry, an open source, freely available computational tool designed specifically for GRN modeling. We also outline our future development plans, and give some examples of current applications of BioTapestry. PMID:18757046
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gallarno, George; Rogers, James H; Maxwell, Don E
The high computational capability of graphics processing units (GPUs) is enabling and driving the scientific discovery process at large-scale. The world s second fastest supercomputer for open science, Titan, has more than 18,000 GPUs that computational scientists use to perform scientific simu- lations and data analysis. Understanding of GPU reliability characteristics, however, is still in its nascent stage since GPUs have only recently been deployed at large-scale. This paper presents a detailed study of GPU errors and their impact on system operations and applications, describing experiences with the 18,688 GPUs on the Titan supercom- puter as well as lessons learnedmore » in the process of efficient operation of GPUs at scale. These experiences are helpful to HPC sites which already have large-scale GPU clusters or plan to deploy GPUs in the future.« less
Reach-scale effects of riparian forest cover on urban stream ecosystems
Roy, A.H.; Faust, C.L.; Freeman, Mary C.; Meyer, J.L.
2005-01-01
We compared habitat and biota between paired open and forested reaches within five small streams (basin area 10?20 km2) in suburban catchments (9%?49% urban land cover) in the Piedmont of Georgia, USA. Stream reaches with open canopies were narrower than forested reaches (4.1 versus 5.0 m, respectively). There were no differences in habitat diversity (variation in velocity, depth, or bed particle size) between open and forested reaches. However, absence of local forest cover corresponded to decreased large wood and increased algal chlorophyll a standing crop biomass. These differences in basal food resources translated into higher densities of fishes in open (9.0 individuals?m?2) versus forested (4.9 individuals?m?2) reaches, primarily attributed to higher densities of the herbivore Campostoma oligolepis. Densities of terrestrial invertebrate inputs were higher in open reaches; however, trends suggested higher biomass of terrestrial inputs in forested reaches and a corresponding higher density of terrestrial prey consumed by water column feeding fishes. Reach-scale biotic integrity (macroinvertebrates, salamanders, and fishes) was largely unaffected by differences in canopy cover. In urbanizing areas where catchment land cover drives habitat and biotic quality, management practices that rely exclusively on forested riparian areas for stream protection are unlikely to be effective at maintaining ecosystem integrity.
Reversible Self-Actuated Thermo-Responsive Pore Membrane
Park, Younggeun; Gutierrez, Maria Paz; Lee, Luke P.
2016-01-01
Smart membranes, which can selectively control the transfer of light, air, humidity and temperature, are important to achieve indoor climate regulation. Even though reversible self-actuation of smart membranes is desirable in large-scale, reversible self-regulation remains challenging. Specifically, reversible 100% opening/closing of pore actuation showing accurate responsiveness, reproducibility and structural flexibility, including uniform structure assembly, is currently very difficult. Here, we report a reversible, thermo-responsive self-activated pore membrane that achieves opening and closing of pores. The reversible, self-actuated thermo-responsive pore membrane was fabricated with hybrid materials of poly (N-isopropylacrylamide), (PNIPAM) within polytetrafluoroethylene (PTFE) to form a multi-dimensional pore array. Using Multiphysics simulation of heat transfer and structural mechanics based on finite element analysis, we demonstrated that pore opening and closing dynamics can be self-activated at environmentally relevant temperatures. Temperature cycle characterizations of the pore structure revealed 100% opening ratio at T = 40 °C and 0% opening ratio at T = 20 °C. The flexibility of the membrane showed an accurate temperature-responsive function at a maximum bending angle of 45°. Addressing the importance of self-regulation, this reversible self-actuated thermo-responsive pore membrane will advance the development of future large-scale smart membranes needed for sustainable indoor climate control. PMID:27991563
Reversible Self-Actuated Thermo-Responsive Pore Membrane
NASA Astrophysics Data System (ADS)
Park, Younggeun; Gutierrez, Maria Paz; Lee, Luke P.
2016-12-01
Smart membranes, which can selectively control the transfer of light, air, humidity and temperature, are important to achieve indoor climate regulation. Even though reversible self-actuation of smart membranes is desirable in large-scale, reversible self-regulation remains challenging. Specifically, reversible 100% opening/closing of pore actuation showing accurate responsiveness, reproducibility and structural flexibility, including uniform structure assembly, is currently very difficult. Here, we report a reversible, thermo-responsive self-activated pore membrane that achieves opening and closing of pores. The reversible, self-actuated thermo-responsive pore membrane was fabricated with hybrid materials of poly (N-isopropylacrylamide), (PNIPAM) within polytetrafluoroethylene (PTFE) to form a multi-dimensional pore array. Using Multiphysics simulation of heat transfer and structural mechanics based on finite element analysis, we demonstrated that pore opening and closing dynamics can be self-activated at environmentally relevant temperatures. Temperature cycle characterizations of the pore structure revealed 100% opening ratio at T = 40 °C and 0% opening ratio at T = 20 °C. The flexibility of the membrane showed an accurate temperature-responsive function at a maximum bending angle of 45°. Addressing the importance of self-regulation, this reversible self-actuated thermo-responsive pore membrane will advance the development of future large-scale smart membranes needed for sustainable indoor climate control.
Image Harvest: an open-source platform for high-throughput plant image processing and analysis
Knecht, Avi C.; Campbell, Malachy T.; Caprez, Adam; Swanson, David R.; Walia, Harkamal
2016-01-01
High-throughput plant phenotyping is an effective approach to bridge the genotype-to-phenotype gap in crops. Phenomics experiments typically result in large-scale image datasets, which are not amenable for processing on desktop computers, thus creating a bottleneck in the image-analysis pipeline. Here, we present an open-source, flexible image-analysis framework, called Image Harvest (IH), for processing images originating from high-throughput plant phenotyping platforms. Image Harvest is developed to perform parallel processing on computing grids and provides an integrated feature for metadata extraction from large-scale file organization. Moreover, the integration of IH with the Open Science Grid provides academic researchers with the computational resources required for processing large image datasets at no cost. Image Harvest also offers functionalities to extract digital traits from images to interpret plant architecture-related characteristics. To demonstrate the applications of these digital traits, a rice (Oryza sativa) diversity panel was phenotyped and genome-wide association mapping was performed using digital traits that are used to describe different plant ideotypes. Three major quantitative trait loci were identified on rice chromosomes 4 and 6, which co-localize with quantitative trait loci known to regulate agronomically important traits in rice. Image Harvest is an open-source software for high-throughput image processing that requires a minimal learning curve for plant biologists to analyzephenomics datasets. PMID:27141917
Large scale obscuration and related climate effects open literature bibliography
DOE Office of Scientific and Technical Information (OSTI.GOV)
Russell, N.A.; Geitgey, J.; Behl, Y.K.
1994-05-01
Large scale obscuration and related climate effects of nuclear detonations first became a matter of concern in connection with the so-called ``Nuclear Winter Controversy`` in the early 1980`s. Since then, the world has changed. Nevertheless, concern remains about the atmospheric effects of nuclear detonations, but the source of concern has shifted. Now it focuses less on global, and more on regional effects and their resulting impacts on the performance of electro-optical and other defense-related systems. This bibliography reflects the modified interest.
NASA Astrophysics Data System (ADS)
Miles, B.; Chepudira, K.; LaBar, W.
2017-12-01
The Open Geospatial Consortium (OGC) SensorThings API (STA) specification, ratified in 2016, is a next-generation open standard for enabling real-time communication of sensor data. Building on over a decade of OGC Sensor Web Enablement (SWE) Standards, STA offers a rich data model that can represent a range of sensor and phenomena types (e.g. fixed sensors sensing fixed phenomena, fixed sensors sensing moving phenomena, mobile sensors sensing fixed phenomena, and mobile sensors sensing moving phenomena) and is data agnostic. Additionally, and in contrast to previous SWE standards, STA is developer-friendly, as is evident from its convenient JSON serialization, and expressive OData-based query language (with support for geospatial queries); with its Message Queue Telemetry Transport (MQTT), STA is also well-suited to efficient real-time data publishing and discovery. All these attributes make STA potentially useful for use in environmental monitoring sensor networks. Here we present Kinota(TM), an Open-Source NoSQL implementation of OGC SensorThings for large-scale high-resolution real-time environmental monitoring. Kinota, which roughly stands for Knowledge from Internet of Things Analyses, relies on Cassandra its underlying data store, which is a horizontally scalable, fault-tolerant open-source database that is often used to store time-series data for Big Data applications (though integration with other NoSQL or rational databases is possible). With this foundation, Kinota can scale to store data from an arbitrary number of sensors collecting data every 500 milliseconds. Additionally, Kinota architecture is very modular allowing for customization by adopters who can choose to replace parts of the existing implementation when desirable. The architecture is also highly portable providing the flexibility to choose between cloud providers like azure, amazon, google etc. The scalable, flexible and cloud friendly architecture of Kinota makes it ideal for use in next-generation large-scale and high-resolution real-time environmental monitoring networks used in domains such as hydrology, geomorphology, and geophysics, as well as management applications such as flood early warning, and regulatory enforcement.
A Web-based Distributed Voluntary Computing Platform for Large Scale Hydrological Computations
NASA Astrophysics Data System (ADS)
Demir, I.; Agliamzanov, R.
2014-12-01
Distributed volunteer computing can enable researchers and scientist to form large parallel computing environments to utilize the computing power of the millions of computers on the Internet, and use them towards running large scale environmental simulations and models to serve the common good of local communities and the world. Recent developments in web technologies and standards allow client-side scripting languages to run at speeds close to native application, and utilize the power of Graphics Processing Units (GPU). Using a client-side scripting language like JavaScript, we have developed an open distributed computing framework that makes it easy for researchers to write their own hydrologic models, and run them on volunteer computers. Users will easily enable their websites for visitors to volunteer sharing their computer resources to contribute running advanced hydrological models and simulations. Using a web-based system allows users to start volunteering their computational resources within seconds without installing any software. The framework distributes the model simulation to thousands of nodes in small spatial and computational sizes. A relational database system is utilized for managing data connections and queue management for the distributed computing nodes. In this paper, we present a web-based distributed volunteer computing platform to enable large scale hydrological simulations and model runs in an open and integrated environment.
Reopening Openness to Experience: A Network Analysis of Four Openness to Experience Inventories.
Christensen, Alexander P; Cotter, Katherine N; Silvia, Paul J
2018-05-10
Openness to Experience is a complex trait, the taxonomic structure of which has been widely debated. Previous research has provided greater clarity of its lower order structure by synthesizing facets across several scales related to Openness to Experience. In this study, we take a finer grained approach by investigating the item-level relations of four Openness to Experience inventories (Big Five Aspects Scale, HEXACO-100, NEO PI-3, and Woo et al.'s Openness to Experience Inventory), using a network science approach, which allowed items to form an emergent taxonomy of facets and aspects. Our results (N = 802) identified 10 distinct facets (variety-seeking, aesthetic appreciation, intellectual curiosity, diversity, openness to emotions, fantasy, imaginative, self-assessed intelligence, intellectual interests, and nontraditionalism) that largely replicate previous findings as well as three higher order aspects: two that are commonly found in the literature (intellect and experiencing; i.e., openness), and one novel aspect (open-mindedness). In addition, we demonstrate that each Openness to Experience inventory offers a unique conceptualization of the trait, and that some inventories provide broader coverage of the network space than others. Our findings establish a broader consensus of Openness to Experience at the aspect and facet level, which has important implications for researchers and the Openness to Experience inventories they use.
Trends in sea ice cover within habitats used by bowhead whales in the western Arctic.
Moore, Sue E; Laidre, Kristin L
2006-06-01
We examined trends in sea ice cover between 1979 and 2002 in four months (March, June, September, and November) for four large (approximately 100,000 km2) and 12 small (approximately 10,000 km2) regions of the western Arctic in habitats used by bowhead whales (Balaena mysticetus). Variation in open water with year was significant in all months except March, but interactions between region and year were not. Open water increased in both large and small regions, but trends were weak with least-squares regression accounting for < or =34% of the total variation. In large regions, positive trends in open water were strongest in September. Linear fits were poor, however, even in the East Siberian, Chukchi, and Beaufort seas, where basin-scale analyses have emphasized dramatic sea ice loss. Small regions also showed weak positive trends in open water and strong interannual variability. Open water increased consistently in five small regions where bowhead whales have been observed feeding or where oceanographic models predict prey entrainment, including: (1) June, along the northern Chukotka coast, near Wrangel Island, and along the Beaufort slope; (2) September, near Wrangel Island, the Barrow Arc, and the Chukchi Borderland; and (3) November, along the Barrow Arc. Conversely, there was very little consistent change in sea ice cover in four small regions considered winter refugia for bowhead whales in the northern Bering Sea, nor in two small regions that include the primary springtime migration corridor in the Chukchi Sea. The effects of sea ice cover on bowhead whale prey availability are unknown but can be modeled via production and advection pathways. Our conceptual model suggests that reductions in sea ice cover will increase prey availability along both pathways for this population. This analysis elucidates the variability inherent in the western Arctic marine ecosystem at scales relevant to bowhead whales and contrasts basin-scale depictions of extreme sea ice retreats, thinning, and wind-driven movements.
Evaluation Findings from High School Reform Efforts in Baltimore
ERIC Educational Resources Information Center
Smerdon, Becky; Cohen, Jennifer
2009-01-01
The Baltimore City Public School System (BCPSS) is one of the first urban districts in the country to undertake large-scale high school reform, phasing in small learning communities by opening new high schools and transforming large, comprehensive high schools into small high schools. With support from the Bill & Melinda Gates Foundation, a…
Large-scale forest composition influences northern goshawk nesting in Wisconsin
Deahn M. Donner; Dean Anderson; Daniel Eklund; Matthew St.Pierre
2013-01-01
The northern goshawk (Accipiter gentilis atricapillus) is a woodland raptor that uses a variety of forest types for nesting across its breeding range, but strongly depends on older forests with large trees and open understories. Goshawks may select nesting locations by maximizing the convergence of nesting and foraging habitats. Insights into goshawk...
A Phenomenology of Learning Large: The Tutorial Sphere of xMOOC Video Lectures
ERIC Educational Resources Information Center
Adams, Catherine; Yin, Yin; Vargas Madriz, Luis Francisco; Mullen, C. Scott
2014-01-01
The current discourse surrounding Massive Open Online Courses (MOOCs) is powerful. Despite their rapid and widespread deployment, research has yet to confirm or refute some of the bold claims rationalizing the popularity and efficacy of these large-scale virtual learning environments. Also, MOOCs' reputed disruptive, game-changing potential…
DistributedFBA.jl: High-level, high-performance flux balance analysis in Julia
DOE Office of Scientific and Technical Information (OSTI.GOV)
Heirendt, Laurent; Thiele, Ines; Fleming, Ronan M. T.
Flux balance analysis and its variants are widely used methods for predicting steady-state reaction rates in biochemical reaction networks. The exploration of high dimensional networks with such methods is currently hampered by software performance limitations. DistributedFBA.jl is a high-level, high-performance, open-source implementation of flux balance analysis in Julia. It is tailored to solve multiple flux balance analyses on a subset or all the reactions of large and huge-scale networks, on any number of threads or nodes. DistributedFBA.jl is a high-level, high-performance, open-source implementation of flux balance analysis in Julia. It is tailored to solve multiple flux balance analyses on amore » subset or all the reactions of large and huge-scale networks, on any number of threads or nodes.« less
DistributedFBA.jl: High-level, high-performance flux balance analysis in Julia
Heirendt, Laurent; Thiele, Ines; Fleming, Ronan M. T.
2017-01-16
Flux balance analysis and its variants are widely used methods for predicting steady-state reaction rates in biochemical reaction networks. The exploration of high dimensional networks with such methods is currently hampered by software performance limitations. DistributedFBA.jl is a high-level, high-performance, open-source implementation of flux balance analysis in Julia. It is tailored to solve multiple flux balance analyses on a subset or all the reactions of large and huge-scale networks, on any number of threads or nodes. DistributedFBA.jl is a high-level, high-performance, open-source implementation of flux balance analysis in Julia. It is tailored to solve multiple flux balance analyses on amore » subset or all the reactions of large and huge-scale networks, on any number of threads or nodes.« less
The structure and evolution of coronal holes
NASA Technical Reports Server (NTRS)
Timothy, A. F.; Krieger, A. S.; Vaiana, G. S.
1975-01-01
Soft X-ray observations of coronal holes are analyzed to determine the structure, temporal evolution, and rotational properties of those features as well as possible mechanisms which may account for their almost rigid rotational characteristics. It is shown that coronal holes are open features with a divergent magnetic-field configuration resulting from a particular large-scale magnetic-field topology. They are apparently formed when the successive emergence and dispersion of active-region fields produce a swath of unipolar field founded by fields of opposite polarity, and they die when large-scale field patterns emerge which significantly distort the original field configuration. Two types of holes are described (compact and elongated), and three possible rotation mechanisms are considered: a rigidly rotating subphotospheric phenomenon, a linking of high and low latitudes by closed field lines, and an interaction between moving coronal material and open field lines.
OMERO and Bio-Formats 5: flexible access to large bioimaging datasets at scale
NASA Astrophysics Data System (ADS)
Moore, Josh; Linkert, Melissa; Blackburn, Colin; Carroll, Mark; Ferguson, Richard K.; Flynn, Helen; Gillen, Kenneth; Leigh, Roger; Li, Simon; Lindner, Dominik; Moore, William J.; Patterson, Andrew J.; Pindelski, Blazej; Ramalingam, Balaji; Rozbicki, Emil; Tarkowska, Aleksandra; Walczysko, Petr; Allan, Chris; Burel, Jean-Marie; Swedlow, Jason
2015-03-01
The Open Microscopy Environment (OME) has built and released Bio-Formats, a Java-based proprietary file format conversion tool and OMERO, an enterprise data management platform under open source licenses. In this report, we describe new versions of Bio-Formats and OMERO that are specifically designed to support large, multi-gigabyte or terabyte scale datasets that are routinely collected across most domains of biological and biomedical research. Bio- Formats reads image data directly from native proprietary formats, bypassing the need for conversion into a standard format. It implements the concept of a file set, a container that defines the contents of multi-dimensional data comprised of many files. OMERO uses Bio-Formats to read files natively, and provides a flexible access mechanism that supports several different storage and access strategies. These new capabilities of OMERO and Bio-Formats make them especially useful for use in imaging applications like digital pathology, high content screening and light sheet microscopy that create routinely large datasets that must be managed and analyzed.
Visualization and modeling of smoke transport over landscape scales
Glenn P. Forney; William Mell
2007-01-01
Computational tools have been developed at the National Institute of Standards and Technology (NIST) for modeling fire spread and smoke transport. These tools have been adapted to address fire scenarios that occur in the wildland urban interface (WUI) over kilometer-scale distances. These models include the smoke plume transport model ALOFT (A Large Open Fire plume...
Biology-Inspired Distributed Consensus in Massively-Deployed Sensor Networks
NASA Technical Reports Server (NTRS)
Jones, Kennie H.; Lodding, Kenneth N.; Olariu, Stephan; Wilson, Larry; Xin, Chunsheng
2005-01-01
Promises of ubiquitous control of the physical environment by large-scale wireless sensor networks open avenues for new applications that are expected to redefine the way we live and work. Most of recent research has concentrated on developing techniques for performing relatively simple tasks in small-scale sensor networks assuming some form of centralized control. The main contribution of this work is to propose a new way of looking at large-scale sensor networks, motivated by lessons learned from the way biological ecosystems are organized. Indeed, we believe that techniques used in small-scale sensor networks are not likely to scale to large networks; that such large-scale networks must be viewed as an ecosystem in which the sensors/effectors are organisms whose autonomous actions, based on local information, combine in a communal way to produce global results. As an example of a useful function, we demonstrate that fully distributed consensus can be attained in a scalable fashion in massively deployed sensor networks where individual motes operate based on local information, making local decisions that are aggregated across the network to achieve globally-meaningful effects.
Pal, Abhro; Anupindi, Kameswararao; Delorme, Yann; Ghaisas, Niranjan; Shetty, Dinesh A; Frankel, Steven H
2014-07-01
In the present study, we performed large eddy simulation (LES) of axisymmetric, and 75% stenosed, eccentric arterial models with steady inflow conditions at a Reynolds number of 1000. The results obtained are compared with the direct numerical simulation (DNS) data (Varghese et al., 2007, "Direct Numerical Simulation of Stenotic Flows. Part 1. Steady Flow," J. Fluid Mech., 582, pp. 253-280). An inhouse code (WenoHemo) employing high-order numerical methods for spatial and temporal terms, along with a 2nd order accurate ghost point immersed boundary method (IBM) (Mark, and Vanwachem, 2008, "Derivation and Validation of a Novel Implicit Second-Order Accurate Immersed Boundary Method," J. Comput. Phys., 227(13), pp. 6660-6680) for enforcing boundary conditions on curved geometries is used for simulations. Three subgrid scale (SGS) models, namely, the classical Smagorinsky model (Smagorinsky, 1963, "General Circulation Experiments With the Primitive Equations," Mon. Weather Rev., 91(10), pp. 99-164), recently developed Vreman model (Vreman, 2004, "An Eddy-Viscosity Subgrid-Scale Model for Turbulent Shear Flow: Algebraic Theory and Applications," Phys. Fluids, 16(10), pp. 3670-3681), and the Sigma model (Nicoud et al., 2011, "Using Singular Values to Build a Subgrid-Scale Model for Large Eddy Simulations," Phys. Fluids, 23(8), 085106) are evaluated in the present study. Evaluation of SGS models suggests that the classical constant coefficient Smagorinsky model gives best agreement with the DNS data, whereas the Vreman and Sigma models predict an early transition to turbulence in the poststenotic region. Supplementary simulations are performed using Open source field operation and manipulation (OpenFOAM) ("OpenFOAM," http://www.openfoam.org/) solver and the results are inline with those obtained with WenoHemo.
ERIC Educational Resources Information Center
Scrivener, Susan; Coghlan, Erin
2011-01-01
Only one-third of all students who enter community colleges with the intent to earn a degree or certificate actually meet this goal within six years. MDRC launched the Opening Doors Demonstration in 2003--the first large-scale random assignment study in a community college setting--to tackle this problem. Partnering with six community colleges,…
ERIC Educational Resources Information Center
Bower, Kirsty; Sheppard, Nick; Bayjoo, Jennifer; Pease, Adele
2017-01-01
This practical article presents findings of a small scale study undertaken at a large U.K. University. The purpose of the study was to encourage academic engagement with Open Access (OA) and the Higher Education Funding Council for England (HEFCE) mandate with the measurable impact being increased engagement with the Repository and dissemination…
Research Notes - An Introduction to Openness and Evolvability Assessment
2016-08-01
importance of different business and technical characteristics that combine to achieve an open solution. The complexity of most large-scale systems of...process characteristic) Granularity of the architecture (size of functional blocks) Modularity (cohesion and coupling) Support for multiple...Description) OV-3 (Operational Information Exchange Matrix) SV-1 (Systems Interface Description) TV-1 ( Technical Standards Profile). Note that there
ERIC Educational Resources Information Center
Marques, Bertil P.; Carvalho, Piedade; Escudeiro, Paula; Barata, Ana; Silva, Ana; Queiros, Sandra
2017-01-01
Promoted by the significant increase of large scale internet access, many audiences have turned to the web and to its resources for learning and inspiration, with diverse sets of skills and intents. In this context, Multimedia Online Open Courses (MOOC) consist in learning models supported on user-friendly web tools that allow anyone with minimum…
Reproducible Large-Scale Neuroimaging Studies with the OpenMOLE Workflow Management System.
Passerat-Palmbach, Jonathan; Reuillon, Romain; Leclaire, Mathieu; Makropoulos, Antonios; Robinson, Emma C; Parisot, Sarah; Rueckert, Daniel
2017-01-01
OpenMOLE is a scientific workflow engine with a strong emphasis on workload distribution. Workflows are designed using a high level Domain Specific Language (DSL) built on top of Scala. It exposes natural parallelism constructs to easily delegate the workload resulting from a workflow to a wide range of distributed computing environments. OpenMOLE hides the complexity of designing complex experiments thanks to its DSL. Users can embed their own applications and scale their pipelines from a small prototype running on their desktop computer to a large-scale study harnessing distributed computing infrastructures, simply by changing a single line in the pipeline definition. The construction of the pipeline itself is decoupled from the execution context. The high-level DSL abstracts the underlying execution environment, contrary to classic shell-script based pipelines. These two aspects allow pipelines to be shared and studies to be replicated across different computing environments. Workflows can be run as traditional batch pipelines or coupled with OpenMOLE's advanced exploration methods in order to study the behavior of an application, or perform automatic parameter tuning. In this work, we briefly present the strong assets of OpenMOLE and detail recent improvements targeting re-executability of workflows across various Linux platforms. We have tightly coupled OpenMOLE with CARE, a standalone containerization solution that allows re-executing on a Linux host any application that has been packaged on another Linux host previously. The solution is evaluated against a Python-based pipeline involving packages such as scikit-learn as well as binary dependencies. All were packaged and re-executed successfully on various HPC environments, with identical numerical results (here prediction scores) obtained on each environment. Our results show that the pair formed by OpenMOLE and CARE is a reliable solution to generate reproducible results and re-executable pipelines. A demonstration of the flexibility of our solution showcases three neuroimaging pipelines harnessing distributed computing environments as heterogeneous as local clusters or the European Grid Infrastructure (EGI).
Reproducible Large-Scale Neuroimaging Studies with the OpenMOLE Workflow Management System
Passerat-Palmbach, Jonathan; Reuillon, Romain; Leclaire, Mathieu; Makropoulos, Antonios; Robinson, Emma C.; Parisot, Sarah; Rueckert, Daniel
2017-01-01
OpenMOLE is a scientific workflow engine with a strong emphasis on workload distribution. Workflows are designed using a high level Domain Specific Language (DSL) built on top of Scala. It exposes natural parallelism constructs to easily delegate the workload resulting from a workflow to a wide range of distributed computing environments. OpenMOLE hides the complexity of designing complex experiments thanks to its DSL. Users can embed their own applications and scale their pipelines from a small prototype running on their desktop computer to a large-scale study harnessing distributed computing infrastructures, simply by changing a single line in the pipeline definition. The construction of the pipeline itself is decoupled from the execution context. The high-level DSL abstracts the underlying execution environment, contrary to classic shell-script based pipelines. These two aspects allow pipelines to be shared and studies to be replicated across different computing environments. Workflows can be run as traditional batch pipelines or coupled with OpenMOLE's advanced exploration methods in order to study the behavior of an application, or perform automatic parameter tuning. In this work, we briefly present the strong assets of OpenMOLE and detail recent improvements targeting re-executability of workflows across various Linux platforms. We have tightly coupled OpenMOLE with CARE, a standalone containerization solution that allows re-executing on a Linux host any application that has been packaged on another Linux host previously. The solution is evaluated against a Python-based pipeline involving packages such as scikit-learn as well as binary dependencies. All were packaged and re-executed successfully on various HPC environments, with identical numerical results (here prediction scores) obtained on each environment. Our results show that the pair formed by OpenMOLE and CARE is a reliable solution to generate reproducible results and re-executable pipelines. A demonstration of the flexibility of our solution showcases three neuroimaging pipelines harnessing distributed computing environments as heterogeneous as local clusters or the European Grid Infrastructure (EGI). PMID:28381997
NASA Technical Reports Server (NTRS)
Over, Thomas, M.; Gupta, Vijay K.
1994-01-01
Under the theory of independent and identically distributed random cascades, the probability distribution of the cascade generator determines the spatial and the ensemble properties of spatial rainfall. Three sets of radar-derived rainfall data in space and time are analyzed to estimate the probability distribution of the generator. A detailed comparison between instantaneous scans of spatial rainfall and simulated cascades using the scaling properties of the marginal moments is carried out. This comparison highlights important similarities and differences between the data and the random cascade theory. Differences are quantified and measured for the three datasets. Evidence is presented to show that the scaling properties of the rainfall can be captured to the first order by a random cascade with a single parameter. The dependence of this parameter on forcing by the large-scale meteorological conditions, as measured by the large-scale spatial average rain rate, is investigated for these three datasets. The data show that this dependence can be captured by a one-to-one function. Since the large-scale average rain rate can be diagnosed from the large-scale dynamics, this relationship demonstrates an important linkage between the large-scale atmospheric dynamics and the statistical cascade theory of mesoscale rainfall. Potential application of this research to parameterization of runoff from the land surface and regional flood frequency analysis is briefly discussed, and open problems for further research are presented.
Toward Enhancing OpenMP's Work-Sharing Directives
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chapman, B M; Huang, L; Jin, H
2006-05-17
OpenMP provides a portable programming interface for shared memory parallel computers (SMPs). Although this interface has proven successful for small SMPs, it requires greater flexibility in light of the steadily growing size of individual SMPs and the recent advent of multithreaded chips. In this paper, we describe two application development experiences that exposed these expressivity problems in the current OpenMP specification. We then propose mechanisms to overcome these limitations, including thread subteams and thread topologies. Thus, we identify language features that improve OpenMP application performance on emerging and large-scale platforms while preserving ease of programming.
Prediction of aerodynamic tonal noise from open rotors
NASA Astrophysics Data System (ADS)
Sharma, Anupam; Chen, Hsuan-nien
2013-08-01
A numerical approach for predicting tonal aerodynamic noise from "open rotors" is presented. "Open rotor" refers to an engine architecture with a pair of counter-rotating propellers. Typical noise spectra from an open rotor consist of dominant tones, which arise due to both the steady loading/thickness and the aerodynamic interaction between the two bladerows. The proposed prediction approach utilizes Reynolds Averaged Navier-Stokes (RANS) Computational Fluid Dynamics (CFD) simulations to obtain near-field description of the noise sources. The near-to-far-field propagation is then carried out by solving the Ffowcs Williams-Hawkings equation. Since the interest of this paper is limited to tone noise, a linearized, frequency domain approach is adopted to solve the wake/vortex-blade interaction problem.This paper focuses primarily on the speed scaling of the aerodynamic tonal noise from open rotors. Even though there is no theoretical mode cut-off due to the absence of nacelle in open rotors, the far-field noise is a strong function of the azimuthal mode order. While the steady loading/thickness noise has circumferential modes of high order, due to the relatively large number of blades (≈10-12), the interaction noise typically has modes of small orders. The high mode orders have very low radiation efficiency and exhibit very strong scaling with Mach number, while the low mode orders show a relatively weaker scaling. The prediction approach is able to capture the speed scaling (observed in experiment) of the overall aerodynamic noise very well.
NASA Astrophysics Data System (ADS)
Rolph, Rebecca J.; Mahoney, Andrew R.; Walsh, John; Loring, Philip A.
2018-05-01
Using thresholds of physical climate variables developed from community observations, together with two large-scale datasets, we have produced local indices directly relevant to the impacts of a reduced sea ice cover on Alaska coastal communities. The indices include the number of false freeze-ups defined by transient exceedances of ice concentration prior to a corresponding exceedance that persists, false break-ups
, timing of freeze-up and break-up, length of the open water duration, number of days when the winds preclude hunting via boat (wind speed threshold exceedances), the number of wind events conducive to geomorphological work or damage to infrastructure from ocean waves, and the number of these wind events with on- and along-shore components promoting water setup along the coastline. We demonstrate how community observations can inform use of large-scale datasets to derive these locally relevant indices. The two primary large-scale datasets are the Historical Sea Ice Atlas for Alaska and the atmospheric output from a regional climate model used to downscale the ERA-Interim atmospheric reanalysis. We illustrate the variability and trends of these indices by application to the rural Alaska communities of Kotzebue, Shishmaref, and Utqiaġvik (previously Barrow), although the same procedure and metrics can be applied to other coastal communities. Over the 1979-2014 time period, there has been a marked increase in the number of combined false freeze-ups and false break-ups as well as the number of days too windy for hunting via boat for all three communities, especially Utqiaġvik. At Utqiaġvik, there has been an approximate tripling of the number of wind events conducive to coastline erosion from 1979 to 2014. We have also found a delay in freeze-up and earlier break-up, leading to a lengthened open water period for all of the communities examined.
Biomass production of multipopulation microalgae in open air pond for biofuel potential.
Selvakumar, P; Umadevi, K
2016-04-01
Biodiesel gains attention as it is made from renewable resources and has considerable environmental benefits. The present investigation has focused on large scale cultivation of multipopulation microalgae in open air pond using natural sea water without any additional nutritive supplements for low cost biomass production as a possible source of biofuel in large scale. Open air algal pond attained average chlorophyll concentration of 11.01 µg/L with the maximum of 43.65 µg/L as well as a higher lipid concentration of 18% (w/w) with lipid content 9.3 mg/L on the 10th day of the culture; and maximum biomass of 0.36 g/L on the 7th day of the culture. Composition analysis of fatty acid methyl ester (FAME) was performed by gas chromatography and mass spectrometry (GCMS). Multipopulation of algal biomass had 18% of total lipid content with 55% of total saturated fatty acids (SFA), 35.3% of monounsaturated fatty acids (MUFA) and 9.7% of polyunsaturated fatty acids (PUFA), revealing a potential source of biofuel production at low cost.
Unstructured-grid coastal ocean modelling in Southern Adriatic and Northern Ionian Seas
NASA Astrophysics Data System (ADS)
Federico, Ivan; Pinardi, Nadia; Coppini, Giovanni; Oddo, Paolo
2016-04-01
The Southern Adriatic Northern Ionian coastal Forecasting System (SANIFS) is a short-term forecasting system based on unstructured grid approach. The model component is built on SHYFEM finite element three-dimensional hydrodynamic model. The operational chain exploits a downscaling approach starting from the Mediterranean oceanographic-scale model MFS (Mediterranean Forecasting System, operated by INGV). The implementation set-up has been designed to provide accurate hydrodynamics and active tracer processes in the coastal waters of Southern Eastern Italy (Apulia, Basilicata and Calabria regions), where the model is characterized by a variable resolution in range of 50-500 m. The horizontal resolution is also high in open-sea areas, where the elements size is approximately 3 km. The model is forced: (i) at the lateral open boundaries through a full nesting strategy directly with the MFS (temperature, salinity, non-tidal sea surface height and currents) and OTPS (tidal forcing) fields; (ii) at surface through two alternative atmospheric forcing datasets (ECMWF and COSMOME) via MFS-bulk-formulae. Given that the coastal fields are driven by a combination of both local/coastal and deep ocean forcings propagating along the shelf, the performance of SANIFS was verified first (i) at the large and shelf-coastal scales by comparing with a large scale CTD survey and then (ii) at the coastal-harbour scale by comparison with CTD, ADCP and tide gauge data. Sensitivity tests were performed on initialization conditions (mainly focused on spin-up procedures) and on surface boundary conditions by assessing the reliability of two alternative datasets at different horizontal resolution (12.5 and 7 km). The present work highlights how downscaling could improve the simulation of the flow field going from typical open-ocean scales of the order of several km to the coastal (and harbour) scales of tens to hundreds of meters.
Karthikeyan, M; Krishnan, S; Pandey, Anil Kumar; Bender, Andreas; Tropsha, Alexander
2008-04-01
We present the application of a Java remote method invocation (RMI) based open source architecture to distributed chemical computing. This architecture was previously employed for distributed data harvesting of chemical information from the Internet via the Google application programming interface (API; ChemXtreme). Due to its open source character and its flexibility, the underlying server/client framework can be quickly adopted to virtually every computational task that can be parallelized. Here, we present the server/client communication framework as well as an application to distributed computing of chemical properties on a large scale (currently the size of PubChem; about 18 million compounds), using both the Marvin toolkit as well as the open source JOELib package. As an application, for this set of compounds, the agreement of log P and TPSA between the packages was compared. Outliers were found to be mostly non-druglike compounds and differences could usually be explained by differences in the underlying algorithms. ChemStar is the first open source distributed chemical computing environment built on Java RMI, which is also easily adaptable to user demands due to its "plug-in architecture". The complete source codes as well as calculated properties along with links to PubChem resources are available on the Internet via a graphical user interface at http://moltable.ncl.res.in/chemstar/.
Open source database of images DEIMOS: extension for large-scale subjective image quality assessment
NASA Astrophysics Data System (ADS)
Vítek, Stanislav
2014-09-01
DEIMOS (Database of Images: Open Source) is an open-source database of images and video sequences for testing, verification and comparison of various image and/or video processing techniques such as compression, reconstruction and enhancement. This paper deals with extension of the database allowing performing large-scale web-based subjective image quality assessment. Extension implements both administrative and client interface. The proposed system is aimed mainly at mobile communication devices, taking into account advantages of HTML5 technology; it means that participants don't need to install any application and assessment could be performed using web browser. The assessment campaign administrator can select images from the large database and then apply rules defined by various test procedure recommendations. The standard test procedures may be fully customized and saved as a template. Alternatively the administrator can define a custom test, using images from the pool and other components, such as evaluating forms and ongoing questionnaires. Image sequence is delivered to the online client, e.g. smartphone or tablet, as a fully automated assessment sequence or viewer can decide on timing of the assessment if required. Environmental data and viewing conditions (e.g. illumination, vibrations, GPS coordinates, etc.), may be collected and subsequently analyzed.
Image Harvest: an open-source platform for high-throughput plant image processing and analysis.
Knecht, Avi C; Campbell, Malachy T; Caprez, Adam; Swanson, David R; Walia, Harkamal
2016-05-01
High-throughput plant phenotyping is an effective approach to bridge the genotype-to-phenotype gap in crops. Phenomics experiments typically result in large-scale image datasets, which are not amenable for processing on desktop computers, thus creating a bottleneck in the image-analysis pipeline. Here, we present an open-source, flexible image-analysis framework, called Image Harvest (IH), for processing images originating from high-throughput plant phenotyping platforms. Image Harvest is developed to perform parallel processing on computing grids and provides an integrated feature for metadata extraction from large-scale file organization. Moreover, the integration of IH with the Open Science Grid provides academic researchers with the computational resources required for processing large image datasets at no cost. Image Harvest also offers functionalities to extract digital traits from images to interpret plant architecture-related characteristics. To demonstrate the applications of these digital traits, a rice (Oryza sativa) diversity panel was phenotyped and genome-wide association mapping was performed using digital traits that are used to describe different plant ideotypes. Three major quantitative trait loci were identified on rice chromosomes 4 and 6, which co-localize with quantitative trait loci known to regulate agronomically important traits in rice. Image Harvest is an open-source software for high-throughput image processing that requires a minimal learning curve for plant biologists to analyzephenomics datasets. © The Author 2016. Published by Oxford University Press on behalf of the Society for Experimental Biology.
Global Distribution of Aerosols Over the Open Ocean as Derived from the Coastal Zone Color Scanner
NASA Technical Reports Server (NTRS)
Stegmann, P. M.; Tindale, N. W.
1999-01-01
Climatological maps of monthly mean aerosol radiance levels derived from the coastal zone color scanner (CZCS) were constructed for the world's ocean basins. This is the first study to use the 7.5.-year CZCS data set to examine the distribution and seasonality of aerosols over the open ocean on a global scale. Examination of our satellite images found the most prominent large-scale patch of elevated aerosol radiances in each month off the coast of northwest Africa. The well-known, large-scale plumes of elevated aerosol levels in the Arabian Sea, the northwest Pacific, and off the east coast of North America were also successfully captured. Radiance data were extracted from 13 major open-ocean zones, ranging from the subpolar to equatorial regions. Results from these extractions revealed the aerosol load in both subpolar and subtropical zones to be higher in the Northern Hemisphere than in the Southern Hemisphere. Aerosol radiances in the subtropics of both hemispheres were about 2 times higher in summer than in winter. In subpolar regions, aerosol radiances in late spring/early summer were almost 3 times that observed in winter. In general, the aerosol signal was higher during the warmer months and lower during the cooler months, irrespective of location. A comparison between our mean monthly aerosol radiance maps with mean monthly chlorophyll maps (also from CZCS) showed similar seasonality between aerosol and chlorophyll levels in the subpolar zones of both hemispheres, i.e., high levels in summer, low levels in winter. In the subtropics of both hemispheres, however, chlorophyll levels were higher in winter months which coincided with a depressed aerosol signal. Our results indicate that the near-IR channel on ocean color sensors can be used to successfully capture well-known, large-scale aerosol plumes on a global scale and that future ocean color sensors may provide a platform for long-term synoptic studies of combined aerosol-phytoplankton productivity interactions.
Large Scale GW Calculations on the Cori System
NASA Astrophysics Data System (ADS)
Deslippe, Jack; Del Ben, Mauro; da Jornada, Felipe; Canning, Andrew; Louie, Steven
The NERSC Cori system, powered by 9000+ Intel Xeon-Phi processors, represents one of the largest HPC systems for open-science in the United States and the world. We discuss the optimization of the GW methodology for this system, including both node level and system-scale optimizations. We highlight multiple large scale (thousands of atoms) case studies and discuss both absolute application performance and comparison to calculations on more traditional HPC architectures. We find that the GW method is particularly well suited for many-core architectures due to the ability to exploit a large amount of parallelism across many layers of the system. This work was supported by the U.S. Department of Energy, Office of Science, Basic Energy Sciences, Materials Sciences and Engineering Division, as part of the Computational Materials Sciences Program.
Porous microwells for geometry-selective, large-scale microparticle arrays
NASA Astrophysics Data System (ADS)
Kim, Jae Jung; Bong, Ki Wan; Reátegui, Eduardo; Irimia, Daniel; Doyle, Patrick S.
2017-01-01
Large-scale microparticle arrays (LSMAs) are key for material science and bioengineering applications. However, previous approaches suffer from trade-offs between scalability, precision, specificity and versatility. Here, we present a porous microwell-based approach to create large-scale microparticle arrays with complex motifs. Microparticles are guided to and pushed into microwells by fluid flow through small open pores at the bottom of the porous well arrays. A scaling theory allows for the rational design of LSMAs to sort and array particles on the basis of their size, shape, or modulus. Sequential particle assembly allows for proximal and nested particle arrangements, as well as particle recollection and pattern transfer. We demonstrate the capabilities of the approach by means of three applications: high-throughput single-cell arrays; microenvironment fabrication for neutrophil chemotaxis; and complex, covert tags by the transfer of an upconversion nanocrystal-laden LSMA.
Large-scale flow experiments for managing river systems
Konrad, Christopher P.; Olden, Julian D.; Lytle, David A.; Melis, Theodore S.; Schmidt, John C.; Bray, Erin N.; Freeman, Mary C.; Gido, Keith B.; Hemphill, Nina P.; Kennard, Mark J.; McMullen, Laura E.; Mims, Meryl C.; Pyron, Mark; Robinson, Christopher T.; Williams, John G.
2011-01-01
Experimental manipulations of streamflow have been used globally in recent decades to mitigate the impacts of dam operations on river systems. Rivers are challenging subjects for experimentation, because they are open systems that cannot be isolated from their social context. We identify principles to address the challenges of conducting effective large-scale flow experiments. Flow experiments have both scientific and social value when they help to resolve specific questions about the ecological action of flow with a clear nexus to water policies and decisions. Water managers must integrate new information into operating policies for large-scale experiments to be effective. Modeling and monitoring can be integrated with experiments to analyze long-term ecological responses. Experimental design should include spatially extensive observations and well-defined, repeated treatments. Large-scale flow manipulations are only a part of dam operations that affect river systems. Scientists can ensure that experimental manipulations continue to be a valuable approach for the scientifically based management of river systems.
Low Pressure Seeder Development for PIV in Large Scale Open Loop Wind Tunnels
NASA Astrophysics Data System (ADS)
Schmit, Ryan
2010-11-01
A low pressure seeding techniques have been developed for Particle Image Velocimetry (PIV) in large scale wind tunnel facilities was performed at the Subsonic Aerodynamic Research Laboratory (SARL) facility at Wright-Patterson Air Force Base. The SARL facility is an open loop tunnel with a 7 by 10 foot octagonal test section that has 56% optical access and the Mach number varies from 0.2 to 0.5. A low pressure seeder sprayer was designed and tested in the inlet of the wind tunnel. The seeder sprayer was designed to produce an even and uniform distribution of seed while reducing the seeders influence in the test section. ViCount Compact 5000 using Smoke Oil 180 was using as the seeding material. The results show that this low pressure seeder does produce streaky seeding but excellent PIV images are produced.
NASA Technical Reports Server (NTRS)
Soderman, Paul T.; Olsen, Larry E.
1990-01-01
An engineering feasibility study was made of aeroacoustic inserts designed for large-scale acoustic research on aircraft models in the 80 by 120 foot Wind Tunnel at NASA Ames Research Center. The advantages and disadvantages of likely designs were analyzed. Results indicate that the required maximum airspeed leads to the design of a particular insert. Using goals of 200, 150, and 100 knots airspeed, the analysis indicated a 30 x 60 ft open-jet test section, a 40 x 80 ft open jet test section, and a 70 x 100 ft closed test section with enhanced wall lining, respectively. The open-jet inserts would be composed of a nozzle, collector, diffuser, and acoutic wedges incorporated in the existing 80 x 120 test section. The closed test section would be composed of approximately 5 ft acoustic wedges covered by a porous plate attached to the test section walls of the existing 80 x 120. All designs would require a double row of acoustic vanes between the test section and fan drive to attenuate fan noise and, in the case of the open-jet designs, to control flow separation at the diffuser downstream end. The inserts would allow virtually anechoic acoustic studies of large helicopter models, jets, and V/STOL aircraft models in simulated flight. Model scale studies would be necessary to optimize the aerodynamic and acoustic performance of any of the designs. In all designs studied, the existing structure would have to be reinforced. Successful development of acoustically transparent walls, though not strictly necessary to the project, would lead to a porous-wall test section that could be substituted for any of the open-jet designs, and thereby eliminate many aerodynamic and acoustic problems characteristic of open-jet shear layers. The larger size of the facility would make installation and removal of the insert components difficult. Consequently, scheduling of the existing 80 x 120 aerodynamic test section and scheduling of the open-jet test section would likely be made on an annual or longer basis. The enhanced wall-lining insert would likely be permanent. Although the modifications are technically feasible, the economic practicality of the project was not evaluated.
NASA Astrophysics Data System (ADS)
Chatterjee, Tanmoy; Peet, Yulia T.
2018-03-01
Length scales of eddies involved in the power generation of infinite wind farms are studied by analyzing the spectra of the turbulent flux of mean kinetic energy (MKE) from large eddy simulations (LES). Large-scale structures with an order of magnitude bigger than the turbine rotor diameter (D ) are shown to have substantial contribution to wind power. Varying dynamics in the intermediate scales (D -10 D ) are also observed from a parametric study involving interturbine distances and hub height of the turbines. Further insight about the eddies responsible for the power generation have been provided from the scaling analysis of two-dimensional premultiplied spectra of MKE flux. The LES code is developed in a high Reynolds number near-wall modeling framework, using an open-source spectral element code Nek5000, and the wind turbines have been modelled using a state-of-the-art actuator line model. The LES of infinite wind farms have been validated against the statistical results from the previous literature. The study is expected to improve our understanding of the complex multiscale dynamics in the domain of large wind farms and identify the length scales that contribute to the power. This information can be useful for design of wind farm layout and turbine placement that take advantage of the large-scale structures contributing to wind turbine power.
NASA Astrophysics Data System (ADS)
Fonseca, R. A.; Vieira, J.; Fiuza, F.; Davidson, A.; Tsung, F. S.; Mori, W. B.; Silva, L. O.
2013-12-01
A new generation of laser wakefield accelerators (LWFA), supported by the extreme accelerating fields generated in the interaction of PW-Class lasers and underdense targets, promises the production of high quality electron beams in short distances for multiple applications. Achieving this goal will rely heavily on numerical modelling to further understand the underlying physics and identify optimal regimes, but large scale modelling of these scenarios is computationally heavy and requires the efficient use of state-of-the-art petascale supercomputing systems. We discuss the main difficulties involved in running these simulations and the new developments implemented in the OSIRIS framework to address these issues, ranging from multi-dimensional dynamic load balancing and hybrid distributed/shared memory parallelism to the vectorization of the PIC algorithm. We present the results of the OASCR Joule Metric program on the issue of large scale modelling of LWFA, demonstrating speedups of over 1 order of magnitude on the same hardware. Finally, scalability to over ˜106 cores and sustained performance over ˜2 P Flops is demonstrated, opening the way for large scale modelling of LWFA scenarios.
SLIDE - a web-based tool for interactive visualization of large-scale -omics data.
Ghosh, Soumita; Datta, Abhik; Tan, Kaisen; Choi, Hyungwon
2018-06-28
Data visualization is often regarded as a post hoc step for verifying statistically significant results in the analysis of high-throughput data sets. This common practice leaves a large amount of raw data behind, from which more information can be extracted. However, existing solutions do not provide capabilities to explore large-scale raw datasets using biologically sensible queries, nor do they allow user interaction based real-time customization of graphics. To address these drawbacks, we have designed an open-source, web-based tool called Systems-Level Interactive Data Exploration, or SLIDE to visualize large-scale -omics data interactively. SLIDE's interface makes it easier for scientists to explore quantitative expression data in multiple resolutions in a single screen. SLIDE is publicly available under BSD license both as an online version as well as a stand-alone version at https://github.com/soumitag/SLIDE. Supplementary Information are available at Bioinformatics online.
NASA Astrophysics Data System (ADS)
Pilone, D.; Cechini, M. F.; Mitchell, A.
2011-12-01
Earth Science applications typically deal with large amounts of data and high throughput rates, if not also high transaction rates. While Open Source is frequently used for smaller scientific applications, large scale, highly available systems frequently fall back to "enterprise" class solutions like Oracle RAC or commercial grade JEE Application Servers. NASA's Earth Observing System Data and Information System (EOSDIS) provides end-to-end capabilities for managing NASA's Earth science data from multiple sources - satellites, aircraft, field measurements, and various other programs. A core capability of EOSDIS, the Earth Observing System (EOS) Clearinghouse (ECHO), is a highly available search and order clearinghouse of over 100 million pieces of science data that has evolved from its early R&D days to a fully operational system. Over the course of this maturity ECHO has largely transitioned from commercial frameworks, databases, and operating systems to Open Source solutions...and in some cases, back. In this talk we discuss the progression of our technological solutions and our lessons learned in the areas of: ? High performance, large scale searching solutions ? GeoSpatial search capabilities and dealing with multiple coordinate systems ? Search and storage of variable format source (science) data ? Highly available deployment solutions ? Scalable (elastic) solutions to visual searching and image handling Throughout the evolution of the ECHO system we have had to evaluate solutions with respect to performance, cost, developer productivity, reliability, and maintainability in the context of supporting global science users. Open Source solutions have played a significant role in our architecture and development but several critical commercial components remain (or have been reinserted) to meet our operational demands.
Importance of rodents for hydrology: lessons learnt from various field experiments
NASA Astrophysics Data System (ADS)
van Schaik, Loes; Zangerlé, Anne; Schneider, Anne-Kathrin; Schröder, Boris; Eccard, Jana
2017-04-01
organisms are known to create soil macropores of different sizes and with varying extent and orientation: most commonly earthworms, rodents, moles and roots. Preferential flow through macropore networks is dynamic and typically occurs when short individual macropores become connected at the hillslope scale as the nodes between the macropores become wet. Large lateral macropores may contribute to rapid subsurface stormflow of water and solutes at hillslope scale and supply a significant part of the catchment scale discharge during high intensity rainfall events even under relatively dry catchment state. Outflow from soil pipes, especially in the valley bottom or along the banking near to streams, is frequently observed, however, it remains a challenge to measure the spatial distribution, extent and connectivity of macropores at hill slope scales. We hypothesize that local information on organism abundances may be used as an indicator for spatial variability in infiltration, water storage and fluxes at the small scale and that knowledge on the landscape scale spatial distribution of organisms can provide information on connectivity of macropores at hillslope scale. Here we summarize the lessons learnt during three years of measurements aimed at determining the influence of rodent burrows on soil hydrology in a meso-scale catchment. Within the Attert Catchment (297 km2) in Luxembourg we performed sprinkling experiments with a brilliant blue tracer on twelve plots, of which six directly above rodent burrow openings and six on a surface without a rodent burrow opening, in order to examine the influence of the burrow openings on the infiltration pattern. Then we tested the extent of flow through mice burrows in different forest types, with varying geology and slope, by supplying 5 Liters of water with brilliant blue tracer directly to 24 burrow openings at soil surface. We excavated the burrows to measure how far the water was transported laterally in the burrow. Though we have serendipitous evidence of lateral water flow through large macropores in deeper soil layers from other projects, with the experiments we performed with the purpose to characterize this, the water did not seem to infiltrate into the burrow openings at the soil surface at all and the infiltration pattern under burrows was not different from that in soils without these openings. The five liter of brilliant blue dyed water which we poured into burrow openings did not flow far into the burrows, it generally infiltrated straight away into the surrounding soil. These results seem to show that the infiltration of water to rodent macropores during high intensity events does not take place at the soil surface but rather through other macropores, e.g. earthworm channels, which connect to deeper lateral channels. Also the lateral flow of water through the rodent burrows is apparently more effective in the deeper soils, where we occasionally saw a burrow with completely blue walls but little infiltration into the surrounding matrix.
LAMMPS strong scaling performance optimization on Blue Gene/Q
DOE Office of Scientific and Technical Information (OSTI.GOV)
Coffman, Paul; Jiang, Wei; Romero, Nichols A.
2014-11-12
LAMMPS "Large-scale Atomic/Molecular Massively Parallel Simulator" is an open-source molecular dynamics package from Sandia National Laboratories. Significant performance improvements in strong-scaling and time-to-solution for this application on IBM's Blue Gene/Q have been achieved through computational optimizations of the OpenMP versions of the short-range Lennard-Jones term of the CHARMM force field and the long-range Coulombic interaction implemented with the PPPM (particle-particle-particle mesh) algorithm, enhanced by runtime parameter settings controlling thread utilization. Additionally, MPI communication performance improvements were made to the PPPM calculation by re-engineering the parallel 3D FFT to use MPICH collectives instead of point-to-point. Performance testing was done using anmore » 8.4-million atom simulation scaling up to 16 racks on the Mira system at Argonne Leadership Computing Facility (ALCF). Speedups resulting from this effort were in some cases over 2x.« less
NASA Astrophysics Data System (ADS)
Coupon, Jean; Leauthaud, Alexie; Kilbinger, Martin; Medezinski, Elinor
2017-07-01
SWOT (Super W Of Theta) computes two-point statistics for very large data sets, based on “divide and conquer” algorithms, mainly, but not limited to data storage in binary trees, approximation at large scale, parellelization (open MPI), and bootstrap and jackknife resampling methods “on the fly”. It currently supports projected and 3D galaxy auto and cross correlations, galaxy-galaxy lensing, and weighted histograms.
`Dhara': An Open Framework for Critical Zone Modeling
NASA Astrophysics Data System (ADS)
Le, P. V.; Kumar, P.
2016-12-01
Processes in the Critical Zone, which sustain terrestrial life, are tightly coupled across hydrological, physical, biological, chemical, pedological, geomorphological and ecological domains over both short and long timescales. Observations and quantification of the Earth's surface across these domains using emerging high resolution measurement technologies such as light detection and ranging (lidar) and hyperspectral remote sensing are enabling us to characterize fine scale landscape attributes over large spatial areas. This presents a unique opportunity to develop novel approaches to model the Critical Zone that can capture fine scale intricate dependencies across the different processes in 3D. The development of interdisciplinary tools that transcend individual disciplines and capture new levels of complexity and emergent properties is at the core of Critical Zone science. Here we introduce an open framework for high-performance computing model (`Dhara') for modeling complex processes in the Critical Zone. The framework is designed to be modular in structure with the aim to create uniform and efficient tools to facilitate and leverage process modeling. It also provides flexibility to maintain, collaborate, and co-develop additional components by the scientific community. We show the essential framework that simulates ecohydrologic dynamics, and surface - sub-surface coupling in 3D using hybrid parallel CPU-GPU. We demonstrate that the open framework in Dhara is feasible for detailed, multi-processes, and large-scale modeling of the Critical Zone, which opens up exciting possibilities. We will also present outcomes from a Modeling Summer Institute led by Intensively Managed Critical Zone Observatory (IMLCZO) with representation from several CZOs and international representatives.
Architecting for Large Scale Agile Software Development: A Risk-Driven Approach
2013-05-01
addressed aspect of scale in agile software development. Practices such as Scrum of Scrums are meant to address orchestration of multiple development...owner, Scrum master) have differing responsibilities from the roles in the existing phase-based waterfall program structures. Such differences may... Scrum . Communication with both internal and external stakeholders must be open and documentation should not be used as a substitute for communication
NASA Astrophysics Data System (ADS)
Appel, Marius; Lahn, Florian; Buytaert, Wouter; Pebesma, Edzer
2018-04-01
Earth observation (EO) datasets are commonly provided as collection of scenes, where individual scenes represent a temporal snapshot and cover a particular region on the Earth's surface. Using these data in complex spatiotemporal modeling becomes difficult as soon as data volumes exceed a certain capacity or analyses include many scenes, which may spatially overlap and may have been recorded at different dates. In order to facilitate analytics on large EO datasets, we combine and extend the geospatial data abstraction library (GDAL) and the array-based data management and analytics system SciDB. We present an approach to automatically convert collections of scenes to multidimensional arrays and use SciDB to scale computationally intensive analytics. We evaluate the approach in three study cases on national scale land use change monitoring with Landsat imagery, global empirical orthogonal function analysis of daily precipitation, and combining historical climate model projections with satellite-based observations. Results indicate that the approach can be used to represent various EO datasets and that analyses in SciDB scale well with available computational resources. To simplify analyses of higher-dimensional datasets as from climate model output, however, a generalization of the GDAL data model might be needed. All parts of this work have been implemented as open-source software and we discuss how this may facilitate open and reproducible EO analyses.
The effects of light, primary production, and temperature on bacterial production at Station ALOHA
NASA Astrophysics Data System (ADS)
Viviani, D. A.; Church, M. J.
2016-02-01
In the open oceans, bacterial metabolism is responsible for a large fraction of the movement of reduced carbon through these ecosystems. While broad meta-analyses suggest that factors such as temperature or primary production control rates of bacterial production over large geographic scales, to date little is known about how these factors influence variability in bacterial production in the open sea. Here we present two years of measurements of 3H-leucine incorporation, a proxy for bacterial production, at the open ocean field site of the Hawaii Ocean Time-series, Station ALOHA (22° 45'N, 158° 00'W). By examining 3H-leucine incorporation over monthly, daily, and hourly scales, this work provides insight into processes controlling bacterial growth in this persistently oligotrophic habitat. Rates of 3H-leucine incorporation were consistently 60% greater when measured in the light than in the dark, highlighting the importance of sunlight in fueling bacterial metabolism in this ecosystem. Over diel time scales, rates of 3H-leucine incorporation were quasi-sinusoidal, with rates in the light higher near midday, while rates in the dark were greatest after sunset. Depth-integrated (0 -125 m) rates of 3H-leucine incorporation in both light and dark were more variable ( 5- and 4-fold, respectively) than coincident measurements of primary production ( 2-fold). On average, rates of bacterial production averaged 2 and 4% of primary production (in the dark and light, respectively). At near-monthly time scales, rates of 3H-leucine incorporation in both light and dark were significantly related to temperature. Our results suggest that in the subtropical oligotrophic Pacific, bacterial production appears decoupled from primary production as a result of seasonal-scale variations in temperature and light.
Open source tools for large-scale neuroscience.
Freeman, Jeremy
2015-06-01
New technologies for monitoring and manipulating the nervous system promise exciting biology but pose challenges for analysis and computation. Solutions can be found in the form of modern approaches to distributed computing, machine learning, and interactive visualization. But embracing these new technologies will require a cultural shift: away from independent efforts and proprietary methods and toward an open source and collaborative neuroscience. Copyright © 2015 The Author. Published by Elsevier Ltd.. All rights reserved.
Cloud-Based Distributed Control of Unmanned Systems
2015-04-01
during mission execution. At best, the data is saved onto hard-drives and is accessible only by the local team. Data history in a form available and...following open source technologies: GeoServer, OpenLayers, PostgreSQL , and PostGIS are chosen to implement the back-end database and server. A brief...geospatial map data. 3. PostgreSQL : An SQL-compliant object-relational database that easily scales to accommodate large amounts of data - upwards to
Hybrid MPI+OpenMP Programming of an Overset CFD Solver and Performance Investigations
NASA Technical Reports Server (NTRS)
Djomehri, M. Jahed; Jin, Haoqiang H.; Biegel, Bryan (Technical Monitor)
2002-01-01
This report describes a two level parallelization of a Computational Fluid Dynamic (CFD) solver with multi-zone overset structured grids. The approach is based on a hybrid MPI+OpenMP programming model suitable for shared memory and clusters of shared memory machines. The performance investigations of the hybrid application on an SGI Origin2000 (O2K) machine is reported using medium and large scale test problems.
A method for data handling numerical results in parallel OpenFOAM simulations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Anton, Alin; Muntean, Sebastian
Parallel computational fluid dynamics simulations produce vast amount of numerical result data. This paper introduces a method for reducing the size of the data by replaying the interprocessor traffic. The results are recovered only in certain regions of interest configured by the user. A known test case is used for several mesh partitioning scenarios using the OpenFOAM toolkit{sup ®}[1]. The space savings obtained with classic algorithms remain constant for more than 60 Gb of floating point data. Our method is most efficient on large simulation meshes and is much better suited for compressing large scale simulation results than the regular algorithms.
NASA Astrophysics Data System (ADS)
Widyaningrum, E.; Gorte, B. G. H.
2017-05-01
LiDAR data acquisition is recognized as one of the fastest solutions to provide basis data for large-scale topographical base maps worldwide. Automatic LiDAR processing is believed one possible scheme to accelerate the large-scale topographic base map provision by the Geospatial Information Agency in Indonesia. As a progressive advanced technology, Geographic Information System (GIS) open possibilities to deal with geospatial data automatic processing and analyses. Considering further needs of spatial data sharing and integration, the one stop processing of LiDAR data in a GIS environment is considered a powerful and efficient approach for the base map provision. The quality of the automated topographic base map is assessed and analysed based on its completeness, correctness, quality, and the confusion matrix.
Friction Angles of Open-Graded Aggregates from Large-Scale Direct Shear Testing : TechBrief
DOT National Transportation Integrated Search
2013-07-08
State and local transportation agencies frequently use opengraded aggregates for wall, roadway, and bridge construction. The primary advantages of using this type of material in wall and abutment applications are ease of constructability, lighter in-...
Magnetic Helicity and the Solar Dynamo
NASA Technical Reports Server (NTRS)
Canfield, Richard C.
1997-01-01
The objective of this investigation is to open a new window into the solar dynamo, convection, and magnetic reconnection through measurement of the helicity density of magnetic fields in the photosphere and tracing of large-scale patterns of magnetic helicity in the corona.
Sharma, Parichit; Mantri, Shrikant S
2014-01-01
The function of a newly sequenced gene can be discovered by determining its sequence homology with known proteins. BLAST is the most extensively used sequence analysis program for sequence similarity search in large databases of sequences. With the advent of next generation sequencing technologies it has now become possible to study genes and their expression at a genome-wide scale through RNA-seq and metagenome sequencing experiments. Functional annotation of all the genes is done by sequence similarity search against multiple protein databases. This annotation task is computationally very intensive and can take days to obtain complete results. The program mpiBLAST, an open-source parallelization of BLAST that achieves superlinear speedup, can be used to accelerate large-scale annotation by using supercomputers and high performance computing (HPC) clusters. Although many parallel bioinformatics applications using the Message Passing Interface (MPI) are available in the public domain, researchers are reluctant to use them due to lack of expertise in the Linux command line and relevant programming experience. With these limitations, it becomes difficult for biologists to use mpiBLAST for accelerating annotation. No web interface is available in the open-source domain for mpiBLAST. We have developed WImpiBLAST, a user-friendly open-source web interface for parallel BLAST searches. It is implemented in Struts 1.3 using a Java backbone and runs atop the open-source Apache Tomcat Server. WImpiBLAST supports script creation and job submission features and also provides a robust job management interface for system administrators. It combines script creation and modification features with job monitoring and management through the Torque resource manager on a Linux-based HPC cluster. Use case information highlights the acceleration of annotation analysis achieved by using WImpiBLAST. Here, we describe the WImpiBLAST web interface features and architecture, explain design decisions, describe workflows and provide a detailed analysis.
NASA Astrophysics Data System (ADS)
Montañez, Isabel P.; Osleger, Dillon J.; Chen, Jitao; Wortham, Barbara E.; Stamm, Robert G.; Nemyrovska, Tamara I.; Griffin, Julie M.; Poletaev, Vladislav I.; Wardlaw, Bruce R.
2018-06-01
Reconstructions of paleo-seawater chemistry are largely inferred from biogenic records of epicontinental seas. Recent studies provide considerable evidence for large-scale spatial and temporal variability in the environmental dynamics of these semi-restricted seas that leads to the decoupling of epicontinental isotopic records from those of the open ocean. We present conodont apatite δ18OPO4 and 87Sr/86Sr records spanning 24 Myr of the late Mississippian through Pennsylvanian derived from the U-Pb calibrated cyclothemic succession of the Donets Basin, eastern Ukraine. On a 2 to 6 Myr-scale, systematic fluctuations in bioapatite δ18OPO4 and 87Sr/86Sr broadly follow major shifts in the Donets onlap-offlap history and inferred regional climate, but are distinct from contemporaneous more open-water δ18OPO4 and global seawater Sr isotope trends. A -1 to -6‰ offset in Donets δ18OPO4 values from those of more open-water conodonts and greater temporal variability in δ18OPO4 and 87Sr/86Sr records are interpreted to primarily record climatically driven changes in local environmental processes in the Donets sea. Systematic isotopic shifts associated with Myr-scale sea-level fluctuations, however, indicate an extrabasinal driver. We propose a mechanistic link to glacioeustasy through a teleconnection between high-latitude ice changes and atmospheric pCO2 and regional monsoonal circulation in the Donets region. Inferred large-magnitude changes in Donets seawater salinity and temperature, not archived in the more open-water or global contemporaneous records, indicate a modification of the global climate signal in the epicontinental sea through amplification or dampening of the climate signal by local and regional environmental processes. This finding of global climate change filtered through local processes has implications for the use of conodont δ18OPO4 and 87Sr/86Sr values as proxies of paleo-seawater composition, mean temperature, and glacioeustasy.
Montanez, Isabel P.; Osleger, Dillon J.; Chen, J.-H.; Wortham, Barbara E.; Stamm, Robert G.; Nemyrovska, Tamara I.; Griffin, Julie M.; Poletaev, Vladislav I.; Wardlaw, Bruce R.
2018-01-01
Reconstructions of paleo-seawater chemistry are largely inferred from biogenic records of epicontinental seas. Recent studies provide considerable evidence for large-scale spatial and temporal variability in the environmental dynamics of these semi-restricted seas that leads to the decoupling of epicontinental isotopic records from those of the open ocean. We present conodont apatite δ18OPO4 and 87Sr/86Sr records spanning 24 Myr of the late Mississippian through Pennsylvanian derived from the U–Pb calibrated cyclothemic succession of the Donets Basin, eastern Ukraine. On a 2 to 6 Myr-scale, systematic fluctuations in bioapatite δ18OPO4 and 87Sr/86Sr broadly follow major shifts in the Donets onlap–offlap history and inferred regional climate, but are distinct from contemporaneous more open-water δ18OPO4 and global seawater Sr isotope trends. A −1 to −6‰ offset in Donets δ18OPO4 values from those of more open-water conodonts and greater temporal variability in δ18OPO4 and 87Sr/86Sr records are interpreted to primarily record climatically driven changes in local environmental processes in the Donets sea. Systematic isotopic shifts associated with Myr-scale sea-level fluctuations, however, indicate an extrabasinal driver. We propose a mechanistic link to glacioeustasy through a teleconnection between high-latitude ice changes and atmospheric pCO2 and regional monsoonal circulation in the Donets region. Inferred large-magnitude changes in Donets seawater salinity and temperature, not archived in the more open-water or global contemporaneous records, indicate a modification of the global climate signal in the epicontinental sea through amplification or dampening of the climate signal by local and regional environmental processes. This finding of global climate change filtered through local processes has implications for the use of conodont δ18OPO4 and 87Sr/86Sr values as proxies of paleo-seawater composition, mean temperature, and glacioeustasy.
Sharma, Parichit; Mantri, Shrikant S.
2014-01-01
The function of a newly sequenced gene can be discovered by determining its sequence homology with known proteins. BLAST is the most extensively used sequence analysis program for sequence similarity search in large databases of sequences. With the advent of next generation sequencing technologies it has now become possible to study genes and their expression at a genome-wide scale through RNA-seq and metagenome sequencing experiments. Functional annotation of all the genes is done by sequence similarity search against multiple protein databases. This annotation task is computationally very intensive and can take days to obtain complete results. The program mpiBLAST, an open-source parallelization of BLAST that achieves superlinear speedup, can be used to accelerate large-scale annotation by using supercomputers and high performance computing (HPC) clusters. Although many parallel bioinformatics applications using the Message Passing Interface (MPI) are available in the public domain, researchers are reluctant to use them due to lack of expertise in the Linux command line and relevant programming experience. With these limitations, it becomes difficult for biologists to use mpiBLAST for accelerating annotation. No web interface is available in the open-source domain for mpiBLAST. We have developed WImpiBLAST, a user-friendly open-source web interface for parallel BLAST searches. It is implemented in Struts 1.3 using a Java backbone and runs atop the open-source Apache Tomcat Server. WImpiBLAST supports script creation and job submission features and also provides a robust job management interface for system administrators. It combines script creation and modification features with job monitoring and management through the Torque resource manager on a Linux-based HPC cluster. Use case information highlights the acceleration of annotation analysis achieved by using WImpiBLAST. Here, we describe the WImpiBLAST web interface features and architecture, explain design decisions, describe workflows and provide a detailed analysis. PMID:24979410
NASA Technical Reports Server (NTRS)
Kashlinsky, A.
1993-01-01
Modified cold dark matter (CDM) models were recently suggested to account for large-scale optical data, which fix the power spectrum on large scales, and the COBE results, which would then fix the bias parameter, b. We point out that all such models have deficit of small-scale power where density fluctuations are presently nonlinear, and should then lead to late epochs of collapse of scales M between 10 exp 9 - 10 exp 10 solar masses and (1-5) x 10 exp 14 solar masses. We compute the probabilities and comoving space densities of various scale objects at high redshifts according to the CDM models and compare these with observations of high-z QSOs, high-z galaxies and the protocluster-size object found recently by Uson et al. (1992) at z = 3.4. We show that the modified CDM models are inconsistent with the observational data on these objects. We thus suggest that in order to account for the high-z objects, as well as the large-scale and COBE data, one needs a power spectrum with more power on small scales than CDM models allow and an open universe.
Multigrid preconditioned conjugate-gradient method for large-scale wave-front reconstruction.
Gilles, Luc; Vogel, Curtis R; Ellerbroek, Brent L
2002-09-01
We introduce a multigrid preconditioned conjugate-gradient (MGCG) iterative scheme for computing open-loop wave-front reconstructors for extreme adaptive optics systems. We present numerical simulations for a 17-m class telescope with n = 48756 sensor measurement grid points within the aperture, which indicate that our MGCG method has a rapid convergence rate for a wide range of subaperture average slope measurement signal-to-noise ratios. The total computational cost is of order n log n. Hence our scheme provides for fast wave-front simulation and control in large-scale adaptive optics systems.
Cyclicity in Upper Mississippian Bangor Limestone, Blount County, Alabama
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bronner, R.L.
1988-01-01
The Upper Mississippian (Chesterian) Bangor Limestone in Alabama consists of a thick, complex sequence of carbonate platform deposits. A continuous core through the Bangor on Blount Mountain in north-central Alabama provides the opportunity to analyze the unit for cyclicity and to identify controls on vertical facies sequence. Lithologies from the core represent four general environments of deposition: (1) subwave-base, open marine, (2) shoal, (3) lagoon, and (4) peritidal. Analysis of the vertical sequence of lithologies in the core indicates the presence of eight large-scale cycles dominated by subtidal deposits, but defined on the basis of peritidal caps. These large-scale cyclesmore » can be subdivided into 16 small-scale cycles that may be entirely subtidal but illustrate upward shallowing followed by rapid deepening. Large-scale cycles range from 33 to 136 ft thick, averaging 68 ft; small-scale cycles range from 5 to 80 ft thick and average 34 ft. Small-scale cycles have an average duration of approximately 125,000 years, which is compatible with Milankovitch periodicity. The large-scale cycles have an average duration of approximately 250,000 years, which may simply reflect variations in amplitude of sea level fluctuation or the influence of tectonic subsidence along the southeastern margin of the North American craton.« less
Organic field effect transistor with ultra high amplification
NASA Astrophysics Data System (ADS)
Torricelli, Fabrizio
2016-09-01
High-gain transistors are essential for the large-scale circuit integration, high-sensitivity sensors and signal amplification in sensing systems. Unfortunately, organic field-effect transistors show limited gain, usually of the order of tens, because of the large contact resistance and channel-length modulation. Here we show organic transistors fabricated on plastic foils enabling unipolar amplifiers with ultra-gain. The proposed approach is general and opens up new opportunities for ultra-large signal amplification in organic circuits and sensors.
ERIC Educational Resources Information Center
Kemple, James J.
2015-01-01
In the first decade of the 21st century, the New York City (NYC) Department of Education implemented a set of large-scale and much debated high school reforms, which included closing large, low-performing schools, opening new small schools, and extending high school choice to students throughout the district. The school closure process was the…
ERIC Educational Resources Information Center
Kemple, James J.
2015-01-01
In the first decade of the 21st century, the New York City (NYC) Department of Education implemented a set of large-scale and much debated high school reforms, which included closing large, low-performing schools, opening new small schools, and extending high school choice to students throughout the district. The school closure process was the…
ERIC Educational Resources Information Center
Kemple, James J.
2015-01-01
In the first decade of the 21st century, the New York City (NYC) Department of Education implemented a set of large-scale and much debated high school reforms, which included closing large, low-performing schools, opening new small schools, and extending high school choice to students throughout the district. The school closure process was the…
Aftermath of the MOOC Wars: Can Commercial Vendors Support Creative Higher Education?
ERIC Educational Resources Information Center
Newfield, Christopher
2016-01-01
The large-scale massive open online course (xMOOC) rose to prominence in 2012-13 on the promise that its outcomes would be better and cheaper than those of face-to-face university instruction. By late 2013, xMOOC educational claims had been largely discredited, though policy interest in ed-tech carried on. What can we learn about the future of…
An overview of the Hadoop/MapReduce/HBase framework and its current applications in bioinformatics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Taylor, Ronald C.
Bioinformatics researchers are increasingly confronted with analysis of ultra large-scale data sets, a problem that will only increase at an alarming rate in coming years. Recent developments in open source software, that is, the Hadoop project and associated software, provide a foundation for scaling to petabyte scale data warehouses on Linux clusters, providing fault-tolerant parallelized analysis on such data using a programming style named MapReduce. An overview is given of the current usage within the bioinformatics community of Hadoop, a top-level Apache Software Foundation project, and of associated open source software projects. The concepts behind Hadoop and the associated HBasemore » project are defined, and current bioinformatics software that employ Hadoop is described. The focus is on next-generation sequencing, as the leading application area to date.« less
A Framework for Daylighting Optimization in Whole Buildings with OpenStudio
DOE Office of Scientific and Technical Information (OSTI.GOV)
None
2016-08-12
We present a toolkit and workflow for leveraging the OpenStudio (Guglielmetti et al. 2010) platform to perform daylighting analysis and optimization in a whole building energy modeling (BEM) context. We have re-implemented OpenStudio's integrated Radiance and EnergyPlus functionality as an OpenStudio Measure. The OpenStudio Radiance Measure works within the OpenStudio Application and Parametric Analysis Tool, as well as the OpenStudio Server large scale analysis framework, allowing a rigorous daylighting simulation to be performed on a single building model or potentially an entire population of programmatically generated models. The Radiance simulation results can automatically inform the broader building energy model, andmore » provide dynamic daylight metrics as a basis for decision. Through introduction and example, this paper illustrates the utility of the OpenStudio building energy modeling platform to leverage existing simulation tools for integrated building energy performance simulation, daylighting analysis, and reportage.« less
The Contextualization of Archetypes: Clinical Template Governance.
Pedersen, Rune; Ulriksen, Gro-Hilde; Ellingsen, Gunnar
2015-01-01
This paper is a status report from a large-scale openEHR-based EPR project from the North Norway Regional Health Authority. It concerns the standardization of a regional ICT portfolio and the ongoing development of a new process oriented EPR systems encouraged by the unfolding of a national repository for openEHR archetypes. Subject of interest; the contextualization of clinical templates is governed over multiple national boundaries which is complex due to the dependency of clinical resources. From the outset of this, we are interested in how local, regional, and national organizers maneuver to standardize while applying OpenEHR technology.
COBE DMR-normalized open inflation cold dark matter cosmogony
NASA Technical Reports Server (NTRS)
Gorski, Krzysztof M.; Ratra, Bharat; Sugiyama, Naoshi; Banday, Anthony J.
1995-01-01
A cut-sky orthogonal mode analysis of the 2 year COBE DMR 53 and 90 GHz sky maps (in Galactic coordinates) is used to determine the normalization of an open inflation model based on the cold dark matter (CDM) scenario. The normalized model is compared to measures of large-scale structure in the universe. Although the DMR data alone does not provide sufficient discriminative power to prefer a particular value of the mass density parameter, the open model appears to be reasonably consistent with observations when Omega(sub 0) is approximately 0.3-0.4 and merits further study.
Olszewski, John; Winona, Linda; Oshima, Kevin H
2005-04-01
The use of ultrafiltration as a concentration method to recover viruses from environmental waters was investigated. Two ultrafiltration systems (hollow fiber and tangential flow) in a large- (100 L) and small-scale (2 L) configuration were able to recover greater than 50% of multiple viruses (bacteriophage PP7 and T1 and poliovirus type 2) from varying water turbidities (10-157 nephelometric turbidity units (NTU)) simultaneously. Mean recoveries (n = 3) in ground and surface water by the large-scale hollow fiber ultrafiltration system (100 L) were comparable to recoveries observed in the small-scale system (2 L). Recovery of seeded viruses in highly turbid waters from small-scale tangential flow (2 L) (screen and open channel) and hollow fiber ultrafilters (2 L) (small pilot) were greater than 70%. Clogging occurred in the hollow fiber pencil module and when particulate concentrations exceeded 1.6 g/L and 5.5 g/L (dry mass) in the screen and open channel filters, respectively. The small pilot module was able to filter all concentrates without clogging. The small pilot hollow fiber ultrafilter was used to test recovery of seeded viruses from surface waters from different geographical regions in 10-L volumes. Recoveries >70% were observed from all locations.
Scale dependence of open c{\\bar{c}} and b{\\bar{b}} production in the low x region
NASA Astrophysics Data System (ADS)
Oliveira, E. G. de; Martin, A. D.; Ryskin, M. G.
2017-03-01
The `optimal' factorization scale μ _0 is calculated for open heavy quark production. We find that the optimal value is μ _F=μ _0˜eq 0.85√{p^2_T+m_Q^2} ; a choice which allows us to resum the double-logarithmic, (α _s ln μ ^2_F ln (1/x))^n corrections (enhanced at LHC energies by large values of ln (1/x)) and to move them into the incoming parton distributions, PDF(x,μ _0^2). Besides this result for the single inclusive cross section (corresponding to an observed heavy quark of transverse momentum p_T), we also determined the scale for processes where the acoplanarity can be measured; that is, events where the azimuthal angle between the quark and the antiquark may be determined experimentally. Moreover, we discuss the important role played by the 2→ 2 subprocesses, gg→ Q\\bar{Q} at NLO and higher orders. In summary, we achieve a better stability of the QCD calculations, so that the data on c{\\bar{c}} and b{\\bar{b}} production can be used to further constrain the gluons in the small x, relatively low scale, domain, where the uncertainties of the global analyses are large at present.
Taking Open Innovation to the Molecular Level - Strengths and Limitations.
Zdrazil, Barbara; Blomberg, Niklas; Ecker, Gerhard F
2012-08-01
The ever-growing availability of large-scale open data and its maturation is having a significant impact on industrial drug-discovery, as well as on academic and non-profit research. As industry is changing to an 'open innovation' business concept, precompetitive initiatives and strong public-private partnerships including academic research cooperation partners are gaining more and more importance. Now, the bioinformatics and cheminformatics communities are seeking for web tools which allow the integration of this large volume of life science datasets available in the public domain. Such a data exploitation tool would ideally be able to answer complex biological questions by formulating only one search query. In this short review/perspective, we outline the use of semantic web approaches for data and knowledge integration. Further, we discuss strengths and current limitations of public available data retrieval tools and integrated platforms.
Pal, Abhro; Anupindi, Kameswararao; Delorme, Yann; Ghaisas, Niranjan; Shetty, Dinesh A.; Frankel, Steven H.
2014-01-01
In the present study, we performed large eddy simulation (LES) of axisymmetric, and 75% stenosed, eccentric arterial models with steady inflow conditions at a Reynolds number of 1000. The results obtained are compared with the direct numerical simulation (DNS) data (Varghese et al., 2007, “Direct Numerical Simulation of Stenotic Flows. Part 1. Steady Flow,” J. Fluid Mech., 582, pp. 253–280). An inhouse code (WenoHemo) employing high-order numerical methods for spatial and temporal terms, along with a 2nd order accurate ghost point immersed boundary method (IBM) (Mark, and Vanwachem, 2008, “Derivation and Validation of a Novel Implicit Second-Order Accurate Immersed Boundary Method,” J. Comput. Phys., 227(13), pp. 6660–6680) for enforcing boundary conditions on curved geometries is used for simulations. Three subgrid scale (SGS) models, namely, the classical Smagorinsky model (Smagorinsky, 1963, “General Circulation Experiments With the Primitive Equations,” Mon. Weather Rev., 91(10), pp. 99–164), recently developed Vreman model (Vreman, 2004, “An Eddy-Viscosity Subgrid-Scale Model for Turbulent Shear Flow: Algebraic Theory and Applications,” Phys. Fluids, 16(10), pp. 3670–3681), and the Sigma model (Nicoud et al., 2011, “Using Singular Values to Build a Subgrid-Scale Model for Large Eddy Simulations,” Phys. Fluids, 23(8), 085106) are evaluated in the present study. Evaluation of SGS models suggests that the classical constant coefficient Smagorinsky model gives best agreement with the DNS data, whereas the Vreman and Sigma models predict an early transition to turbulence in the poststenotic region. Supplementary simulations are performed using Open source field operation and manipulation (OpenFOAM) (“OpenFOAM,” http://www.openfoam.org/) solver and the results are inline with those obtained with WenoHemo. PMID:24801556
On distributed wavefront reconstruction for large-scale adaptive optics systems.
de Visser, Cornelis C; Brunner, Elisabeth; Verhaegen, Michel
2016-05-01
The distributed-spline-based aberration reconstruction (D-SABRE) method is proposed for distributed wavefront reconstruction with applications to large-scale adaptive optics systems. D-SABRE decomposes the wavefront sensor domain into any number of partitions and solves a local wavefront reconstruction problem on each partition using multivariate splines. D-SABRE accuracy is within 1% of a global approach with a speedup that scales quadratically with the number of partitions. The D-SABRE is compared to the distributed cumulative reconstruction (CuRe-D) method in open-loop and closed-loop simulations using the YAO adaptive optics simulation tool. D-SABRE accuracy exceeds CuRe-D for low levels of decomposition, and D-SABRE proved to be more robust to variations in the loop gain.
Horiguchi, Hiromasa; Yasunaga, Hideo; Hashimoto, Hideki; Ohe, Kazuhiko
2012-12-22
Secondary use of large scale administrative data is increasingly popular in health services and clinical research, where a user-friendly tool for data management is in great demand. MapReduce technology such as Hadoop is a promising tool for this purpose, though its use has been limited by the lack of user-friendly functions for transforming large scale data into wide table format, where each subject is represented by one row, for use in health services and clinical research. Since the original specification of Pig provides very few functions for column field management, we have developed a novel system called GroupFilterFormat to handle the definition of field and data content based on a Pig Latin script. We have also developed, as an open-source project, several user-defined functions to transform the table format using GroupFilterFormat and to deal with processing that considers date conditions. Having prepared dummy discharge summary data for 2.3 million inpatients and medical activity log data for 950 million events, we used the Elastic Compute Cloud environment provided by Amazon Inc. to execute processing speed and scaling benchmarks. In the speed benchmark test, the response time was significantly reduced and a linear relationship was observed between the quantity of data and processing time in both a small and a very large dataset. The scaling benchmark test showed clear scalability. In our system, doubling the number of nodes resulted in a 47% decrease in processing time. Our newly developed system is widely accessible as an open resource. This system is very simple and easy to use for researchers who are accustomed to using declarative command syntax for commercial statistical software and Structured Query Language. Although our system needs further sophistication to allow more flexibility in scripts and to improve efficiency in data processing, it shows promise in facilitating the application of MapReduce technology to efficient data processing with large scale administrative data in health services and clinical research.
NASA Astrophysics Data System (ADS)
Kang, Jingoo; Keinonen, Tuula
2017-04-01
Since students have lost their interest in school science, several student-centered approaches, such as using topics that are relevant for students, inquiry-based learning, and discussion-based learning have been implemented to attract pupils into science. However, the effect of these approaches was usually measured in small-scale research, and thus, the large-scale evidence supporting student-centered approaches in general use is insufficient. Accordingly, this study aimed to investigate the effect of student-centered approaches on students' interest and achievement by analyzing a large-scale data set derived from Program for International Student Assessment (PISA) 2006, to add evidence for advocating these approaches in school science, and to generalize the effects on a large population. We used Finnish PISA 2006 data, which is the most recent data that measures science literacy and that contains relevant variables for the constructs of this study. As a consequence of the factor analyses, four teaching methods were grouped as student-centered approaches (relevant topic-based, open and guided inquiry-based, and discussion-based approaches in school science) from the Finnish PISA 2006 sample. The structural equation modeling result indicated that using topics relevant for students positively affected students' interest and achievement in science. Guided inquiry-based learning was also indicated as a strong positive predictor for students' achievement, and its effect was also positively associated with students' interest. On the other hand, open inquiry-based learning was indicated as a strong negative predictor for students' achievement, as was using discussion in school science. Implications and limitations of the study were discussed.
NASA Astrophysics Data System (ADS)
Wang, Kunpeng; Ji, Weidong; Zhang, Feifei; Yu, Wei; Zheng, Runqing
2018-02-01
This thesis, based on the closed reconstruction project of the coal storage yard of Shengli Power Plant which is affiliated to Sinopec Shengli Petroleum Administration, first makes an analysis on the significance of current dustfall reconstruction of open coal yard, then summarizes the methods widely adopted in the dustfall of large-scale open coal storage yard of current thermal power plant as well as their advantages and disadvantages, and finally focuses on this project, aiming at providing some reference and assistance to the future closed reconstruction project of open coal storage yard in thermal power plant.
3D reconstruction software comparison for short sequences
NASA Astrophysics Data System (ADS)
Strupczewski, Adam; Czupryński, BłaŻej
2014-11-01
Large scale multiview reconstruction is recently a very popular area of research. There are many open source tools that can be downloaded and run on a personal computer. However, there are few, if any, comparisons between all the available software in terms of accuracy on small datasets that a single user can create. The typical datasets for testing of the software are archeological sites or cities, comprising thousands of images. This paper presents a comparison of currently available open source multiview reconstruction software for small datasets. It also compares the open source solutions with a simple structure from motion pipeline developed by the authors from scratch with the use of OpenCV and Eigen libraries.
CImbinator: a web-based tool for drug synergy analysis in small- and large-scale datasets.
Flobak, Åsmund; Vazquez, Miguel; Lægreid, Astrid; Valencia, Alfonso
2017-08-01
Drug synergies are sought to identify combinations of drugs particularly beneficial. User-friendly software solutions that can assist analysis of large-scale datasets are required. CImbinator is a web-service that can aid in batch-wise and in-depth analyzes of data from small-scale and large-scale drug combination screens. CImbinator offers to quantify drug combination effects, using both the commonly employed median effect equation, as well as advanced experimental mathematical models describing dose response relationships. CImbinator is written in Ruby and R. It uses the R package drc for advanced drug response modeling. CImbinator is available at http://cimbinator.bioinfo.cnio.es , the source-code is open and available at https://github.com/Rbbt-Workflows/combination_index . A Docker image is also available at https://hub.docker.com/r/mikisvaz/rbbt-ci_mbinator/ . asmund.flobak@ntnu.no or miguel.vazquez@cnio.es. Supplementary data are available at Bioinformatics online. © The Author(s) 2017. Published by Oxford University Press.
Structure and modeling of turbulence
DOE Office of Scientific and Technical Information (OSTI.GOV)
Novikov, E.A.
The {open_quotes}vortex strings{close_quotes} scale l{sub s} {approximately} LRe{sup -3/10} (L-external scale, Re - Reynolds number) is suggested as a grid scale for the large-eddy simulation. Various aspects of the structure of turbulence and subgrid modeling are described in terms of conditional averaging, Markov processes with dependent increments and infinitely divisible distributions. The major request from the energy, naval, aerospace and environmental engineering communities to the theory of turbulence is to reduce the enormous number of degrees of freedom in turbulent flows to a level manageable by computer simulations. The vast majority of these degrees of freedom is in the small-scalemore » motion. The study of the structure of turbulence provides a basis for subgrid-scale (SGS) models, which are necessary for the large-eddy simulations (LES).« less
Zhao, Shanrong; Prenger, Kurt; Smith, Lance
2013-01-01
RNA-Seq is becoming a promising replacement to microarrays in transcriptome profiling and differential gene expression study. Technical improvements have decreased sequencing costs and, as a result, the size and number of RNA-Seq datasets have increased rapidly. However, the increasing volume of data from large-scale RNA-Seq studies poses a practical challenge for data analysis in a local environment. To meet this challenge, we developed Stormbow, a cloud-based software package, to process large volumes of RNA-Seq data in parallel. The performance of Stormbow has been tested by practically applying it to analyse 178 RNA-Seq samples in the cloud. In our test, it took 6 to 8 hours to process an RNA-Seq sample with 100 million reads, and the average cost was $3.50 per sample. Utilizing Amazon Web Services as the infrastructure for Stormbow allows us to easily scale up to handle large datasets with on-demand computational resources. Stormbow is a scalable, cost effective, and open-source based tool for large-scale RNA-Seq data analysis. Stormbow can be freely downloaded and can be used out of box to process Illumina RNA-Seq datasets. PMID:25937948
Zhao, Shanrong; Prenger, Kurt; Smith, Lance
2013-01-01
RNA-Seq is becoming a promising replacement to microarrays in transcriptome profiling and differential gene expression study. Technical improvements have decreased sequencing costs and, as a result, the size and number of RNA-Seq datasets have increased rapidly. However, the increasing volume of data from large-scale RNA-Seq studies poses a practical challenge for data analysis in a local environment. To meet this challenge, we developed Stormbow, a cloud-based software package, to process large volumes of RNA-Seq data in parallel. The performance of Stormbow has been tested by practically applying it to analyse 178 RNA-Seq samples in the cloud. In our test, it took 6 to 8 hours to process an RNA-Seq sample with 100 million reads, and the average cost was $3.50 per sample. Utilizing Amazon Web Services as the infrastructure for Stormbow allows us to easily scale up to handle large datasets with on-demand computational resources. Stormbow is a scalable, cost effective, and open-source based tool for large-scale RNA-Seq data analysis. Stormbow can be freely downloaded and can be used out of box to process Illumina RNA-Seq datasets.
Development of fire test methods for airplane interior materials
NASA Technical Reports Server (NTRS)
Tustin, E. A.
1978-01-01
Fire tests were conducted in a 737 airplane fuselage at NASA-JSC to characterize jet fuel fires in open steel pans (simulating post-crash fire sources and a ruptured airplane fuselage) and to characterize fires in some common combustibles (simulating in-flight fire sources). Design post-crash and in-flight fire source selections were based on these data. Large panels of airplane interior materials were exposed to closely-controlled large scale heating simulations of the two design fire sources in a Boeing fire test facility utilizing a surplused 707 fuselage section. Small samples of the same airplane materials were tested by several laboratory fire test methods. Large scale and laboratory scale data were examined for correlative factors. Published data for dangerous hazard levels in a fire environment were used as the basis for developing a method to select the most desirable material where trade-offs in heat, smoke and gaseous toxicant evolution must be considered.
A synthesis and comparative evaluation of drainage water management
USDA-ARS?s Scientific Manuscript database
Viable large-scale crop production in the United States requires artificial drainage in humid and poorly drained agricultural regions. Excess water removal is generally achieved by installing tile drains that export water to open ditches that eventually flow into streams. Drainage water management...
Introducing Large-Scale Innovation in Schools
ERIC Educational Resources Information Center
Sotiriou, Sofoklis; Riviou, Katherina; Cherouvis, Stephanos; Chelioti, Eleni; Bogner, Franz X.
2016-01-01
Education reform initiatives tend to promise higher effectiveness in classrooms especially when emphasis is given to e-learning and digital resources. Practical changes in classroom realities or school organization, however, are lacking. A major European initiative entitled Open Discovery Space (ODS) examined the challenge of modernizing school…
NASA Astrophysics Data System (ADS)
Longair, Malcolm S.
2013-04-01
Part I. Stars and Stellar Evolution up to the Second World War: 1. The legacy of the nineteenth century; 2. The classification of stellar spectra; 3. Stellar structure and evolution; 4. The end points of stellar evolution; Part II. The Large-Scale Structure of the Universe, 1900-1939: 5. The Galaxy and the nature of spiral nebulae; 6. The origins of astrophysical cosmology; Part III. The Opening up of the Electromagnetic Spectrum: 7. The opening up of the electromagnetic spectrum and the new astronomies; Part IV. The Astrophysics of Stars and Galaxies since 1945: 8. Stars and stellar evolution; 9. The physics of the interstellar medium; 10. The physics of galaxies and clusters of galaxies; 11. High-energy astrophysics; Part V. Astrophysical Cosmology since 1945: 12. Astrophysical cosmology; 13. The determination of cosmological parameters; 14. The evolution of galaxies and active galaxies with cosmic epoch; 15. The origin of galaxies and the large-scale structure of the Universe; 16. The very early Universe; References; Name index; Object index; Subject index.
LSSGalPy: Interactive Visualization of the Large-scale Environment Around Galaxies
NASA Astrophysics Data System (ADS)
Argudo-Fernández, M.; Duarte Puertas, S.; Ruiz, J. E.; Sabater, J.; Verley, S.; Bergond, G.
2017-05-01
New tools are needed to handle the growth of data in astrophysics delivered by recent and upcoming surveys. We aim to build open-source, light, flexible, and interactive software designed to visualize extensive three-dimensional (3D) tabular data. Entirely written in the Python language, we have developed interactive tools to browse and visualize the positions of galaxies in the universe and their positions with respect to its large-scale structures (LSS). Motivated by a previous study, we created two codes using Mollweide projection and wedge diagram visualizations, where survey galaxies can be overplotted on the LSS of the universe. These are interactive representations where the visualizations can be controlled by widgets. We have released these open-source codes that have been designed to be easily re-used and customized by the scientific community to fulfill their needs. The codes are adaptable to other kinds of 3D tabular data and are robust enough to handle several millions of objects. .
DOE Office of Scientific and Technical Information (OSTI.GOV)
Grossman, Max; Pritchard Jr., Howard Porter; Budimlic, Zoran
2016-12-22
Graph500 [14] is an effort to offer a standardized benchmark across large-scale distributed platforms which captures the behavior of common communicationbound graph algorithms. Graph500 differs from other large-scale benchmarking efforts (such as HPL [6] or HPGMG [7]) primarily in the irregularity of its computation and data access patterns. The core computational kernel of Graph500 is a breadth-first search (BFS) implemented on an undirected graph. The output of Graph500 is a spanning tree of the input graph, usually represented by a predecessor mapping for every node in the graph. The Graph500 benchmark defines several pre-defined input sizes for implementers to testmore » against. This report summarizes investigation into implementing the Graph500 benchmark on OpenSHMEM, and focuses on first building a strong and practical understanding of the strengths and limitations of past work before proposing and developing novel extensions.« less
Ectopically tethered CP190 induces large-scale chromatin decondensation
NASA Astrophysics Data System (ADS)
Ahanger, Sajad H.; Günther, Katharina; Weth, Oliver; Bartkuhn, Marek; Bhonde, Ramesh R.; Shouche, Yogesh S.; Renkawitz, Rainer
2014-01-01
Insulator mediated alteration in higher-order chromatin and/or nucleosome organization is an important aspect of epigenetic gene regulation. Recent studies have suggested a key role for CP190 in such processes. In this study, we analysed the effects of ectopically tethered insulator factors on chromatin structure and found that CP190 induces large-scale decondensation when targeted to a condensed lacO array in mammalian and Drosophila cells. In contrast, dCTCF alone, is unable to cause such a decondensation, however, when CP190 is present, dCTCF recruits it to the lacO array and mediates chromatin unfolding. The CP190 induced opening of chromatin may not be correlated with transcriptional activation, as binding of CP190 does not enhance luciferase activity in reporter assays. We propose that CP190 may mediate histone modification and chromatin remodelling activity to induce an open chromatin state by its direct recruitment or targeting by a DNA binding factor such as dCTCF.
An innovative large scale integration of silicon nanowire-based field effect transistors
NASA Astrophysics Data System (ADS)
Legallais, M.; Nguyen, T. T. T.; Mouis, M.; Salem, B.; Robin, E.; Chenevier, P.; Ternon, C.
2018-05-01
Since the early 2000s, silicon nanowire field effect transistors are emerging as ultrasensitive biosensors while offering label-free, portable and rapid detection. Nevertheless, their large scale production remains an ongoing challenge due to time consuming, complex and costly technology. In order to bypass these issues, we report here on the first integration of silicon nanowire networks, called nanonet, into long channel field effect transistors using standard microelectronic process. A special attention is paid to the silicidation of the contacts which involved a large number of SiNWs. The electrical characteristics of these FETs constituted by randomly oriented silicon nanowires are also studied. Compatible integration on the back-end of CMOS readout and promising electrical performances open new opportunities for sensing applications.
NASA Astrophysics Data System (ADS)
Plebe, Alice; Grasso, Giorgio
2016-12-01
This paper describes a system developed for the simulation of flames inside an open-source 3D computer graphic software, Blender, with the aim of analyzing in virtual reality scenarios of hazards in large-scale industrial plants. The advantages of Blender are of rendering at high resolution the very complex structure of large industrial plants, and of embedding a physical engine based on smoothed particle hydrodynamics. This particle system is used to evolve a simulated fire. The interaction of this fire with the components of the plant is computed using polyhedron separation distance, adopting a Voronoi-based strategy that optimizes the number of feature distance computations. Results on a real oil and gas refining industry are presented.
NASA Astrophysics Data System (ADS)
Tecle, Amanuel Sebhatu
Hurricane is one of the most destructive and costly natural hazard to the built environment and its impact on low-rise buildings, particularity, is beyond acceptable. The major objective of this research was to perform a parametric evaluation of internal pressure (IP) for wind-resistant design of low-rise buildings and wind-driven natural ventilation applications. For this purpose, a multi-scale experimental, i.e. full-scale at Wall of Wind (WoW) and small-scale at Boundary Layer Wind Tunnel (BLWT), and a Computational Fluid Dynamics (CFD) approach was adopted. This provided new capability to assess wind pressures realistically on internal volumes ranging from small spaces formed between roof tiles and its deck to attic to room partitions. Effects of sudden breaching, existing dominant openings on building envelopes as well as compartmentalization of building interior on the IP were systematically investigated. Results of this research indicated: (i) for sudden breaching of dominant openings, the transient overshooting response was lower than the subsequent steady state peak IP and internal volume correction for low-wind-speed testing facilities was necessary. For example a building without volume correction experienced a response four times faster and exhibited 30--40% lower mean and peak IP; (ii) for existing openings, vent openings uniformly distributed along the roof alleviated, whereas one sided openings aggravated the IP; (iii) larger dominant openings exhibited a higher IP on the building envelope, and an off-center opening on the wall exhibited (30--40%) higher IP than center located openings; (iv) compartmentalization amplified the intensity of IP and; (v) significant underneath pressure was measured for field tiles, warranting its consideration during net pressure evaluations. The study aimed at wind driven natural ventilation indicated: (i) the IP due to cross ventilation was 1.5 to 2.5 times higher for Ainlet/Aoutlet>1 compared to cases where Ainlet/Aoutlet<1, this in effect reduced the mixing of air inside the building and hence the ventilation effectiveness; (ii) the presence of multi-room partitioning increased the pressure differential and consequently the air exchange rate. Overall good agreement was found between the observed large-scale, small-scale and CFD based IP responses. Comparisons with ASCE 7-10 consistently demonstrated that the code underestimated peak positive and suction IP.
Yilmaz Eroglu, Duygu; Caglar Gencosman, Burcu; Cavdur, Fatih; Ozmutlu, H. Cenk
2014-01-01
In this paper, we analyze a real-world OVRP problem for a production company. Considering real-world constrains, we classify our problem as multicapacitated/heterogeneous fleet/open vehicle routing problem with split deliveries and multiproduct (MCHF/OVRP/SDMP) which is a novel classification of an OVRP. We have developed a mixed integer programming (MIP) model for the problem and generated test problems in different size (10–90 customers) considering real-world parameters. Although MIP is able to find optimal solutions of small size (10 customers) problems, when the number of customers increases, the problem gets harder to solve, and thus MIP could not find optimal solutions for problems that contain more than 10 customers. Moreover, MIP fails to find any feasible solution of large-scale problems (50–90 customers) within time limits (7200 seconds). Therefore, we have developed a genetic algorithm (GA) based solution approach for large-scale problems. The experimental results show that the GA based approach reaches successful solutions with 9.66% gap in 392.8 s on average instead of 7200 s for the problems that contain 10–50 customers. For large-scale problems (50–90 customers), GA reaches feasible solutions of problems within time limits. In conclusion, for the real-world applications, GA is preferable rather than MIP to reach feasible solutions in short time periods. PMID:25045735
GoFFish: A Sub-Graph Centric Framework for Large-Scale Graph Analytics1
DOE Office of Scientific and Technical Information (OSTI.GOV)
Simmhan, Yogesh; Kumbhare, Alok; Wickramaarachchi, Charith
2014-08-25
Large scale graph processing is a major research area for Big Data exploration. Vertex centric programming models like Pregel are gaining traction due to their simple abstraction that allows for scalable execution on distributed systems naturally. However, there are limitations to this approach which cause vertex centric algorithms to under-perform due to poor compute to communication overhead ratio and slow convergence of iterative superstep. In this paper we introduce GoFFish a scalable sub-graph centric framework co-designed with a distributed persistent graph storage for large scale graph analytics on commodity clusters. We introduce a sub-graph centric programming abstraction that combines themore » scalability of a vertex centric approach with the flexibility of shared memory sub-graph computation. We map Connected Components, SSSP and PageRank algorithms to this model to illustrate its flexibility. Further, we empirically analyze GoFFish using several real world graphs and demonstrate its significant performance improvement, orders of magnitude in some cases, compared to Apache Giraph, the leading open source vertex centric implementation. We map Connected Components, SSSP and PageRank algorithms to this model to illustrate its flexibility. Further, we empirically analyze GoFFish using several real world graphs and demonstrate its significant performance improvement, orders of magnitude in some cases, compared to Apache Giraph, the leading open source vertex centric implementation.« less
Large-scale virtual screening on public cloud resources with Apache Spark.
Capuccini, Marco; Ahmed, Laeeq; Schaal, Wesley; Laure, Erwin; Spjuth, Ola
2017-01-01
Structure-based virtual screening is an in-silico method to screen a target receptor against a virtual molecular library. Applying docking-based screening to large molecular libraries can be computationally expensive, however it constitutes a trivially parallelizable task. Most of the available parallel implementations are based on message passing interface, relying on low failure rate hardware and fast network connection. Google's MapReduce revolutionized large-scale analysis, enabling the processing of massive datasets on commodity hardware and cloud resources, providing transparent scalability and fault tolerance at the software level. Open source implementations of MapReduce include Apache Hadoop and the more recent Apache Spark. We developed a method to run existing docking-based screening software on distributed cloud resources, utilizing the MapReduce approach. We benchmarked our method, which is implemented in Apache Spark, docking a publicly available target receptor against [Formula: see text]2.2 M compounds. The performance experiments show a good parallel efficiency (87%) when running in a public cloud environment. Our method enables parallel Structure-based virtual screening on public cloud resources or commodity computer clusters. The degree of scalability that we achieve allows for trying out our method on relatively small libraries first and then to scale to larger libraries. Our implementation is named Spark-VS and it is freely available as open source from GitHub (https://github.com/mcapuccini/spark-vs).Graphical abstract.
The state of OA: a large-scale analysis of the prevalence and impact of Open Access articles.
Piwowar, Heather; Priem, Jason; Larivière, Vincent; Alperin, Juan Pablo; Matthias, Lisa; Norlander, Bree; Farley, Ashley; West, Jevin; Haustein, Stefanie
2018-01-01
Despite growing interest in Open Access (OA) to scholarly literature, there is an unmet need for large-scale, up-to-date, and reproducible studies assessing the prevalence and characteristics of OA. We address this need using oaDOI, an open online service that determines OA status for 67 million articles. We use three samples, each of 100,000 articles, to investigate OA in three populations: (1) all journal articles assigned a Crossref DOI, (2) recent journal articles indexed in Web of Science, and (3) articles viewed by users of Unpaywall, an open-source browser extension that lets users find OA articles using oaDOI. We estimate that at least 28% of the scholarly literature is OA (19M in total) and that this proportion is growing, driven particularly by growth in Gold and Hybrid. The most recent year analyzed (2015) also has the highest percentage of OA (45%). Because of this growth, and the fact that readers disproportionately access newer articles, we find that Unpaywall users encounter OA quite frequently: 47% of articles they view are OA. Notably, the most common mechanism for OA is not Gold, Green, or Hybrid OA, but rather an under-discussed category we dub Bronze: articles made free-to-read on the publisher website, without an explicit Open license. We also examine the citation impact of OA articles, corroborating the so-called open-access citation advantage: accounting for age and discipline, OA articles receive 18% more citations than average, an effect driven primarily by Green and Hybrid OA. We encourage further research using the free oaDOI service, as a way to inform OA policy and practice.
The state of OA: a large-scale analysis of the prevalence and impact of Open Access articles
Larivière, Vincent; Alperin, Juan Pablo; Matthias, Lisa; Norlander, Bree; Farley, Ashley; West, Jevin; Haustein, Stefanie
2018-01-01
Despite growing interest in Open Access (OA) to scholarly literature, there is an unmet need for large-scale, up-to-date, and reproducible studies assessing the prevalence and characteristics of OA. We address this need using oaDOI, an open online service that determines OA status for 67 million articles. We use three samples, each of 100,000 articles, to investigate OA in three populations: (1) all journal articles assigned a Crossref DOI, (2) recent journal articles indexed in Web of Science, and (3) articles viewed by users of Unpaywall, an open-source browser extension that lets users find OA articles using oaDOI. We estimate that at least 28% of the scholarly literature is OA (19M in total) and that this proportion is growing, driven particularly by growth in Gold and Hybrid. The most recent year analyzed (2015) also has the highest percentage of OA (45%). Because of this growth, and the fact that readers disproportionately access newer articles, we find that Unpaywall users encounter OA quite frequently: 47% of articles they view are OA. Notably, the most common mechanism for OA is not Gold, Green, or Hybrid OA, but rather an under-discussed category we dub Bronze: articles made free-to-read on the publisher website, without an explicit Open license. We also examine the citation impact of OA articles, corroborating the so-called open-access citation advantage: accounting for age and discipline, OA articles receive 18% more citations than average, an effect driven primarily by Green and Hybrid OA. We encourage further research using the free oaDOI service, as a way to inform OA policy and practice. PMID:29456894
NASA Astrophysics Data System (ADS)
Staff, J. E.; Koning, N.; Ouyed, R.; Thompson, A.; Pudritz, R. E.
2015-02-01
We present the results of large scale, three-dimensional magnetohydrodynamics simulations of disc winds for different initial magnetic field configurations. The jets are followed from the source to 90 au scale, which covers several pixels of Hubble Space Telescope images of nearby protostellar jets. Our simulations show that jets are heated along their length by many shocks. We compute the emission lines that are produced, and find excellent agreement with observations. The jet width is found to be between 20 and 30 au while the maximum velocities perpendicular to the jet are found to be up to above 100 km s-1. The initially less open magnetic field configuration simulations result in a wider, two-component jet; a cylindrically shaped outer jet surrounding a narrow and much faster, inner jet. These simulations preserve the underlying Keplerian rotation profile of the inner jet to large distances from the source. However, for the initially most open magnetic field configuration the kink mode creates a narrow corkscrew-like jet without a clear Keplerian rotation profile and even regions where we observe rotation opposite to the disc (counter-rotating). The RW Aur jet is narrow, indicating that the disc field in that case is very open meaning the jet can contain a counter-rotating component that we suggest explains why observations of rotation in this jet have given confusing results. Thus magnetized disc winds from underlying Keplerian discs can develop rotation profiles far down the jet that is not Keplerian.
Wenchi Jin; Hong S. He; Stephen R. Shifley; Wen J. Wang; John M. Kabrick; Brian K. Davidson
2018-01-01
Historical fire regimes in the central United States maintained open-canopy shortleaf pine-oak woodlands on xeric sites. Following large-scale harvest and fire suppression, those woodlands grew denser with more continuous canopy cover, and they gained mesic species at the expense of shortleaf pine. There is high interest in restoring shortleaf pine-oak woodlands; most...
Small-scale open ocean currents have large effects on wind wave heights
NASA Astrophysics Data System (ADS)
Ardhuin, Fabrice; Gille, Sarah T.; Menemenlis, Dimitris; Rocha, Cesar B.; Rascle, Nicolas; Chapron, Bertrand; Gula, Jonathan; Molemaker, Jeroen
2017-06-01
Tidal currents and large-scale oceanic currents are known to modify ocean wave properties, causing extreme sea states that are a hazard to navigation. Recent advances in the understanding and modeling capability of open ocean currents have revealed the ubiquitous presence of eddies, fronts, and filaments at scales 10-100 km. Based on realistic numerical models, we show that these structures can be the main source of variability in significant wave heights at scales less than 200 km, including important variations down to 10 km. Model results are consistent with wave height variations along satellite altimeter tracks, resolved at scales larger than 50 km. The spectrum of significant wave heights is found to be of the order of 70>
Epigenome data release: a participant-centered approach to privacy protection.
Dyke, Stephanie O M; Cheung, Warren A; Joly, Yann; Ammerpohl, Ole; Lutsik, Pavlo; Rothstein, Mark A; Caron, Maxime; Busche, Stephan; Bourque, Guillaume; Rönnblom, Lars; Flicek, Paul; Beck, Stephan; Hirst, Martin; Stunnenberg, Henk; Siebert, Reiner; Walter, Jörn; Pastinen, Tomi
2015-07-17
Large-scale epigenome mapping by the NIH Roadmap Epigenomics Project, the ENCODE Consortium and the International Human Epigenome Consortium (IHEC) produces genome-wide DNA methylation data at one base-pair resolution. We examine how such data can be made open-access while balancing appropriate interpretation and genomic privacy. We propose guidelines for data release that both reduce ambiguity in the interpretation of open-access data and limit immediate access to genetic variation data that are made available through controlled access.
Integrating Cloud-Computing-Specific Model into Aircraft Design
NASA Astrophysics Data System (ADS)
Zhimin, Tian; Qi, Lin; Guangwen, Yang
Cloud Computing is becoming increasingly relevant, as it will enable companies involved in spreading this technology to open the door to Web 3.0. In the paper, the new categories of services introduced will slowly replace many types of computational resources currently used. In this perspective, grid computing, the basic element for the large scale supply of cloud services, will play a fundamental role in defining how those services will be provided. The paper tries to integrate cloud computing specific model into aircraft design. This work has acquired good results in sharing licenses of large scale and expensive software, such as CFD (Computational Fluid Dynamics), UG, CATIA, and so on.
Learning Deep Representations for Ground to Aerial Geolocalization (Open Access)
2015-10-15
proposed approach, Where-CNN, is inspired by deep learning success in face verification and achieves significant improvements over tra- ditional hand...crafted features and existing deep features learned from other large-scale databases. We show the ef- fectiveness of Where-CNN in finding matches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kanno, Sugumi; IKERBASQUE, Basque Foundation for Science, Maria Diaz de Haro 3, 48013, Bilbao; Laboratory for Quantum Gravity & Strings and Astrophysics, Cosmology & Gravity Center, Department of Mathematics & Applied Mathematics, University of Cape Town, Private Bag, Rondebosch 7701
We explore quantum entanglement between two causally disconnected regions in the multiverse. We first consider a free massive scalar field, and compute the entanglement negativity between two causally separated open charts in de Sitter space. The qualitative feature of it turns out to be in agreement with that of the entanglement entropy. We then introduce two observers who determine the entanglement between two causally disconnected de Sitter spaces. When one of the observers remains constrained to a region of the open chart in a de Sitter space, we find that the scale dependence enters into the entanglement. We show thatmore » a state which is initially maximally entangled becomes more entangled or less entangled on large scales depending on the mass of the scalar field and recovers the initial entanglement in the small scale limit. We argue that quantum entanglement may provide some evidence for the existence of the multiverse.« less
Entanglement negativity in the multiverse
NASA Astrophysics Data System (ADS)
Kanno, Sugumi; Shock, Jonathan P.; Soda, Jiro
2015-03-01
We explore quantum entanglement between two causally disconnected regions in the multiverse. We first consider a free massive scalar field, and compute the entanglement negativity between two causally separated open charts in de Sitter space. The qualitative feature of it turns out to be in agreement with that of the entanglement entropy. We then introduce two observers who determine the entanglement between two causally disconnected de Sitter spaces. When one of the observers remains constrained to a region of the open chart in a de Sitter space, we find that the scale dependence enters into the entanglement. We show that a state which is initially maximally entangled becomes more entangled or less entangled on large scales depending on the mass of the scalar field and recovers the initial entanglement in the small scale limit. We argue that quantum entanglement may provide some evidence for the existence of the multiverse.
Entanglement negativity in the multiverse
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kanno, Sugumi; Shock, Jonathan P.; Soda, Jiro, E-mail: sugumi.kanno@ehu.es, E-mail: jonathan.shock@uct.ac.za, E-mail: jiro@phys.sci.kobe-u.ac.jp
2015-03-01
We explore quantum entanglement between two causally disconnected regions in the multiverse. We first consider a free massive scalar field, and compute the entanglement negativity between two causally separated open charts in de Sitter space. The qualitative feature of it turns out to be in agreement with that of the entanglement entropy. We then introduce two observers who determine the entanglement between two causally disconnected de Sitter spaces. When one of the observers remains constrained to a region of the open chart in a de Sitter space, we find that the scale dependence enters into the entanglement. We show thatmore » a state which is initially maximally entangled becomes more entangled or less entangled on large scales depending on the mass of the scalar field and recovers the initial entanglement in the small scale limit. We argue that quantum entanglement may provide some evidence for the existence of the multiverse.« less
Constructing Flexible, Configurable, ETL Pipelines for the Analysis of "Big Data" with Apache OODT
NASA Astrophysics Data System (ADS)
Hart, A. F.; Mattmann, C. A.; Ramirez, P.; Verma, R.; Zimdars, P. A.; Park, S.; Estrada, A.; Sumarlidason, A.; Gil, Y.; Ratnakar, V.; Krum, D.; Phan, T.; Meena, A.
2013-12-01
A plethora of open source technologies for manipulating, transforming, querying, and visualizing 'big data' have blossomed and matured in the last few years, driven in large part by recognition of the tremendous value that can be derived by leveraging data mining and visualization techniques on large data sets. One facet of many of these tools is that input data must often be prepared into a particular format (e.g.: JSON, CSV), or loaded into a particular storage technology (e.g.: HDFS) before analysis can take place. This process, commonly known as Extract-Transform-Load, or ETL, often involves multiple well-defined steps that must be executed in a particular order, and the approach taken for a particular data set is generally sensitive to the quantity and quality of the input data, as well as the structure and complexity of the desired output. When working with very large, heterogeneous, unstructured or semi-structured data sets, automating the ETL process and monitoring its progress becomes increasingly important. Apache Object Oriented Data Technology (OODT) provides a suite of complementary data management components called the Process Control System (PCS) that can be connected together to form flexible ETL pipelines as well as browser-based user interfaces for monitoring and control of ongoing operations. The lightweight, metadata driven middleware layer can be wrapped around custom ETL workflow steps, which themselves can be implemented in any language. Once configured, it facilitates communication between workflow steps and supports execution of ETL pipelines across a distributed cluster of compute resources. As participants in a DARPA-funded effort to develop open source tools for large-scale data analysis, we utilized Apache OODT to rapidly construct custom ETL pipelines for a variety of very large data sets to prepare them for analysis and visualization applications. We feel that OODT, which is free and open source software available through the Apache Software Foundation, is particularly well suited to developing and managing arbitrary large-scale ETL processes both for the simplicity and flexibility of its wrapper framework, as well as the detailed provenance information it exposes throughout the process. Our experience using OODT to manage processing of large-scale data sets in domains as diverse as radio astronomy, life sciences, and social network analysis demonstrates the flexibility of the framework, and the range of potential applications to a broad array of big data ETL challenges.
Geophysical potential for wind energy over the open oceans
2017-01-01
Wind turbines continuously remove kinetic energy from the lower troposphere, thereby reducing the wind speed near hub height. The rate of electricity generation in large wind farms containing multiple wind arrays is, therefore, constrained by the rate of kinetic energy replenishment from the atmosphere above. In recent years, a growing body of research argues that the rate of generated power is limited to around 1.5 W m−2 within large wind farms. However, in this study, we show that considerably higher power generation rates may be sustainable over some open ocean areas. In particular, the North Atlantic is identified as a region where the downward transport of kinetic energy may sustain extraction rates of 6 W m−2 and above over large areas in the annual mean. Furthermore, our results indicate that the surface heat flux from the oceans to the atmosphere may play an important role in creating regions where sustained high rates of downward transport of kinetic energy and thus, high rates of kinetic energy extraction may be geophysical possible. While no commercial-scale deep water wind farms yet exist, our results suggest that such technologies, if they became technically and economically feasible, could potentially provide civilization-scale power. PMID:29073053
Geophysical potential for wind energy over the open oceans.
Possner, Anna; Caldeira, Ken
2017-10-24
Wind turbines continuously remove kinetic energy from the lower troposphere, thereby reducing the wind speed near hub height. The rate of electricity generation in large wind farms containing multiple wind arrays is, therefore, constrained by the rate of kinetic energy replenishment from the atmosphere above. In recent years, a growing body of research argues that the rate of generated power is limited to around 1.5 W m -2 within large wind farms. However, in this study, we show that considerably higher power generation rates may be sustainable over some open ocean areas. In particular, the North Atlantic is identified as a region where the downward transport of kinetic energy may sustain extraction rates of 6 W m -2 and above over large areas in the annual mean. Furthermore, our results indicate that the surface heat flux from the oceans to the atmosphere may play an important role in creating regions where sustained high rates of downward transport of kinetic energy and thus, high rates of kinetic energy extraction may be geophysical possible. While no commercial-scale deep water wind farms yet exist, our results suggest that such technologies, if they became technically and economically feasible, could potentially provide civilization-scale power.
Open-label placebo treatment in chronic low back pain: a randomized controlled trial
Carvalho, Cláudia; Caetano, Joaquim Machado; Cunha, Lidia; Rebouta, Paula; Kaptchuk, Ted J.; Kirsch, Irving
2016-01-01
Abstract This randomized controlled trial was performed to investigate whether placebo effects in chronic low back pain could be harnessed ethically by adding open-label placebo (OLP) treatment to treatment as usual (TAU) for 3 weeks. Pain severity was assessed on three 0- to 10-point Numeric Rating Scales, scoring maximum pain, minimum pain, and usual pain, and a composite, primary outcome, total pain score. Our other primary outcome was back-related dysfunction, assessed on the Roland–Morris Disability Questionnaire. In an exploratory follow-up, participants on TAU received placebo pills for 3 additional weeks. We randomized 97 adults reporting persistent low back pain for more than 3 months' duration and diagnosed by a board-certified pain specialist. Eighty-three adults completed the trial. Compared to TAU, OLP elicited greater pain reduction on each of the three 0- to 10-point Numeric Rating Scales and on the 0- to 10-point composite pain scale (P < 0.001), with moderate to large effect sizes. Pain reduction on the composite Numeric Rating Scales was 1.5 (95% confidence interval: 1.0-2.0) in the OLP group and 0.2 (−0.3 to 0.8) in the TAU group. Open-label placebo treatment also reduced disability compared to TAU (P < 0.001), with a large effect size. Improvement in disability scores was 2.9 (1.7-4.0) in the OLP group and 0.0 (−1.1 to 1.2) in the TAU group. After being switched to OLP, the TAU group showed significant reductions in both pain (1.5, 0.8-2.3) and disability (3.4, 2.2-4.5). Our findings suggest that OLP pills presented in a positive context may be helpful in chronic low back pain. PMID:27755279
Open-label placebo treatment in chronic low back pain: a randomized controlled trial.
Carvalho, Cláudia; Caetano, Joaquim Machado; Cunha, Lidia; Rebouta, Paula; Kaptchuk, Ted J; Kirsch, Irving
2016-12-01
This randomized controlled trial was performed to investigate whether placebo effects in chronic low back pain could be harnessed ethically by adding open-label placebo (OLP) treatment to treatment as usual (TAU) for 3 weeks. Pain severity was assessed on three 0- to 10-point Numeric Rating Scales, scoring maximum pain, minimum pain, and usual pain, and a composite, primary outcome, total pain score. Our other primary outcome was back-related dysfunction, assessed on the Roland-Morris Disability Questionnaire. In an exploratory follow-up, participants on TAU received placebo pills for 3 additional weeks. We randomized 97 adults reporting persistent low back pain for more than 3 months' duration and diagnosed by a board-certified pain specialist. Eighty-three adults completed the trial. Compared to TAU, OLP elicited greater pain reduction on each of the three 0- to 10-point Numeric Rating Scales and on the 0- to 10-point composite pain scale (P < 0.001), with moderate to large effect sizes. Pain reduction on the composite Numeric Rating Scales was 1.5 (95% confidence interval: 1.0-2.0) in the OLP group and 0.2 (-0.3 to 0.8) in the TAU group. Open-label placebo treatment also reduced disability compared to TAU (P < 0.001), with a large effect size. Improvement in disability scores was 2.9 (1.7-4.0) in the OLP group and 0.0 (-1.1 to 1.2) in the TAU group. After being switched to OLP, the TAU group showed significant reductions in both pain (1.5, 0.8-2.3) and disability (3.4, 2.2-4.5). Our findings suggest that OLP pills presented in a positive context may be helpful in chronic low back pain.
Estimating planktonic diversity through spatial dominance patterns in a model ocean.
Soccodato, Alice; d'Ovidio, Francesco; Lévy, Marina; Jahn, Oliver; Follows, Michael J; De Monte, Silvia
2016-10-01
In the open ocean, the observation and quantification of biodiversity patterns is challenging. Marine ecosystems are indeed largely composed by microbial planktonic communities whose niches are affected by highly dynamical physico-chemical conditions, and whose observation requires advanced methods for morphological and molecular classification. Optical remote sensing offers an appealing complement to these in-situ techniques. Global-scale coverage at high spatiotemporal resolution is however achieved at the cost of restrained information on the local assemblage. Here, we use a coupled physical and ecological model ocean simulation to explore one possible metrics for comparing measures performed on such different scales. We show that a large part of the local diversity of the virtual plankton ecosystem - corresponding to what accessible by genomic methods - can be inferred from crude, but spatially extended, information - as conveyed by remote sensing. Shannon diversity of the local community is indeed highly correlated to a 'seascape' index, which quantifies the surrounding spatial heterogeneity of the most abundant functional group. The error implied in drastically reducing the resolution of the plankton community is shown to be smaller in frontal regions as well as in regions of intermediate turbulent energy. On the spatial scale of hundreds of kms, patterns of virtual plankton diversity are thus largely sustained by mixing communities that occupy adjacent niches. We provide a proof of principle that in the open ocean information on spatial variability of communities can compensate for limited local knowledge, suggesting the possibility of integrating in-situ and satellite observations to monitor biodiversity distribution at the global scale. Copyright © 2016 Elsevier B.V. All rights reserved.
Introducing Large-Scale Innovation in Schools
NASA Astrophysics Data System (ADS)
Sotiriou, Sofoklis; Riviou, Katherina; Cherouvis, Stephanos; Chelioti, Eleni; Bogner, Franz X.
2016-08-01
Education reform initiatives tend to promise higher effectiveness in classrooms especially when emphasis is given to e-learning and digital resources. Practical changes in classroom realities or school organization, however, are lacking. A major European initiative entitled Open Discovery Space (ODS) examined the challenge of modernizing school education via a large-scale implementation of an open-scale methodology in using technology-supported innovation. The present paper describes this innovation scheme which involved schools and teachers all over Europe, embedded technology-enhanced learning into wider school environments and provided training to teachers. Our implementation scheme consisted of three phases: (1) stimulating interest, (2) incorporating the innovation into school settings and (3) accelerating the implementation of the innovation. The scheme's impact was monitored for a school year using five indicators: leadership and vision building, ICT in the curriculum, development of ICT culture, professional development support, and school resources and infrastructure. Based on about 400 schools, our study produced four results: (1) The growth in digital maturity was substantial, even for previously high scoring schools. This was even more important for indicators such as vision and leadership" and "professional development." (2) The evolution of networking is presented graphically, showing the gradual growth of connections achieved. (3) These communities became core nodes, involving numerous teachers in sharing educational content and experiences: One out of three registered users (36 %) has shared his/her educational resources in at least one community. (4) Satisfaction scores ranged from 76 % (offer of useful support through teacher academies) to 87 % (good environment to exchange best practices). Initiatives such as ODS add substantial value to schools on a large scale.
NASA Astrophysics Data System (ADS)
McCoy, Isabel; Wood, Robert; Fletcher, Jennifer
Marine low clouds are key influencers of the climate and contribute significantly to uncertainty in model climate sensitivity due to their small scale and complex processes. Many low clouds occur in large-scale cellular patterns, known as open and closed mesoscale cellular convection (MCC), which have significantly different radiative and microphysical properties. Investigating MCC development and meteorological controls will improve our understanding of their impacts on the climate. We conducted an examination of time-varying meteorological conditions associated with satellite-determined open and closed MCC. The spatial and temporal patterns of MCC clouds were compared with key meteorological control variables calculated from ERA-Interim Reanalysis to highlight dependencies and major differences. This illustrated the influence of environmental stability and surface forcing as well as the role of marine cold air outbreaks (MCAO, the movement of cold air from polar-regions across warmer waters) in MCC cloud formation. Such outbreaks are important to open MCC development and may also influence the transition from open to closed MCC. Our results may lead to improvements in the parameterization of cloudiness and advance the simulation of marine low clouds. National Science Foundation Graduate Research Fellowship Grant (DGE-1256082).
NASA Astrophysics Data System (ADS)
Piao, Chunhui; Han, Xufang; Wu, Harris
2010-08-01
We provide a formal definition of an e-commerce transaction network. Agent-based modelling is used to simulate e-commerce transaction networks. For real-world analysis, we studied the open application programming interfaces (APIs) from eBay and Taobao e-commerce websites and captured real transaction data. Pajek is used to visualise the agent relationships in the transaction network. We derived one-mode networks from the transaction network and analysed them using degree and betweenness centrality. Integrating multi-agent modelling, open APIs and social network analysis, we propose a new way to study large-scale e-commerce systems.
Gouhier, Tarik C; Guichard, Frédéric
2007-03-01
In marine systems, the occurrence and implications of disturbance-recovery cycles have been revealed at the landscape level, but only in demographically open or closed systems where landscape-level dynamics are assumed to have no feedback effect on regional dynamics. We present a mussel metapopulation model to elucidate the role of landscape-level disturbance cycles for regional response of mussel populations to onshore productivity and larval transport. Landscape dynamics are generated through spatially explicit rules, and each landscape is connected to its neighbor through unidirectional larval dispersal. The role of landscape disturbance cycles in the regional system behavior is elucidated (1) in demographically open vs. demographically coupled systems, in relation to (2) onshore reproductive output and (3) the temporal scale of landscape disturbance dynamics. By controlling for spatial structure at the landscape and metapopulation levels, we first demonstrate the interaction between landscape and oceanographic connectivity. The temporal scale of disturbance cycles, as controlled by mussel colonization rate, plays a critical role in the regional behavior of the system. Indeed, fast disturbance cycles are responsible for regional synchrony in relation to onshore reproductive output. Slow disturbance cycles, however, lead to increased robustness to changes in productivity and to demographic coupling. These testable predictions indicate that the occurrence and temporal scale of local disturbance-recovery dynamics can drive large-scale variability in demographically open systems, and the response of metapopulations to changes in nearshore productivity.
The impact of radiatively active water-ice clouds on Martian mesoscale atmospheric circulations
NASA Astrophysics Data System (ADS)
Spiga, A.; Madeleine, J.-B.; Hinson, D.; Navarro, T.; Forget, F.
2014-04-01
Background and Goals Water ice clouds are a key component of the Martian climate [1]. Understanding the properties of the Martian water ice clouds is crucial to constrain the Red Planet's climate and hydrological cycle both in the present and in the past [2]. In recent years, this statement have become all the more true as it was shown that the radiative effects of water ice clouds is far from being as negligible as hitherto believed; water ice clouds plays instead a key role in the large-scale thermal structure and dynamics of the Martian atmosphere [3, 4, 5]. Nevertheless, the radiative effect of water ice clouds at lower scales than the large synoptic scale (the so-called meso-scales) is still left to be explored. Here we use for the first time mesoscale modeling with radiatively active water ice clouds to address this open question.
ERIC Educational Resources Information Center
Pavlu, Virgil
2008-01-01
Today, search engines are embedded into all aspects of digital world: in addition to Internet search, all operating systems have integrated search engines that respond even as you type, even over the network, even on cell phones; therefore the importance of their efficacy and efficiency cannot be overstated. There are many open possibilities for…
Kuipers, Jeroen; Kalicharan, Ruby D; Wolters, Anouk H G; van Ham, Tjakko J; Giepmans, Ben N G
2016-05-25
Large-scale 2D electron microscopy (EM), or nanotomy, is the tissue-wide application of nanoscale resolution electron microscopy. Others and we previously applied large scale EM to human skin pancreatic islets, tissue culture and whole zebrafish larvae(1-7). Here we describe a universally applicable method for tissue-scale scanning EM for unbiased detection of sub-cellular and molecular features. Nanotomy was applied to investigate the healthy and a neurodegenerative zebrafish brain. Our method is based on standardized EM sample preparation protocols: Fixation with glutaraldehyde and osmium, followed by epoxy-resin embedding, ultrathin sectioning and mounting of ultrathin-sections on one-hole grids, followed by post staining with uranyl and lead. Large-scale 2D EM mosaic images are acquired using a scanning EM connected to an external large area scan generator using scanning transmission EM (STEM). Large scale EM images are typically ~ 5 - 50 G pixels in size, and best viewed using zoomable HTML files, which can be opened in any web browser, similar to online geographical HTML maps. This method can be applied to (human) tissue, cross sections of whole animals as well as tissue culture(1-5). Here, zebrafish brains were analyzed in a non-invasive neuronal ablation model. We visualize within a single dataset tissue, cellular and subcellular changes which can be quantified in various cell types including neurons and microglia, the brain's macrophages. In addition, nanotomy facilitates the correlation of EM with light microscopy (CLEM)(8) on the same tissue, as large surface areas previously imaged using fluorescent microscopy, can subsequently be subjected to large area EM, resulting in the nano-anatomy (nanotomy) of tissues. In all, nanotomy allows unbiased detection of features at EM level in a tissue-wide quantifiable manner.
Kuipers, Jeroen; Kalicharan, Ruby D.; Wolters, Anouk H. G.
2016-01-01
Large-scale 2D electron microscopy (EM), or nanotomy, is the tissue-wide application of nanoscale resolution electron microscopy. Others and we previously applied large scale EM to human skin pancreatic islets, tissue culture and whole zebrafish larvae1-7. Here we describe a universally applicable method for tissue-scale scanning EM for unbiased detection of sub-cellular and molecular features. Nanotomy was applied to investigate the healthy and a neurodegenerative zebrafish brain. Our method is based on standardized EM sample preparation protocols: Fixation with glutaraldehyde and osmium, followed by epoxy-resin embedding, ultrathin sectioning and mounting of ultrathin-sections on one-hole grids, followed by post staining with uranyl and lead. Large-scale 2D EM mosaic images are acquired using a scanning EM connected to an external large area scan generator using scanning transmission EM (STEM). Large scale EM images are typically ~ 5 - 50 G pixels in size, and best viewed using zoomable HTML files, which can be opened in any web browser, similar to online geographical HTML maps. This method can be applied to (human) tissue, cross sections of whole animals as well as tissue culture1-5. Here, zebrafish brains were analyzed in a non-invasive neuronal ablation model. We visualize within a single dataset tissue, cellular and subcellular changes which can be quantified in various cell types including neurons and microglia, the brain's macrophages. In addition, nanotomy facilitates the correlation of EM with light microscopy (CLEM)8 on the same tissue, as large surface areas previously imaged using fluorescent microscopy, can subsequently be subjected to large area EM, resulting in the nano-anatomy (nanotomy) of tissues. In all, nanotomy allows unbiased detection of features at EM level in a tissue-wide quantifiable manner. PMID:27285162
Interchange Reconnection and Coronal Hole Dynamics
NASA Technical Reports Server (NTRS)
Edmondson, J. K.; Antiochos, S. K.; DeVore, C. R.; Lynch, B. J.; Zurbuchen, T. H.
2011-01-01
We investigate the effect of magnetic reconnection between open and closed field, (often referred to as "interchange" reconnection), on the dynamics and topology of coronal hole boundaries. The most important and most prevalent 3D topology of the interchange process is that of a small-scale bipolar magnetic field interacting with a large-scale background field. We determine the evolution of such a magnetic topology by numerical solution of the fully 3D MHD equations in spherical coordinates. First, we calculate the evolution of a small-scale bipole that initially is completely inside an open field region and then is driven across a coronal hole boundary by photospheric motions. Next the reverse situation is calculated in which the bipole is initially inside the closed region and driven toward the coronal hole boundary. In both cases we find that the stress imparted by the photospheric motions results in deformation of the separatrix surface between the closed field of the bipole and the background field, leading to rapid current sheet formation and to efficient reconnection. When the bipole is inside the open field region, the reconnection is of the interchange type in that it exchanges open and closed field. We examine, in detail, the topology of the field as the bipole moves across the coronal hole boundary, and find that the field remains well-connected throughout this process. Our results imply that open flux cannot penetrate deeply into the closed field region below a helmet streamer and, hence, support the quasi-steady models in which open and closed flux remain topologically distinct. Our results also support the uniqueness hypothesis for open field regions as postulated by Antiochos et al. We discuss the implications of this work for coronal observations. Subject Headings: Sun: corona Sun: magnetic fields Sun: reconnection Sun: coronal hole
Decompositions of large-scale biological systems based on dynamical properties.
Soranzo, Nicola; Ramezani, Fahimeh; Iacono, Giovanni; Altafini, Claudio
2012-01-01
Given a large-scale biological network represented as an influence graph, in this article we investigate possible decompositions of the network aimed at highlighting specific dynamical properties. The first decomposition we study consists in finding a maximal directed acyclic subgraph of the network, which dynamically corresponds to searching for a maximal open-loop subsystem of the given system. Another dynamical property investigated is strong monotonicity. We propose two methods to deal with this property, both aimed at decomposing the system into strongly monotone subsystems, but with different structural characteristics: one method tends to produce a single large strongly monotone component, while the other typically generates a set of smaller disjoint strongly monotone subsystems. Original heuristics for the methods investigated are described in the article. altafini@sissa.it
OpenCluster: A Flexible Distributed Computing Framework for Astronomical Data Processing
NASA Astrophysics Data System (ADS)
Wei, Shoulin; Wang, Feng; Deng, Hui; Liu, Cuiyin; Dai, Wei; Liang, Bo; Mei, Ying; Shi, Congming; Liu, Yingbo; Wu, Jingping
2017-02-01
The volume of data generated by modern astronomical telescopes is extremely large and rapidly growing. However, current high-performance data processing architectures/frameworks are not well suited for astronomers because of their limitations and programming difficulties. In this paper, we therefore present OpenCluster, an open-source distributed computing framework to support rapidly developing high-performance processing pipelines of astronomical big data. We first detail the OpenCluster design principles and implementations and present the APIs facilitated by the framework. We then demonstrate a case in which OpenCluster is used to resolve complex data processing problems for developing a pipeline for the Mingantu Ultrawide Spectral Radioheliograph. Finally, we present our OpenCluster performance evaluation. Overall, OpenCluster provides not only high fault tolerance and simple programming interfaces, but also a flexible means of scaling up the number of interacting entities. OpenCluster thereby provides an easily integrated distributed computing framework for quickly developing a high-performance data processing system of astronomical telescopes and for significantly reducing software development expenses.
Dynamics of the Transition Corona
NASA Technical Reports Server (NTRS)
Masson, Sophie; McCauley, Patrick; Golub, Leon; Reeves, Katharine K.; DeLuca, Edward E.
2014-01-01
Magnetic reconnection between the open and closed magnetic fields in the corona is believed to play a crucial role in the corona/heliosphere coupling. At large scale, the exchange of open/closed connectivity is expected to occur in pseudo-streamer (PS) structures. However, there is neither clear observational evidence of how such coupling occurs in PSs, nor evidence for how the magnetic reconnection evolves. Using a newly developed technique, we enhance the off-limb magnetic fine structures observed with the Atmospheric Imaging Assembly and identify a PS-like feature located close to the northern coronal hole. We first identify that the magnetic topology associated with the observation is a PS, null-point (NP) related topology bounded by the open field. By comparing the magnetic field configuration with the extreme ultraviolet (EUV) emission regions, we determined that most of the magnetic flux associated with plasma emission are small loops below the PS basic NP and open field bounding the PS topology. In order to interpret the evolution of the PS, we referred to a three-dimensional MHD interchange reconnection modeling the exchange of connectivity between small closed loops and the open field. The observed PS fine structures follow the dynamics of the magnetic field before and after reconnecting at the NP obtained by the interchange model. Moreover, the pattern of the EUV plasma emission is the same as the shape of the expected plasma emission location derived from the simulation. These morphological and dynamical similarities between the PS observations and the results from the simulation strongly suggest that the evolution of the PS, and in particular the opening/closing of the field, occurs via interchange/slipping reconnection at the basic NP of the PS. Besides identifying the mechanism at work in the large-scale coupling between the open and closed fields, our results highlight that interchange reconnection in PSs is a gradual physical process that differs from the impulsive reconnection of the solar-jet model.
Open source Modeling and optimization tools for Planning
DOE Office of Scientific and Technical Information (OSTI.GOV)
Peles, S.
Open source modeling and optimization tools for planning The existing tools and software used for planning and analysis in California are either expensive, difficult to use, or not generally accessible to a large number of participants. These limitations restrict the availability of participants for larger scale energy and grid studies in the state. The proposed initiative would build upon federal and state investments in open source software, and create and improve open source tools for use in the state planning and analysis activities. Computational analysis and simulation frameworks in development at national labs and universities can be brought forward tomore » complement existing tools. An open source platform would provide a path for novel techniques and strategies to be brought into the larger community and reviewed by a broad set of stakeholders.« less
NASA Astrophysics Data System (ADS)
Wang, Chunzai; Wang, Xidong; Weisberg, Robert H.; Black, Michael L.
2017-12-01
The paper uses observational data from 1950 to 2014 to investigate rapid intensification (RI) variability of tropical cyclones (TCs) in the North Atlantic and its relationships with large-scale climate variations. RI is defined as a TC intensity increase of at least 15.4 m/s (30 knots) in 24 h. The seasonal RI distribution follows the seasonal TC distribution, with the highest number in September. Although an RI event can occur anywhere over the tropical North Atlantic (TNA), there are three regions of maximum RI occurrence: (1) the western TNA of 12°N-18°N and 60°W-45°W, (2) the Gulf of Mexico and the western Caribbean Sea, and (3) the open ocean southeast and east of Florida. RI events also show a minimum value in the eastern Caribbean Sea north of South America—a place called a hurricane graveyard due to atmospheric divergence and subsidence. On longer time scales, RI displays both interannual and multidecadal variability, but RI does not show a long-term trend due to global warming. The top three climate indices showing high correlations with RI are the June-November ENSO and Atlantic warm pool indices, and the January-March North Atlantic oscillation index. It is found that variabilities of vertical wind shear and TC heat potential are important for TC RI in the hurricane main development region, whereas relative humidity at 500 hPa is the main factor responsible for TC RI in the eastern TNA. However, the large-scale oceanic and atmospheric variables analyzed in this study do not show an important role in TC RI in the Gulf of Mexico and the open ocean southeast and east of Florida. This suggests that other factors such as small-scale changes of oceanic and atmospheric variables or TC internal processes may be responsible for TC RI in these two regions. Additionally, the analyses indicate that large-scale atmospheric and oceanic variables are not critical to TC genesis and formation; however, once a tropical depression forms, large-scale climate variations play a role in TC intensification.
The HYPE Open Source Community
NASA Astrophysics Data System (ADS)
Strömbäck, L.; Pers, C.; Isberg, K.; Nyström, K.; Arheimer, B.
2013-12-01
The Hydrological Predictions for the Environment (HYPE) model is a dynamic, semi-distributed, process-based, integrated catchment model. It uses well-known hydrological and nutrient transport concepts and can be applied for both small and large scale assessments of water resources and status. In the model, the landscape is divided into classes according to soil type, vegetation and altitude. The soil representation is stratified and can be divided in up to three layers. Water and substances are routed through the same flow paths and storages (snow, soil, groundwater, streams, rivers, lakes) considering turn-over and transformation on the way towards the sea. HYPE has been successfully used in many hydrological applications at SMHI. For Europe, we currently have three different models; The S-HYPE model for Sweden; The BALT-HYPE model for the Baltic Sea; and the E-HYPE model for the whole Europe. These models simulate hydrological conditions and nutrients for their respective areas and are used for characterization, forecasts, and scenario analyses. Model data can be downloaded from hypeweb.smhi.se. In addition, we provide models for the Arctic region, the Arab (Middle East and Northern Africa) region, India, the Niger River basin, the La Plata Basin. This demonstrates the applicability of the HYPE model for large scale modeling in different regions of the world. An important goal with our work is to make our data and tools available as open data and services. For this aim we created the HYPE Open Source Community (OSC) that makes the source code of HYPE available for anyone interested in further development of HYPE. The HYPE OSC (hype.sourceforge.net) is an open source initiative under the Lesser GNU Public License taken by SMHI to strengthen international collaboration in hydrological modeling and hydrological data production. The hypothesis is that more brains and more testing will result in better models and better code. The code is transparent and can be changed and learnt from. New versions of the main code are delivered frequently. HYPE OSC is open to everyone interested in hydrology, hydrological modeling and code development - e.g. scientists, authorities, and consultancies. By joining the HYPE OSC you get access a state-of-the-art operational hydrological model. The HYPE source code is designed to efficiently handle large scale modeling for forecast, hindcast and climate applications. The code is under constant development to improve the hydrological processes, efficiency and readability. In the beginning of 2013 we released a version with new and better modularization based on hydrological processes. This will make the code easier to understand and further develop for a new user. An important challenge in this process is to produce code that is easy for anyone to understand and work with, but still maintain the properties that make the code efficient enough for large scale applications. Input from the HYPE Open Source Community is an important source for future improvements of the HYPE model. Therefore, by joining the community you become an active part of the development, get access to the latest features and can influence future versions of the model.
Towards Large-area Field-scale Operational Evapotranspiration for Water Use Mapping
NASA Astrophysics Data System (ADS)
Senay, G. B.; Friedrichs, M.; Morton, C.; Huntington, J. L.; Verdin, J.
2017-12-01
Field-scale evapotranspiration (ET) estimates are needed for improving surface and groundwater use and water budget studies. Ideally, field-scale ET estimates would be at regional to national levels and cover long time periods. As a result of large data storage and computational requirements associated with processing field-scale satellite imagery such as Landsat, numerous challenges remain to develop operational ET estimates over large areas for detailed water use and availability studies. However, the combination of new science, data availability, and cloud computing technology is enabling unprecedented capabilities for ET mapping. To demonstrate this capability, we used Google's Earth Engine cloud computing platform to create nationwide annual ET estimates with 30-meter resolution Landsat ( 16,000 images) and gridded weather data using the Operational Simplified Surface Energy Balance (SSEBop) model in support of the National Water Census, a USGS research program designed to build decision support capacity for water management agencies and other natural resource managers. By leveraging Google's Earth Engine Application Programming Interface (API) and developing software in a collaborative, open-platform environment, we rapidly advance from research towards applications for large-area field-scale ET mapping. Cloud computing of the Landsat image archive combined with other satellite, climate, and weather data, is creating never imagined opportunities for assessing ET model behavior and uncertainty, and ultimately providing the ability for more robust operational monitoring and assessment of water use at field-scales.
McCrae, Robert R.; Scally, Matthew; Terracciano, Antonio; Abecasis, Gonçalo R.; Costa, Paul T.
2011-01-01
There is growing evidence that personality traits are affected by many genes, all of which have very small effects. As an alternative to the largely-unsuccessful search for individual polymorphisms associated with personality traits, we identified large sets of potentially related single nucleotide polymorphisms (SNPs) and summed them to form molecular personality scales (MPSs) with from 4 to 2,497 SNPs. Scales were derived from two-thirds of a large (N = 3,972) sample of individuals from Sardinia who completed the Revised NEO Personality Inventory and were assessed in a genome-wide association scan. When MPSs were correlated with the phenotype in the remaining third of the sample, very small but significant associations were found for four of the five personality factors when the longest scales were examined. These data suggest that MPSs for Neuroticism, Openness to Experience, Agreeableness, and Conscientiousness (but not Extraversion) contain genetic information that can be refined in future studies, and the procedures described here should be applicable to other quantitative traits. PMID:21114353
Streamline curvature in supersonic shear layers
NASA Technical Reports Server (NTRS)
Kibens, V.
1992-01-01
Results of an experimental investigation in which a curved shear layer was generated between supersonic flow from a rectangular converging/diverging nozzle and the freestream in a series of open channels with varying radii of curvature are reported. The shear layers exhibit unsteady large-scale activity at supersonic pressure ratios, indicating increased mixing efficiency. This effect contrasts with supersonic flow in a straight channel, for which no large-scale vortical structure development occurs. Curvature must exceed a minimum level before it begins to affect the dynamics of the supersonic shear layer appreciably. The curved channel flows are compared with reference flows consisting of a free jet, a straight channel, and wall jets without sidewalls on a flat and a curved plate.
Automated microscopy for high-content RNAi screening
2010-01-01
Fluorescence microscopy is one of the most powerful tools to investigate complex cellular processes such as cell division, cell motility, or intracellular trafficking. The availability of RNA interference (RNAi) technology and automated microscopy has opened the possibility to perform cellular imaging in functional genomics and other large-scale applications. Although imaging often dramatically increases the content of a screening assay, it poses new challenges to achieve accurate quantitative annotation and therefore needs to be carefully adjusted to the specific needs of individual screening applications. In this review, we discuss principles of assay design, large-scale RNAi, microscope automation, and computational data analysis. We highlight strategies for imaging-based RNAi screening adapted to different library and assay designs. PMID:20176920
NASA Technical Reports Server (NTRS)
Bennett, Charles V.
1947-01-01
An investigation of the low-speed, power-off stability and control characteristics of a 1/20-scale model of the Consolidated Vultee XB-53 airplane has been conducted in the Langley free-flight tunnel. In the investigation it was found that with flaps neutral satisfactory flight behavior at low speeds was obtainable with an increase in height of the vertical tail and with the inboard slats opened. In the flap-down slat-open condition the longitudinal stability was satisfactory, but it was impossible to obtain satisfactory lateral-flight characteristics even with the increase in height of the vertical tail because of the negative effective dihedral, low directional stability, and large-adverse yawing moments of the ailerons.
Open shop scheduling problem to minimize total weighted completion time
NASA Astrophysics Data System (ADS)
Bai, Danyu; Zhang, Zhihai; Zhang, Qiang; Tang, Mengqian
2017-01-01
A given number of jobs in an open shop scheduling environment must each be processed for given amounts of time on each of a given set of machines in an arbitrary sequence. This study aims to achieve a schedule that minimizes total weighted completion time. Owing to the strong NP-hardness of the problem, the weighted shortest processing time block (WSPTB) heuristic is presented to obtain approximate solutions for large-scale problems. Performance analysis proves the asymptotic optimality of the WSPTB heuristic in the sense of probability limits. The largest weight block rule is provided to seek optimal schedules in polynomial time for a special case. A hybrid discrete differential evolution algorithm is designed to obtain high-quality solutions for moderate-scale problems. Simulation experiments demonstrate the effectiveness of the proposed algorithms.
Global assessment of human losses due to earthquakes
Silva, Vitor; Jaiswal, Kishor; Weatherill, Graeme; Crowley, Helen
2014-01-01
Current studies have demonstrated a sharp increase in human losses due to earthquakes. These alarming levels of casualties suggest the need for large-scale investment in seismic risk mitigation, which, in turn, requires an adequate understanding of the extent of the losses, and location of the most affected regions. Recent developments in global and uniform datasets such as instrumental and historical earthquake catalogues, population spatial distribution and country-based vulnerability functions, have opened an unprecedented possibility for a reliable assessment of earthquake consequences at a global scale. In this study, a uniform probabilistic seismic hazard assessment (PSHA) model was employed to derive a set of global seismic hazard curves, using the open-source software OpenQuake for seismic hazard and risk analysis. These results were combined with a collection of empirical fatality vulnerability functions and a population dataset to calculate average annual human losses at the country level. The results from this study highlight the regions/countries in the world with a higher seismic risk, and thus where risk reduction measures should be prioritized.
Parameterization and scaling of Arctic ice conditions in the context of ice-atmosphere processes
NASA Technical Reports Server (NTRS)
Barry, R. G.; Heinrichs, J.; Steffen, K.; Maslanik, J. A.; Key, J.; Serreze, M. C.; Weaver, R. W.
1994-01-01
This report summarizes achievements during year three of our project to investigate the use of ERS-1 SAR data to study Arctic ice and ice/atmosphere processes. The project was granted a one year extension, and goals for the final year are outlined. The specific objects of the project are to determine how the development and evolution of open water/thin ice areas within the interior ice pack vary under different atmospheric synoptic regimes; compare how open water/thin ice fractions estimated from large-area divergence measurements differ from fractions determined by summing localized openings in the pack; relate these questions of scale and process to methods of observation, modeling, and averaging over time and space; determine whether SAR data might be used to calibrate ice concentration estimates from medium and low-rate bit sensors (AVHRR and DMSP-OLS) and the special sensor microwave imager (SSM/I); and investigate methods to integrate SAR data for turbulent heat flux parametrization at the atmosphere interface with other satellite data.
Soenksen, L R; Kassis, T; Noh, M; Griffith, L G; Trumper, D L
2018-03-13
Precise fluid height sensing in open-channel microfluidics has long been a desirable feature for a wide range of applications. However, performing accurate measurements of the fluid level in small-scale reservoirs (<1 mL) has proven to be an elusive goal, especially if direct fluid-sensor contact needs to be avoided. In particular, gravity-driven systems used in several microfluidic applications to establish pressure gradients and impose flow remain open-loop and largely unmonitored due to these sensing limitations. Here we present an optimized self-shielded coplanar capacitive sensor design and automated control system to provide submillimeter fluid-height resolution (∼250 μm) and control of small-scale open reservoirs without the need for direct fluid contact. Results from testing and validation of our optimized sensor and system also suggest that accurate fluid height information can be used to robustly characterize, calibrate and dynamically control a range of microfluidic systems with complex pumping mechanisms, even in cell culture conditions. Capacitive sensing technology provides a scalable and cost-effective way to enable continuous monitoring and closed-loop feedback control of fluid volumes in small-scale gravity-dominated wells in a variety of microfluidic applications.
Jiang, Hanlun; Sheong, Fu Kit; Zhu, Lizhe; Gao, Xin; Bernauer, Julie; Huang, Xuhui
2015-07-01
Argonaute (Ago) proteins and microRNAs (miRNAs) are central components in RNA interference, which is a key cellular mechanism for sequence-specific gene silencing. Despite intensive studies, molecular mechanisms of how Ago recognizes miRNA remain largely elusive. In this study, we propose a two-step mechanism for this molecular recognition: selective binding followed by structural re-arrangement. Our model is based on the results of a combination of Markov State Models (MSMs), large-scale protein-RNA docking, and molecular dynamics (MD) simulations. Using MSMs, we identify an open state of apo human Ago-2 in fast equilibrium with partially open and closed states. Conformations in this open state are distinguished by their largely exposed binding grooves that can geometrically accommodate miRNA as indicated in our protein-RNA docking studies. miRNA may then selectively bind to these open conformations. Upon the initial binding, the complex may perform further structural re-arrangement as shown in our MD simulations and eventually reach the stable binary complex structure. Our results provide novel insights in Ago-miRNA recognition mechanisms and our methodology holds great potential to be widely applied in the studies of other important molecular recognition systems.
NASA Astrophysics Data System (ADS)
Draper, Martin; Usera, Gabriel
2015-04-01
The Scale Dependent Dynamic Model (SDDM) has been widely validated in large-eddy simulations using pseudo-spectral codes [1][2][3]. The scale dependency, particularly the potential law, has been proved also in a priori studies [4][5]. To the authors' knowledge there have been only few attempts to use the SDDM in finite difference (FD) and finite volume (FV) codes [6][7], finding some improvements with the dynamic procedures (scale independent or scale dependent approach), but not showing the behavior of the scale-dependence parameter when using the SDDM. The aim of the present paper is to evaluate the SDDM in the open source code caffa3d.MBRi, an updated version of the code presented in [8]. caffa3d.MBRi is a FV code, second-order accurate, parallelized with MPI, in which the domain is divided in unstructured blocks of structured grids. To accomplish this, 2 cases are considered: flow between flat plates and flow over a rough surface with the presence of a model wind turbine, taking for this case the experimental data presented in [9]. In both cases the standard Smagorinsky Model (SM), the Scale Independent Dynamic Model (SIDM) and the SDDM are tested. As presented in [6][7] slight improvements are obtained with the SDDM. Nevertheless, the behavior of the scale-dependence parameter supports the generalization of the dynamic procedure proposed in the SDDM, particularly taking into account that no explicit filter is used (the implicit filter is unknown). [1] F. Porté-Agel, C. Meneveau, M.B. Parlange. "A scale-dependent dynamic model for large-eddy simulation: application to a neutral atmospheric boundary layer". Journal of Fluid Mechanics, 2000, 415, 261-284. [2] E. Bou-Zeid, C. Meneveau, M. Parlante. "A scale-dependent Lagrangian dynamic model for large eddy simulation of complex turbulent flows". Physics of Fluids, 2005, 17, 025105 (18p). [3] R. Stoll, F. Porté-Agel. "Dynamic subgrid-scale models for momentum and scalar fluxes in large-eddy simulations of neutrally stratified atmospheric boundary layers over heterogeneous terrain". Water Resources Research, 2006, 42, WO1409 (18 p). [4] J. Keissl, M. Parlange, C. Meneveau. "Field experimental study of dynamic Smagorinsky models in the atmospheric surface layer". Journal of the Atmospheric Science, 2004, 61, 2296-2307. [5] E. Bou-Zeid, N. Vercauteren, M.B. Parlange, C. Meneveau. "Scale dependence of subgrid-scale model coefficients: An a priori study". Physics of Fluids, 2008, 20, 115106. [6] G. Kirkil, J. Mirocha, E. Bou-Zeid, F.K. Chow, B. Kosovic, "Implementation and evaluation of dynamic subfilter - scale stress models for large - eddy simulation using WRF". Monthly Weather Review, 2012, 140, 266-284. [7] S. Radhakrishnan, U. Piomelli. "Large-eddy simulation of oscillating boundary layers: model comparison and validation". Journal of Geophysical Research, 2008, 113, C02022. [8] G. Usera, A. Vernet, J.A. Ferré. "A parallel block-structured finite volume method for flows in complex geometry with sliding interfaces". Flow, Turbulence and Combustion, 2008, 81, 471-495. [9] Y-T. Wu, F. Porté-Agel. "Large-eddy simulation of wind-turbine wakes: evaluation of turbine parametrisations". BoundaryLayerMeteorology, 2011, 138, 345-366.
Li, Chen; Yongbo, Lv; Chi, Chen
2015-01-01
Based on the data from 30 provincial regions in China, an assessment and empirical analysis was carried out on the utilizing and sharing of the large-scale scientific equipment with a comprehensive assessment model established on the three dimensions, namely, equipment, utilization and sharing. The assessment results were interpreted in light of relevant policies. The results showed that on the whole, the overall development level in the provincial regions in eastern and central China is higher than that in western China. This is mostly because of the large gap among the different provincial regions with respect to the equipped level. But in terms of utilizing and sharing, some of the Western provincial regions, such as Ningxia, perform well, which is worthy of our attention. Policy adjustment targeting at the differentiation, elevation of the capacity of the equipment management personnel, perfection of the sharing and cooperation platform, and the promotion of the establishment of open sharing funds, are all important measures to promote the utilization and sharing of the large-scale scientific equipment and to narrow the gap among different regions. PMID:25937850
Panoptes: web-based exploration of large scale genome variation data.
Vauterin, Paul; Jeffery, Ben; Miles, Alistair; Amato, Roberto; Hart, Lee; Wright, Ian; Kwiatkowski, Dominic
2017-10-15
The size and complexity of modern large-scale genome variation studies demand novel approaches for exploring and sharing the data. In order to unlock the potential of these data for a broad audience of scientists with various areas of expertise, a unified exploration framework is required that is accessible, coherent and user-friendly. Panoptes is an open-source software framework for collaborative visual exploration of large-scale genome variation data and associated metadata in a web browser. It relies on technology choices that allow it to operate in near real-time on very large datasets. It can be used to browse rich, hybrid content in a coherent way, and offers interactive visual analytics approaches to assist the exploration. We illustrate its application using genome variation data of Anopheles gambiae, Plasmodium falciparum and Plasmodium vivax. Freely available at https://github.com/cggh/panoptes, under the GNU Affero General Public License. paul.vauterin@gmail.com. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com
The Relationship Between Galaxies and the Large-Scale Structure of the Universe
NASA Astrophysics Data System (ADS)
Coil, Alison L.
2018-06-01
I will describe our current understanding of the relationship between galaxies and the large-scale structure of the Universe, often called the galaxy-halo connection. Galaxies are thought to form and evolve in the centers of dark matter halos, which grow along with the galaxies they host. Large galaxy redshift surveys have revealed clear observational signatures of connections between galaxy properties and their clustering properties on large scales. For example, older, quiescent galaxies are known to cluster more strongly than younger, star-forming galaxies, which are more likely to be found in galactic voids and filaments rather than the centers of galaxy clusters. I will show how cosmological numerical simulations have aided our understanding of this galaxy-halo connection and what is known from a statistical point of view about how galaxies populate dark matter halos. This knowledge both helps us learn about galaxy evolution and is fundamental to our ability to use galaxy surveys to reveal cosmological information. I will talk briefly about some of the current open questions in the field, including galactic conformity and assembly bias.
Li, Wenke; Wayne, Gregory S; Lallaman, John E; Chang, Sou-Jen; Wittenberger, Steven J
2006-02-17
Ketoester 1 is cyclized to give pyran-3,5-dione 2 in 78% yield using a parallel addition of ketoester 1 and base NaO(t)Bu in refluxing THF. Compared to the previously reported procedures, these optimized conditions have significantly increased the yield of this transformation and the quality of pyran 2 and prove to be suitable for large-scale preparation. An application of 2 to the synthesis of ABT-598, a potassium channel opener, is demonstrated.
1983-07-01
complex suite of physiological and 39 respiratory adaptations (Ultch 1976) enable this species to occupy both open water and littoral zone environments...be an adaptation to avoid fish predation on juveniles in open water and may explain the spring peak in funnel trap captures near shore (Fig. 10). 79...population of C. picta probably is not established on Lake Conway. The one collected individual 103 defecated gastropod (Viviparous sp.) shells and
Harris, Michael J; Woo, Hyung-June
2008-11-01
Energetics of conformational changes experienced by an ATP-bound myosin head detached from actin was studied by all-atom explicit water umbrella sampling simulations. The statistics of coupling between large scale domain movements and smaller scale structural features were examined, including the closing of the ATP binding pocket, and a number of key hydrogen bond formations shown to play roles in structural and biochemical studies. The statistics for the ATP binding pocket open/close transition show an evolution of the relative stability from the open state in the early stages of the recovery stroke to the stable closed state after the stroke. The change in solvation environment of the fluorescence probe Trp507 (scallop numbering; 501 in Dictyostelium discoideum) indicates that the probe faithfully reflects the closing of the binding pocket as previously shown in experimental studies, while being directly coupled to roughly the early half of the overall large scale conformational change of the converter domain rotation. The free energy change of this solvation environment change, in particular, is -1.3 kcal/mol, in close agreement with experimental estimates. In addition, our results provide direct molecular level data allowing for interpretations of the fluorescence experiments of myosin conformational change in terms of the de-solvation of Trp side chain.
Designing artificial 2D crystals with site and size controlled quantum dots.
Xie, Xuejun; Kang, Jiahao; Cao, Wei; Chu, Jae Hwan; Gong, Yongji; Ajayan, Pulickel M; Banerjee, Kaustav
2017-08-30
Ordered arrays of quantum dots in two-dimensional (2D) materials would make promising optical materials, but their assembly could prove challenging. Here we demonstrate a scalable, site and size controlled fabrication of quantum dots in monolayer molybdenum disulfide (MoS 2 ), and quantum dot arrays with nanometer-scale spatial density by focused electron beam irradiation induced local 2H to 1T phase change in MoS 2 . By designing the quantum dots in a 2D superlattice, we show that new energy bands form where the new band gap can be controlled by the size and pitch of the quantum dots in the superlattice. The band gap can be tuned from 1.81 eV to 1.42 eV without loss of its photoluminescence performance, which provides new directions for fabricating lasers with designed wavelengths. Our work constitutes a photoresist-free, top-down method to create large-area quantum dot arrays with nanometer-scale spatial density that allow the quantum dots to interfere with each other and create artificial crystals. This technique opens up new pathways for fabricating light emitting devices with 2D materials at desired wavelengths. This demonstration can also enable the assembly of large scale quantum information systems and open up new avenues for the design of artificial 2D materials.
Sachem: a chemical cartridge for high-performance substructure search.
Kratochvíl, Miroslav; Vondrášek, Jiří; Galgonek, Jakub
2018-05-23
Structure search is one of the valuable capabilities of small-molecule databases. Fingerprint-based screening methods are usually employed to enhance the search performance by reducing the number of calls to the verification procedure. In substructure search, fingerprints are designed to capture important structural aspects of the molecule to aid the decision about whether the molecule contains a given substructure. Currently available cartridges typically provide acceptable search performance for processing user queries, but do not scale satisfactorily with dataset size. We present Sachem, a new open-source chemical cartridge that implements two substructure search methods: The first is a performance-oriented reimplementation of substructure indexing based on the OrChem fingerprint, and the second is a novel method that employs newly designed fingerprints stored in inverted indices. We assessed the performance of both methods on small, medium, and large datasets containing 1, 10, and 94 million compounds, respectively. Comparison of Sachem with other freely available cartridges revealed improvements in overall performance, scaling potential and screen-out efficiency. The Sachem cartridge allows efficient substructure searches in databases of all sizes. The sublinear performance scaling of the second method and the ability to efficiently query large amounts of pre-extracted information may together open the door to new applications for substructure searches.
Investigation of multilayer domains in large-scale CVD monolayer graphene by optical imaging
NASA Astrophysics Data System (ADS)
Yu, Yuanfang; Li, Zhenzhen; Wang, Wenhui; Guo, Xitao; Jiang, Jie; Nan, Haiyan; Ni, Zhenhua
2017-03-01
CVD graphene is a promising candidate for optoelectronic applications due to its high quality and high yield. However, multi-layer domains could inevitably form at the nucleation centers during the growth. Here, we propose an optical imaging technique to precisely identify the multilayer domains and also the ratio of their coverage in large-scale CVD monolayer graphene. We have also shown that the stacking disorder in twisted bilayer graphene as well as the impurities on the graphene surface could be distinguished by optical imaging. Finally, we investigated the effects of bilayer domains on the optical and electrical properties of CVD graphene, and found that the carrier mobility of CVD graphene is seriously limited by scattering from bilayer domains. Our results could be useful for guiding future optoelectronic applications of large-scale CVD graphene. Project supported by the National Natural Science Foundation of China (Nos. 61422503, 61376104), the Open Research Funds of Key Laboratory of MEMS of Ministry of Education (SEU, China), and the Fundamental Research Funds for the Central Universities.
Energetics and Structural Characterization of the large-scale Functional Motion of Adenylate Kinase
Formoso, Elena; Limongelli, Vittorio; Parrinello, Michele
2015-01-01
Adenylate Kinase (AK) is a signal transducing protein that regulates cellular energy homeostasis balancing between different conformations. An alteration of its activity can lead to severe pathologies such as heart failure, cancer and neurodegenerative diseases. A comprehensive elucidation of the large-scale conformational motions that rule the functional mechanism of this enzyme is of great value to guide rationally the development of new medications. Here using a metadynamics-based computational protocol we elucidate the thermodynamics and structural properties underlying the AK functional transitions. The free energy estimation of the conformational motions of the enzyme allows characterizing the sequence of events that regulate its action. We reveal the atomistic details of the most relevant enzyme states, identifying residues such as Arg119 and Lys13, which play a key role during the conformational transitions and represent druggable spots to design enzyme inhibitors. Our study offers tools that open new areas of investigation on large-scale motion in proteins. PMID:25672826
Energetics and Structural Characterization of the large-scale Functional Motion of Adenylate Kinase
NASA Astrophysics Data System (ADS)
Formoso, Elena; Limongelli, Vittorio; Parrinello, Michele
2015-02-01
Adenylate Kinase (AK) is a signal transducing protein that regulates cellular energy homeostasis balancing between different conformations. An alteration of its activity can lead to severe pathologies such as heart failure, cancer and neurodegenerative diseases. A comprehensive elucidation of the large-scale conformational motions that rule the functional mechanism of this enzyme is of great value to guide rationally the development of new medications. Here using a metadynamics-based computational protocol we elucidate the thermodynamics and structural properties underlying the AK functional transitions. The free energy estimation of the conformational motions of the enzyme allows characterizing the sequence of events that regulate its action. We reveal the atomistic details of the most relevant enzyme states, identifying residues such as Arg119 and Lys13, which play a key role during the conformational transitions and represent druggable spots to design enzyme inhibitors. Our study offers tools that open new areas of investigation on large-scale motion in proteins.
Open Reading Frame Phylogenetic Analysis on the Cloud
2013-01-01
Phylogenetic analysis has become essential in researching the evolutionary relationships between viruses. These relationships are depicted on phylogenetic trees, in which viruses are grouped based on sequence similarity. Viral evolutionary relationships are identified from open reading frames rather than from complete sequences. Recently, cloud computing has become popular for developing internet-based bioinformatics tools. Biocloud is an efficient, scalable, and robust bioinformatics computing service. In this paper, we propose a cloud-based open reading frame phylogenetic analysis service. The proposed service integrates the Hadoop framework, virtualization technology, and phylogenetic analysis methods to provide a high-availability, large-scale bioservice. In a case study, we analyze the phylogenetic relationships among Norovirus. Evolutionary relationships are elucidated by aligning different open reading frame sequences. The proposed platform correctly identifies the evolutionary relationships between members of Norovirus. PMID:23671843
Zaehringer, Julie G; Wambugu, Grace; Kiteme, Boniface; Eckert, Sandra
2018-05-01
Africa has been heavily targeted by large-scale agricultural investments (LAIs) throughout the last decade, with scarcely known impacts on local social-ecological systems. In Kenya, a large number of LAIs were made in the region northwest of Mount Kenya. These large-scale farms produce vegetables and flowers mainly for European markets. However, land use in the region remains dominated by small-scale crop and livestock farms with less than 1 ha of land each, who produce both for their own subsistence and for the local markets. We interviewed 100 small-scale farmers living near five different LAIs to elicit their perceptions of the impacts that these LAIs have on their land use and the overall environment. Furthermore, we analyzed remotely sensed land cover and land use data to assess land use change in the vicinity of the five LAIs. While land use change did not follow a clear trend, a number of small-scale farmers did adapt their crop management to environmental changes such as a reduced river water flows and increased pests, which they attributed to the presence of LAIs. Despite the high number of open conflicts between small-scale land users and LAIs around the issue of river water abstraction, the main environmental impact, felt by almost half of the interviewed land users, was air pollution with agrochemicals sprayed on the LAIs' land. Even though only a low percentage of local land users and their household members were directly involved with LAIs, a large majority of respondents favored the presence of LAIs nearby, as they are believed to contribute to the region's overall economic development. Copyright © 2018 Elsevier Ltd. All rights reserved.
Keeping Connected: A Review of the Research Relationship
ERIC Educational Resources Information Center
Moss, Julianne; Hay, Trevor
2014-01-01
In this paper, some key findings of the Keeping Connected project are discussed in light of the methodological challenges of developing an analytical approach in a large-scale study, particularly in starting with open-ended, participant-selected, digital still visual images as part of 31 longitudinal case studies. The paper works to clarify the…
Improving Mobile MOOC Learning via Implicit Physiological Signal Sensing
ERIC Educational Resources Information Center
Xiao, Xiang
2017-01-01
Massive Open Online Courses (MOOCs) are becoming a promising solution for delivering high-quality education on a large scale at low cost in recent years. Despite the great potential, today's MOOCs also suffer from challenges such as low student engagement, lack of personalization, and most importantly, lack of direct, immediate feedback channels…
Mentors' Perceptions of Their Own Professional Development during Mentoring
ERIC Educational Resources Information Center
Lopez-Real, Francis; Kwan, Tammy
2005-01-01
During 2001-2002 the Faculty of Education at the University of Hong Kong undertook a large-scale evaluation of the school-university partnership scheme. As part of this study, all mentors were asked how, if at all, the mentoring process had enhanced their own professional development. An open-response questionnaire was administered with a…
Purposes and Effects of Lying.
ERIC Educational Resources Information Center
Hample, Dale
Three exploratory studies were aimed at describing the purposes of lies and the consequences of lying. Data were collected through a partly open-ended questionnaire, a content analysis of several tape-recorded interviews, and a large-scale survey. The results showed that two of every three lies were told for selfish reasons, while three of every…
The Opening of Higher Education
ERIC Educational Resources Information Center
Matkin, Gary W.
2012-01-01
In a 1974 report presented to the Organisation for Economic Co-operation and Development (OECD), Martin Trow laid out a framework for understanding large-scale, worldwide changes in higher education. Trow's essay also pointed to the problems that "arise out of the transition from one phase to another in a broad pattern of development of higher…
Deutsch, Eric W.; Mendoza, Luis; Shteynberg, David; Slagel, Joseph; Sun, Zhi; Moritz, Robert L.
2015-01-01
Democratization of genomics technologies has enabled the rapid determination of genotypes. More recently the democratization of comprehensive proteomics technologies is enabling the determination of the cellular phenotype and the molecular events that define its dynamic state. Core proteomic technologies include mass spectrometry to define protein sequence, protein:protein interactions, and protein post-translational modifications. Key enabling technologies for proteomics are bioinformatic pipelines to identify, quantitate, and summarize these events. The Trans-Proteomics Pipeline (TPP) is a robust open-source standardized data processing pipeline for large-scale reproducible quantitative mass spectrometry proteomics. It supports all major operating systems and instrument vendors via open data formats. Here we provide a review of the overall proteomics workflow supported by the TPP, its major tools, and how it can be used in its various modes from desktop to cloud computing. We describe new features for the TPP, including data visualization functionality. We conclude by describing some common perils that affect the analysis of tandem mass spectrometry datasets, as well as some major upcoming features. PMID:25631240
Deutsch, Eric W; Mendoza, Luis; Shteynberg, David; Slagel, Joseph; Sun, Zhi; Moritz, Robert L
2015-08-01
Democratization of genomics technologies has enabled the rapid determination of genotypes. More recently the democratization of comprehensive proteomics technologies is enabling the determination of the cellular phenotype and the molecular events that define its dynamic state. Core proteomic technologies include MS to define protein sequence, protein:protein interactions, and protein PTMs. Key enabling technologies for proteomics are bioinformatic pipelines to identify, quantitate, and summarize these events. The Trans-Proteomics Pipeline (TPP) is a robust open-source standardized data processing pipeline for large-scale reproducible quantitative MS proteomics. It supports all major operating systems and instrument vendors via open data formats. Here, we provide a review of the overall proteomics workflow supported by the TPP, its major tools, and how it can be used in its various modes from desktop to cloud computing. We describe new features for the TPP, including data visualization functionality. We conclude by describing some common perils that affect the analysis of MS/MS datasets, as well as some major upcoming features. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Karim, Md Rezaul; Michel, Audrey; Zappa, Achille; Baranov, Pavel; Sahay, Ratnesh; Rebholz-Schuhmann, Dietrich
2017-04-16
Data workflow systems (DWFSs) enable bioinformatics researchers to combine components for data access and data analytics, and to share the final data analytics approach with their collaborators. Increasingly, such systems have to cope with large-scale data, such as full genomes (about 200 GB each), public fact repositories (about 100 TB of data) and 3D imaging data at even larger scales. As moving the data becomes cumbersome, the DWFS needs to embed its processes into a cloud infrastructure, where the data are already hosted. As the standardized public data play an increasingly important role, the DWFS needs to comply with Semantic Web technologies. This advancement to DWFS would reduce overhead costs and accelerate the progress in bioinformatics research based on large-scale data and public resources, as researchers would require less specialized IT knowledge for the implementation. Furthermore, the high data growth rates in bioinformatics research drive the demand for parallel and distributed computing, which then imposes a need for scalability and high-throughput capabilities onto the DWFS. As a result, requirements for data sharing and access to public knowledge bases suggest that compliance of the DWFS with Semantic Web standards is necessary. In this article, we will analyze the existing DWFS with regard to their capabilities toward public open data use as well as large-scale computational and human interface requirements. We untangle the parameters for selecting a preferable solution for bioinformatics research with particular consideration to using cloud services and Semantic Web technologies. Our analysis leads to research guidelines and recommendations toward the development of future DWFS for the bioinformatics research community. © The Author 2017. Published by Oxford University Press.
Kasperowski, Dick; Hillman, Thomas
2018-05-01
In the past decade, some areas of science have begun turning to masses of online volunteers through open calls for generating and classifying very large sets of data. The purpose of this study is to investigate the epistemic culture of a large-scale online citizen science project, the Galaxy Zoo, that turns to volunteers for the classification of images of galaxies. For this task, we chose to apply the concepts of programs and antiprograms to examine the 'essential tensions' that arise in relation to the mobilizing values of a citizen science project and the epistemic subjects and cultures that are enacted by its volunteers. Our premise is that these tensions reveal central features of the epistemic subjects and distributed cognition of epistemic cultures in these large-scale citizen science projects.
Tethys – A Python Package for Spatial and Temporal Downscaling of Global Water Withdrawals
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, Xinya; Vernon, Chris R.; Hejazi, Mohamad I.
Downscaling of water withdrawals from regional/national to local scale is a fundamental step and also a common problem when integrating large scale economic and integrated assessment models with high-resolution detailed sectoral models. Tethys, an open-access software written in Python, is developed with statistical downscaling algorithms, to spatially and temporally downscale water withdrawal data to a finer scale. The spatial resolution will be downscaled from region/basin scale to grid (0.5 geographic degree) scale and the temporal resolution will be downscaled from year to month. Tethys is used to produce monthly global gridded water withdrawal products based on estimates from the Globalmore » Change Assessment Model (GCAM).« less
Tethys – A Python Package for Spatial and Temporal Downscaling of Global Water Withdrawals
Li, Xinya; Vernon, Chris R.; Hejazi, Mohamad I.; ...
2018-02-09
Downscaling of water withdrawals from regional/national to local scale is a fundamental step and also a common problem when integrating large scale economic and integrated assessment models with high-resolution detailed sectoral models. Tethys, an open-access software written in Python, is developed with statistical downscaling algorithms, to spatially and temporally downscale water withdrawal data to a finer scale. The spatial resolution will be downscaled from region/basin scale to grid (0.5 geographic degree) scale and the temporal resolution will be downscaled from year to month. Tethys is used to produce monthly global gridded water withdrawal products based on estimates from the Globalmore » Change Assessment Model (GCAM).« less
Distributions and Changes of Carbonate Parameters Along the U.S. East Coast
NASA Astrophysics Data System (ADS)
Xu, Y. Y.; Cai, W. J.; Wanninkhof, R. H.; Salisbury, J., II
2017-12-01
On top of anthropogenic climate change, upwelling, eutrophication, river discharge, and interactions with the open ocean have affected carbonate chemistry in coastal waters. In this study, we present the large-scale variations of carbonate parameters along the U.S. east coast using in situ observations obtained during an East Coast Ocean Acidification (ECOA) cruise in summer 2015. Compare with previous large-scale cruises along the east coast, the ECOA cruise increases spatial coverage in the Gulf of Marine region and has more offshore stations for a better understanding of carbon dynamics in coastal waters and their interactions with open ocean waters. Our results show that the spatial distribution of water mass properties set up the large-scale advection of salt and heat and the distribution of total alkalinity (TA). However, dissolved inorganic carbon (DIC) shows a distinct pattern. Coastal water pH displays high variability in the Gulf of Maine and the Mid-Atlantic Bight (MAB). But it is relatively homogeneous in the South Atlantic Bight (SAB). In contrast, the distribution of aragonite saturation state (Ω) has an increase pattern from north to south similar to those of TA, SST, and SSS. A mechanistic discussion will be presented to understand the controls on Ω in eastern coastal waters. A comparison with previous cruises also suggests very different changes of pH and Ω in the MAB and SAB. Preliminary analysis suggests an overall increase in surface pH and Ω in the MAB. In contrast, pH and Ω in the SAB surface waters decrease over the past two decades. This work serves as a platform for the monitoring of large-scale carbon cycling in the U.S. east coast. It is also important to identify the physical and biogeochemical processes that affect these distributions and changes over time for a better understanding of carbon cycling and ocean acidification in coastal waters.
NASA Astrophysics Data System (ADS)
Alvera-Azcarate, A.; Barth, A.; Virmani, J. I.; Weisberg, R. H.
2007-05-01
The Intra-Americas Sea (IAS) surface circulation is characterized by large scale currents. The Caribbean current, which originates in the Lesser Antilles, travels westwards through the Caribbean Sea and eastern Mexico and passes through the Gulf of Mexico to finally form the Gulf Stream. This complex system of currents is also characterized by a high mesoscale variability, such as eddies and meanders. The objectives of this work are twofold: first, the multi-scale surface circulation of the IAS is described using satellite altimetry. The topographic influence of the different basins forming the IAS, the characteristic time and spatial scales, and the time variability of the surface circulation will be addressed. The second objective is to analyze the influence of this large scale circulation on a small scale coastal domain with a ROMS-based model of the Cariaco basin (Venezuela). Cariaco is a deep (1400 m), semi-enclosed basin connected to the open ocean by two shallow channels (Tortuga and Centinela Channels). Its connection with the open sea, and therefore the ventilation of the basin, occurs in the surface layers. The Cariaco ROMS model will be used to study the exchanges of mass, heat and salt through the channels. A 1/60 degree ROMS model nested in the global 1/12 degree HYCOM model from the Naval Research Laboratory will be used for this study. In addition, a series of observations (satellite altimetry and in situ temperature, salinity and velocity data), will be used to assess the influence of the Caribbean circulation on the basin.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kraus, Michaela; Nickeler, Dieter H.; Liimets, Tiina
The Galactic object MWC 137 has been suggested to belong to the group of B[e] supergiants. However, with its large-scale optical bipolar ring nebula and high-velocity jet and knots, it is a rather atypical representative of this class. We performed multiwavelength observations spreading from the optical to the radio regimes. Based on optical imaging and long-slit spectroscopic data, we found that the northern parts of the large-scale nebula are predominantly blueshifted, while the southern regions appear mostly redshifted. We developed a geometrical model consisting of two double cones. Although various observational features can be approximated with such a scenario, themore » observed velocity pattern is more complex. Using near-infrared integral-field unit spectroscopy, we studied the hot molecular gas in the vicinity of the star. The emission from the hot CO gas arises in a small-scale disk revolving around the star on Keplerian orbits. Although the disk itself cannot be spatially resolved, its emission is reflected by the dust arranged in arc-like structures and the clumps surrounding MWC 137 on small scales. In the radio regime, we mapped the cold molecular gas in the outskirts of the optical nebula. We found that large amounts of cool molecular gas and warm dust embrace the optical nebula in the east, south, and west. No cold gas or dust was detected in the north and northwestern regions. Despite the new insights into the nebula kinematics gained from our studies, the real formation scenario of the large-scale nebula remains an open issue.« less
NASA Astrophysics Data System (ADS)
Carollo, Federico; Garrahan, Juan P.; Lesanovsky, Igor; Pérez-Espigares, Carlos
2017-11-01
We consider a class of either fermionic or bosonic noninteracting open quantum chains driven by dissipative interactions at the boundaries and study the interplay of coherent transport and dissipative processes, such as bulk dephasing and diffusion. Starting from the microscopic formulation, we show that the dynamics on large scales can be described in terms of fluctuating hydrodynamics. This is an important simplification as it allows us to apply the methods of macroscopic fluctuation theory to compute the large deviation (LD) statistics of time-integrated currents. In particular, this permits us to show that fermionic open chains display a third-order dynamical phase transition in LD functions. We show that this transition is manifested in a singular change in the structure of trajectories: while typical trajectories are diffusive, rare trajectories associated with atypical currents are ballistic and hyperuniform in their spatial structure. We confirm these results by numerically simulating ensembles of rare trajectories via the cloning method, and by exact numerical diagonalization of the microscopic quantum generator.
Carollo, Federico; Garrahan, Juan P; Lesanovsky, Igor; Pérez-Espigares, Carlos
2017-11-01
We consider a class of either fermionic or bosonic noninteracting open quantum chains driven by dissipative interactions at the boundaries and study the interplay of coherent transport and dissipative processes, such as bulk dephasing and diffusion. Starting from the microscopic formulation, we show that the dynamics on large scales can be described in terms of fluctuating hydrodynamics. This is an important simplification as it allows us to apply the methods of macroscopic fluctuation theory to compute the large deviation (LD) statistics of time-integrated currents. In particular, this permits us to show that fermionic open chains display a third-order dynamical phase transition in LD functions. We show that this transition is manifested in a singular change in the structure of trajectories: while typical trajectories are diffusive, rare trajectories associated with atypical currents are ballistic and hyperuniform in their spatial structure. We confirm these results by numerically simulating ensembles of rare trajectories via the cloning method, and by exact numerical diagonalization of the microscopic quantum generator.
NASA Astrophysics Data System (ADS)
Bianchelli, S.; Gambi, C.; Zeppilli, D.; Danovaro, R.
2010-03-01
Metazoan meiofaunal abundance, total biomass, nematode size and the richness of taxa were investigated along bathymetric gradients (from the shelf break down to ca. 5000-m depth) in six submarine canyons and on five adjacent open slopes of three deep-sea regions. The investigated areas were distributed along >2500 km, on the Portuguese to the Catalan and South Adriatic margins. The Portuguese and Catalan margins displayed the highest abundances, biomass and richness of taxa, while the lowest values were observed in the Central Mediterranean Sea. The comparison between canyons and the nearby open slopes showed the lack of significant differences in terms of meiofaunal abundance and biomass at any sampling depth. In most canyons and on most slopes, meiofaunal variables did not display consistent bathymetric patterns. Conversely, we found that the different topographic features were apparently responsible for significant differences in the abundance and distribution of the rare meiofaunal taxa (i.e. taxa accounting for <1% of total meiofaunal abundance). Several taxa belonging to the temporary meiofauna, such as larvae/juveniles of Priapulida, Holothuroidea, Ascidiacea and Cnidaria, were encountered exclusively on open slopes, while others (including the Tanaidacea and Echinodea larvae) were found exclusively in canyons sediments. Results reported here indicate that, at large spatial scales, differences in deep-sea meiofaunal abundance and biomass are not only controlled by the available food sources, but also by the region or habitat specific topographic features, which apparently play a key role in the distribution of rare benthic taxa.
OpenMP Parallelization and Optimization of Graph-Based Machine Learning Algorithms
Meng, Zhaoyi; Koniges, Alice; He, Yun Helen; ...
2016-09-21
In this paper, we investigate the OpenMP parallelization and optimization of two novel data classification algorithms. The new algorithms are based on graph and PDE solution techniques and provide significant accuracy and performance advantages over traditional data classification algorithms in serial mode. The methods leverage the Nystrom extension to calculate eigenvalue/eigenvectors of the graph Laplacian and this is a self-contained module that can be used in conjunction with other graph-Laplacian based methods such as spectral clustering. We use performance tools to collect the hotspots and memory access of the serial codes and use OpenMP as the parallelization language to parallelizemore » the most time-consuming parts. Where possible, we also use library routines. We then optimize the OpenMP implementations and detail the performance on traditional supercomputer nodes (in our case a Cray XC30), and test the optimization steps on emerging testbed systems based on Intel’s Knights Corner and Landing processors. We show both performance improvement and strong scaling behavior. Finally, a large number of optimization techniques and analyses are necessary before the algorithm reaches almost ideal scaling.« less
NASA Astrophysics Data System (ADS)
Federico, Ivan; Oddo, Paolo; Pinardi, Nadia; Coppini, Giovanni
2014-05-01
The Southern Adriatic Northern Ionian Forecasting System (SANIFS) operational chain is based on a nesting approach. The large scale model for the entire Mediterranean basin (MFS, Mediterranean Forecasting system, operated by INGV, e.g. Tonani et al. 2008, Oddo et al. 2009) provides lateral open boundary conditions to the regional model for Adriatic and Ionian seas (AIFS, Adriatic Ionian Forecasting System) which provides the open-sea fields (initial conditions and lateral open boundary conditions) to SANIFS. The latter, here presented, is a coastal ocean model based on SHYFEM (Shallow HYdrodynamics Finite Element Model) code, which is an unstructured grid, finite element three-dimensional hydrodynamic model (e.g. Umgiesser et al., 2004, Ferrarin et al., 2013). The SANIFS hydrodynamic model component has been designed to provide accurate information of hydrodynamics and active tracer fields in the coastal waters of Southern Eastern Italy (Apulia, Basilicata and Calabria regions), where the model is characterized by a resolution of about of 200-500 m. The horizontal resolution is also accurate in open-sea areas, where the elements size is approximately 3 km. During the development phase the model has been initialized and forced at the lateral open boundaries through a full nesting strategy directly with the MFS fields. The heat fluxes has been computed by bulk formulae using as input data the operational analyses of European Centre for Medium-Range Weather Forecasts. Short range pre-operational forecast tests have been performed in different seasons to evaluate the robustness of the implemented model in different oceanographic conditions. Model results are validated by means of comparison with MFS operational results and observations. The model is able to reproduce the large-scale oceanographic structures of the area (keeping similar structures of MFS in open sea), while in the coastal area significant improvements in terms of reproduced structures and dynamics are evident.
Rey-Villamizar, Nicolas; Somasundar, Vinay; Megjhani, Murad; Xu, Yan; Lu, Yanbin; Padmanabhan, Raghav; Trett, Kristen; Shain, William; Roysam, Badri
2014-01-01
In this article, we describe the use of Python for large-scale automated server-based bio-image analysis in FARSIGHT, a free and open-source toolkit of image analysis methods for quantitative studies of complex and dynamic tissue microenvironments imaged by modern optical microscopes, including confocal, multi-spectral, multi-photon, and time-lapse systems. The core FARSIGHT modules for image segmentation, feature extraction, tracking, and machine learning are written in C++, leveraging widely used libraries including ITK, VTK, Boost, and Qt. For solving complex image analysis tasks, these modules must be combined into scripts using Python. As a concrete example, we consider the problem of analyzing 3-D multi-spectral images of brain tissue surrounding implanted neuroprosthetic devices, acquired using high-throughput multi-spectral spinning disk step-and-repeat confocal microscopy. The resulting images typically contain 5 fluorescent channels. Each channel consists of 6000 × 10,000 × 500 voxels with 16 bits/voxel, implying image sizes exceeding 250 GB. These images must be mosaicked, pre-processed to overcome imaging artifacts, and segmented to enable cellular-scale feature extraction. The features are used to identify cell types, and perform large-scale analysis for identifying spatial distributions of specific cell types relative to the device. Python was used to build a server-based script (Dell 910 PowerEdge servers with 4 sockets/server with 10 cores each, 2 threads per core and 1TB of RAM running on Red Hat Enterprise Linux linked to a RAID 5 SAN) capable of routinely handling image datasets at this scale and performing all these processing steps in a collaborative multi-user multi-platform environment. Our Python script enables efficient data storage and movement between computers and storage servers, logs all the processing steps, and performs full multi-threaded execution of all codes, including open and closed-source third party libraries.
An Open-Source Galaxy Redshift Survey Simulator for next-generation Large Scale Structure Surveys
NASA Astrophysics Data System (ADS)
Seijak, Uros
Galaxy redshift surveys produce three-dimensional maps of the galaxy distribution. On large scales these maps trace the underlying matter fluctuations in a relatively simple manner, so that the properties of the primordial fluctuations along with the overall expansion history and growth of perturbations can be extracted. The BAO standard ruler method to measure the expansion history of the universe using galaxy redshift surveys is thought to be robust to observational artifacts and understood theoretically with high precision. These same surveys can offer a host of additional information, including a measurement of the growth rate of large scale structure through redshift space distortions, the possibility of measuring the sum of neutrino masses, tighter constraints on the expansion history through the Alcock-Paczynski effect, and constraints on the scale-dependence and non-Gaussianity of the primordial fluctuations. Extracting this broadband clustering information hinges on both our ability to minimize and subtract observational systematics to the observed galaxy power spectrum, and our ability to model the broadband behavior of the observed galaxy power spectrum with exquisite precision. Rapid development on both fronts is required to capitalize on WFIRST's data set. We propose to develop an open-source computational toolbox that will propel development in both areas by connecting large scale structure modeling and instrument and survey modeling with the statistical inference process. We will use the proposed simulator to both tailor perturbation theory and fully non-linear models of the broadband clustering of WFIRST galaxies and discover novel observables in the non-linear regime that are robust to observational systematics and able to distinguish between a wide range of spatial and dynamic biasing models for the WFIRST galaxy redshift survey sources. We have demonstrated the utility of this approach in a pilot study of the SDSS-III BOSS galaxies, in which we improved the redshift space distortion growth rate measurement precision by a factor of 2.5 using customized clustering statistics in the non-linear regime that were immunized against observational systematics. We look forward to addressing the unique challenges of modeling and empirically characterizing the WFIRST galaxies and observational systematics.
The spatial and temporal domains of modern ecology.
Estes, Lyndon; Elsen, Paul R; Treuer, Timothy; Ahmed, Labeeb; Caylor, Kelly; Chang, Jason; Choi, Jonathan J; Ellis, Erle C
2018-05-01
To understand ecological phenomena, it is necessary to observe their behaviour across multiple spatial and temporal scales. Since this need was first highlighted in the 1980s, technology has opened previously inaccessible scales to observation. To help to determine whether there have been corresponding changes in the scales observed by modern ecologists, we analysed the resolution, extent, interval and duration of observations (excluding experiments) in 348 studies that have been published between 2004 and 2014. We found that observational scales were generally narrow, because ecologists still primarily use conventional field techniques. In the spatial domain, most observations had resolutions ≤1 m 2 and extents ≤10,000 ha. In the temporal domain, most observations were either unreplicated or infrequently repeated (>1 month interval) and ≤1 year in duration. Compared with studies conducted before 2004, observational durations and resolutions appear largely unchanged, but intervals have become finer and extents larger. We also found a large gulf between the scales at which phenomena are actually observed and the scales those observations ostensibly represent, raising concerns about observational comprehensiveness. Furthermore, most studies did not clearly report scale, suggesting that it remains a minor concern. Ecologists can better understand the scales represented by observations by incorporating autocorrelation measures, while journals can promote attentiveness to scale by implementing scale-reporting standards.
NASA Astrophysics Data System (ADS)
Li, Lee; Liu, Lun; Liu, Yun-Long; Bin, Yu; Ge, Ya-Feng; Lin, Fo-Chang
2014-01-01
Atmospheric air diffuse plasmas have enormous application potential in various fields of science and technology. Without dielectric barrier, generating large-scale air diffuse plasmas is always a challenging issue. This paper discusses and analyses the formation mechanism of cold homogenous plasma. It is proposed that generating stable diffuse atmospheric plasmas in open air should meet the three conditions: high transient power with low average power, excitation in low average E-field with locally high E-field region, and multiple overlapping electron avalanches. Accordingly, an experimental configuration of generating large-scale barrier-free diffuse air plasmas is designed. Based on runaway electron theory, a low duty-ratio, high voltage repetitive nanosecond pulse generator is chosen as a discharge excitation source. Using the wire-electrodes with small curvature radius, the gaps with highly non-uniform E-field are structured. Experimental results show that the volume-scaleable, barrier-free, homogeneous air non-thermal plasmas have been obtained between the gap spacing with the copper-wire electrodes. The area of air cold plasmas has been up to hundreds of square centimeters. The proposed formation conditions of large-scale barrier-free diffuse air plasmas are proved to be reasonable and feasible.
A survey on routing protocols for large-scale wireless sensor networks.
Li, Changle; Zhang, Hanxiao; Hao, Binbin; Li, Jiandong
2011-01-01
With the advances in micro-electronics, wireless sensor devices have been made much smaller and more integrated, and large-scale wireless sensor networks (WSNs) based the cooperation among the significant amount of nodes have become a hot topic. "Large-scale" means mainly large area or high density of a network. Accordingly the routing protocols must scale well to the network scope extension and node density increases. A sensor node is normally energy-limited and cannot be recharged, and thus its energy consumption has a quite significant effect on the scalability of the protocol. To the best of our knowledge, currently the mainstream methods to solve the energy problem in large-scale WSNs are the hierarchical routing protocols. In a hierarchical routing protocol, all the nodes are divided into several groups with different assignment levels. The nodes within the high level are responsible for data aggregation and management work, and the low level nodes for sensing their surroundings and collecting information. The hierarchical routing protocols are proved to be more energy-efficient than flat ones in which all the nodes play the same role, especially in terms of the data aggregation and the flooding of the control packets. With focus on the hierarchical structure, in this paper we provide an insight into routing protocols designed specifically for large-scale WSNs. According to the different objectives, the protocols are generally classified based on different criteria such as control overhead reduction, energy consumption mitigation and energy balance. In order to gain a comprehensive understanding of each protocol, we highlight their innovative ideas, describe the underlying principles in detail and analyze their advantages and disadvantages. Moreover a comparison of each routing protocol is conducted to demonstrate the differences between the protocols in terms of message complexity, memory requirements, localization, data aggregation, clustering manner and other metrics. Finally some open issues in routing protocol design in large-scale wireless sensor networks and conclusions are proposed.
Electrokinetic decontamination of concrete. Final report, August 3, 1993--September 15, 1996
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
1998-12-31
The ELECTROSORB{reg_sign} {open_quotes}C{close_quotes} process is an electrokinetic process for decontaminating concrete. ELECTROSORB{reg_sign} {open_quotes}C{close_quotes} uses a carpet-like extraction pad which is placed on the contaminated concrete surface. An electrolyte solution is circulated from a supporting module. This module keeps the electrolyte solution clean. The work is advancing through the engineering development stage with steady progress toward a full scale demonstration unit which will be ready for incorporation in the DOE Large Scale Demonstration Program by Summer 1997. A demonstration was carried out at the Mound Facility in Miamisburg, Ohio, in June 1996. Third party verification by EG&G verified the effectiveness ofmore » the process. Results of this work and the development work that proceeded are described herein.« less
NASA Technical Reports Server (NTRS)
Moog, R. D.; Bacchus, D. L.; Utreja, L. R.
1979-01-01
The aerodynamic performance characteristics have been determined for the Space Shuttle Solid Rocket Booster drogue, main, and pilot parachutes. The performance evaluation on the 20-degree conical ribbon parachutes is based primarily on air drop tests of full scale prototype parachutes. In addition, parametric wind tunnel tests were performed and used in parachute configuration development and preliminary performance assessments. The wind tunnel test data are compared to the drop test results and both sets of data are used to determine the predicted performance of the Solid Rocket Booster flight parachutes. Data from other drop tests of large ribbon parachutes are also compared with the Solid Rocket Booster parachute performance characteristics. Parameters assessed include full open terminal drag coefficients, reefed drag area, opening characteristics, clustering effects, and forebody interference.
Inside Success: Strategies of 25 Effective Small High Schools in NYC. Executive Summary
ERIC Educational Resources Information Center
Villavicencio, Adriana; Marinell, William H.
2014-01-01
For decades, New York City's high school graduation rates hovered at or below 50 percent. In attempt to turn around these disappointing results, the NYC Department of Education enacted a series of large-scale reforms, including opening hundreds of new "small schools of choice" (SSCs). Recent research by MDRC has shown that these schools…
Inside Success: Strategies of 25 Effective Small High Schools in NYC. Technical Appendices
ERIC Educational Resources Information Center
Villavicencio, Adriana; Marinell, William H.
2014-01-01
For decades, New York City's high school graduation rates hovered at or below 50 percent. In attempt to turn around these disappointing results, the NYC Department of Education enacted a series of large-scale reforms, including opening hundreds of new "small schools of choice" (SSCs). Recent research by MDRC has shown that these schools…
Inside Success: Strategies of 25 Effective Small High Schools in NYC. Report
ERIC Educational Resources Information Center
Villavicencio, Adriana; Marinell, William H.
2014-01-01
For decades, New York City's high school graduation rates hovered at or below 50 percent. In attempt to turn around these disappointing results, the NYC Department of Education enacted a series of large-scale reforms, including opening hundreds of new "small schools of choice" (SSCs). Recent research by MDRC has shown that these schools…
Chemical process simulation has long been used as a design tool in the development of chemical plants, and has long been considered a means to evaluate different design options. With the advent of large scale computer networks and interface models for program components, it is po...
Experience of public procurement of Open Compute servers
NASA Astrophysics Data System (ADS)
Bärring, Olof; Guerri, Marco; Bonfillou, Eric; Valsan, Liviu; Grigore, Alexandru; Dore, Vincent; Gentit, Alain; Clement, Benoît; Grossir, Anthony
2015-12-01
The Open Compute Project. OCP (http://www.opencompute.org/). was launched by Facebook in 2011 with the objective of building efficient computing infrastructures at the lowest possible cost. The technologies are released as open hardware. with the goal to develop servers and data centres following the model traditionally associated with open source software projects. In 2013 CERN acquired a few OCP servers in order to compare performance and power consumption with standard hardware. The conclusions were that there are sufficient savings to motivate an attempt to procure a large scale installation. One objective is to evaluate if the OCP market is sufficiently mature and broad enough to meet the constraints of a public procurement. This paper summarizes this procurement. which started in September 2014 and involved the Request for information (RFI) to qualify bidders and Request for Tender (RFT).
Leveraging human oversight and intervention in large-scale parallel processing of open-source data
NASA Astrophysics Data System (ADS)
Casini, Enrico; Suri, Niranjan; Bradshaw, Jeffrey M.
2015-05-01
The popularity of cloud computing along with the increased availability of cheap storage have led to the necessity of elaboration and transformation of large volumes of open-source data, all in parallel. One way to handle such extensive volumes of information properly is to take advantage of distributed computing frameworks like Map-Reduce. Unfortunately, an entirely automated approach that excludes human intervention is often unpredictable and error prone. Highly accurate data processing and decision-making can be achieved by supporting an automatic process through human collaboration, in a variety of environments such as warfare, cyber security and threat monitoring. Although this mutual participation seems easily exploitable, human-machine collaboration in the field of data analysis presents several challenges. First, due to the asynchronous nature of human intervention, it is necessary to verify that once a correction is made, all the necessary reprocessing is done in chain. Second, it is often needed to minimize the amount of reprocessing in order to optimize the usage of resources due to limited availability. In order to improve on these strict requirements, this paper introduces improvements to an innovative approach for human-machine collaboration in the processing of large amounts of open-source data in parallel.
Depressed emission between magnetic arcades near a sunspot
NASA Astrophysics Data System (ADS)
Ryabov, B. I.; Shibasaki, K.
The locations of the depressed emission in microwaves, EUV and soft X-rays are compared with each other and with the location of the plasma outflow in the active region (AR) 8535 on the Sun. We found that two open-field regions overlap the regions of depressed emission near the AR's sunspot. These two open-field regions are simulated with the potential-field source-surface (PFSS) model under radial distances of RSS = 1.8 R⊙ and RSS = 2.5 R⊙. Each open-field region is located between the arcades of the loops of the same magnetic polarity. The former open-field region covers the region of the plasma outflow, which is thus useful for the tests on connection to the heliosphere. The utmost microwave depression of the intensity in the ordinary mode (the Very Large Array 15 GHz observations) also overlaps the region of the plasma outflow and thus indicates this outflow. The lasting for eight days depression in soft X-rays and the SOHO EIT 2.84× 10-8 m images are attributed to the evacuation of as hot coronal plasma as T≥ 2× 106 K from the extended in height (``open") magnetic structures. We conclude that the AR 8535 presents the sunspot atmosphere affected by the large-scale magnetic fields.
Repurposing of open data through large scale hydrological modelling - hypeweb.smhi.se
NASA Astrophysics Data System (ADS)
Strömbäck, Lena; Andersson, Jafet; Donnelly, Chantal; Gustafsson, David; Isberg, Kristina; Pechlivanidis, Ilias; Strömqvist, Johan; Arheimer, Berit
2015-04-01
Hydrological modelling demands large amounts of spatial data, such as soil properties, land use, topography, lakes and reservoirs, ice and snow coverage, water management (e.g. irrigation patterns and regulations), meteorological data and observed water discharge in rivers. By using such data, the hydrological model will in turn provide new data that can be used for new purposes (i.e. re-purposing). This presentation will give an example of how readily available open data from public portals have been re-purposed by using the Hydrological Predictions for the Environment (HYPE) model in a number of large-scale model applications covering numerous subbasins and rivers. HYPE is a dynamic, semi-distributed, process-based, and integrated catchment model. The model output is launched as new Open Data at the web site www.hypeweb.smhi.se to be used for (i) Climate change impact assessments on water resources and dynamics; (ii) The European Water Framework Directive (WFD) for characterization and development of measure programs to improve the ecological status of water bodies; (iii) Design variables for infrastructure constructions; (iv) Spatial water-resource mapping; (v) Operational forecasts (1-10 days and seasonal) on floods and droughts; (vi) Input to oceanographic models for operational forecasts and marine status assessments; (vii) Research. The following regional domains have been modelled so far with different resolutions (number of subbasins within brackets): Sweden (37 000), Europe (35 000), Arctic basin (30 000), La Plata River (6 000), Niger River (800), Middle-East North-Africa (31 000), and the Indian subcontinent (6 000). The Hype web site provides several interactive web applications for exploring results from the models. The user can explore an overview of various water variables for historical and future conditions. Moreover the user can explore and download historical time series of discharge for each basin and explore the performance of the model towards observed river flow. The presentation will describe the Open Data sources used, show the functionality of the web site and discuss model performance and experience from this world-wide hydrological modelling of multi-basins using open data.
McCrae, Robert R; Scally, Matthew; Terracciano, Antonio; Abecasis, Gonçalo R; Costa, Paul T
2010-12-01
There is growing evidence that personality traits are affected by many genes, all of which have very small effects. As an alternative to the largely unsuccessful search for individual polymorphisms associated with personality traits, the authors identified large sets of potentially related single nucleotide polymorphisms (SNPs) and summed them to form molecular personality scales (MPSs) with from 4 to 2,497 SNPs. Scales were derived from two thirds of a large (N = 3,972) sample of individuals from Sardinia who completed the Revised NEO Personality Inventory (P. T. Costa, Jr., & R. R. McCrae, 1992) and were assessed in a genomewide association scan. When MPSs were correlated with the phenotype in the remaining one third of the sample, very small but significant associations were found for 4 of the 5e personality factors when the longest scales were examined. These data suggest that MPSs for Neuroticism, Openness to Experience, Agreeableness, and Conscientiousness (but not Extraversion) contain genetic information that can be refined in future studies, and the procedures described here should be applicable to other quantitative traits. PsycINFO Database Record (c) 2010 APA, all rights reserved.
Weighing trees with lasers: advances, challenges and opportunities
Boni Vicari, M.; Burt, A.; Calders, K.; Lewis, S. L.; Raumonen, P.; Wilkes, P.
2018-01-01
Terrestrial laser scanning (TLS) is providing exciting new ways to quantify tree and forest structure, particularly above-ground biomass (AGB). We show how TLS can address some of the key uncertainties and limitations of current approaches to estimating AGB based on empirical allometric scaling equations (ASEs) that underpin all large-scale estimates of AGB. TLS provides extremely detailed non-destructive measurements of tree form independent of tree size and shape. We show examples of three-dimensional (3D) TLS measurements from various tropical and temperate forests and describe how the resulting TLS point clouds can be used to produce quantitative 3D models of branch and trunk size, shape and distribution. These models can drastically improve estimates of AGB, provide new, improved large-scale ASEs, and deliver insights into a range of fundamental tree properties related to structure. Large quantities of detailed measurements of individual 3D tree structure also have the potential to open new and exciting avenues of research in areas where difficulties of measurement have until now prevented statistical approaches to detecting and understanding underlying patterns of scaling, form and function. We discuss these opportunities and some of the challenges that remain to be overcome to enable wider adoption of TLS methods. PMID:29503726
Effect of small scale transport processes on phytoplankton distribution in coastal seas.
Hernández-Carrasco, Ismael; Orfila, Alejandro; Rossi, Vincent; Garçon, Veronique
2018-06-05
Coastal ocean ecosystems are major contributors to the global biogeochemical cycles and biological productivity. Physical factors induced by the turbulent flow play a crucial role in regulating marine ecosystems. However, while large-scale open-ocean dynamics is well described by geostrophy, the role of multiscale transport processes in coastal regions is still poorly understood due to the lack of continuous high-resolution observations. Here, the influence of small-scale dynamics (O(3.5-25) km, i.e. spanning upper submesoscale and mesoscale processes) on surface phytoplankton derived from satellite chlorophyll-a (Chl-a) is studied using Lagrangian metrics computed from High-Frequency Radar currents. The combination of complementary Lagrangian diagnostics, including the Lagrangian divergence along fluid trajectories, provides an improved description of the 3D flow geometry which facilitates the interpretation of two non-exclusive physical mechanisms affecting phytoplankton dynamics and patchiness. Attracting small-scale fronts, unveiled by backwards Lagrangian Coherent Structures, are associated to negative divergence where particles and Chl-a standing stocks cluster. Filaments of positive divergence, representing large accumulated upward vertical velocities and suggesting accrued injection of subsurface nutrients, match areas with large Chl-a concentrations. Our findings demonstrate that an accurate characterization of small-scale transport processes is necessary to comprehend bio-physical interactions in coastal seas.
Saitow, Masaaki; Becker, Ute; Riplinger, Christoph; Valeev, Edward F; Neese, Frank
2017-04-28
The Coupled-Cluster expansion, truncated after single and double excitations (CCSD), provides accurate and reliable molecular electronic wave functions and energies for many molecular systems around their equilibrium geometries. However, the high computational cost, which is well-known to scale as O(N 6 ) with system size N, has limited its practical application to small systems consisting of not more than approximately 20-30 atoms. To overcome these limitations, low-order scaling approximations to CCSD have been intensively investigated over the past few years. In our previous work, we have shown that by combining the pair natural orbital (PNO) approach and the concept of orbital domains it is possible to achieve fully linear scaling CC implementations (DLPNO-CCSD and DLPNO-CCSD(T)) that recover around 99.9% of the total correlation energy [C. Riplinger et al., J. Chem. Phys. 144, 024109 (2016)]. The production level implementations of the DLPNO-CCSD and DLPNO-CCSD(T) methods were shown to be applicable to realistic systems composed of a few hundred atoms in a routine, black-box fashion on relatively modest hardware. In 2011, a reduced-scaling CCSD approach for high-spin open-shell unrestricted Hartree-Fock reference wave functions was proposed (UHF-LPNO-CCSD) [A. Hansen et al., J. Chem. Phys. 135, 214102 (2011)]. After a few years of experience with this method, a few shortcomings of UHF-LPNO-CCSD were noticed that required a redesign of the method, which is the subject of this paper. To this end, we employ the high-spin open-shell variant of the N-electron valence perturbation theory formalism to define the initial guess wave function, and consequently also the open-shell PNOs. The new PNO ansatz properly converges to the closed-shell limit since all truncations and approximations have been made in strict analogy to the closed-shell case. Furthermore, given the fact that the formalism uses a single set of orbitals, only a single PNO integral transformation is necessary, which offers large computational savings. We show that, with the default PNO truncation parameters, approximately 99.9% of the total CCSD correlation energy is recovered for open-shell species, which is comparable to the performance of the method for closed-shells. UHF-DLPNO-CCSD shows a linear scaling behavior for closed-shell systems, while linear to quadratic scaling is obtained for open-shell systems. The largest systems we have considered contain more than 500 atoms and feature more than 10 000 basis functions with a triple-ζ quality basis set.
NASA Astrophysics Data System (ADS)
Saitow, Masaaki; Becker, Ute; Riplinger, Christoph; Valeev, Edward F.; Neese, Frank
2017-04-01
The Coupled-Cluster expansion, truncated after single and double excitations (CCSD), provides accurate and reliable molecular electronic wave functions and energies for many molecular systems around their equilibrium geometries. However, the high computational cost, which is well-known to scale as O(N6) with system size N, has limited its practical application to small systems consisting of not more than approximately 20-30 atoms. To overcome these limitations, low-order scaling approximations to CCSD have been intensively investigated over the past few years. In our previous work, we have shown that by combining the pair natural orbital (PNO) approach and the concept of orbital domains it is possible to achieve fully linear scaling CC implementations (DLPNO-CCSD and DLPNO-CCSD(T)) that recover around 99.9% of the total correlation energy [C. Riplinger et al., J. Chem. Phys. 144, 024109 (2016)]. The production level implementations of the DLPNO-CCSD and DLPNO-CCSD(T) methods were shown to be applicable to realistic systems composed of a few hundred atoms in a routine, black-box fashion on relatively modest hardware. In 2011, a reduced-scaling CCSD approach for high-spin open-shell unrestricted Hartree-Fock reference wave functions was proposed (UHF-LPNO-CCSD) [A. Hansen et al., J. Chem. Phys. 135, 214102 (2011)]. After a few years of experience with this method, a few shortcomings of UHF-LPNO-CCSD were noticed that required a redesign of the method, which is the subject of this paper. To this end, we employ the high-spin open-shell variant of the N-electron valence perturbation theory formalism to define the initial guess wave function, and consequently also the open-shell PNOs. The new PNO ansatz properly converges to the closed-shell limit since all truncations and approximations have been made in strict analogy to the closed-shell case. Furthermore, given the fact that the formalism uses a single set of orbitals, only a single PNO integral transformation is necessary, which offers large computational savings. We show that, with the default PNO truncation parameters, approximately 99.9% of the total CCSD correlation energy is recovered for open-shell species, which is comparable to the performance of the method for closed-shells. UHF-DLPNO-CCSD shows a linear scaling behavior for closed-shell systems, while linear to quadratic scaling is obtained for open-shell systems. The largest systems we have considered contain more than 500 atoms and feature more than 10 000 basis functions with a triple-ζ quality basis set.
X-ray structures of LeuT in substrate-free outward-open and apo inward-open states
DOE Office of Scientific and Technical Information (OSTI.GOV)
Krishnamurthy, Harini; Gouaux, Eric
2012-08-09
Neurotransmitter sodium symporters are integral membrane proteins that remove chemical transmitters from the synapse and terminate neurotransmission mediated by serotonin, dopamine, noradrenaline, glycine and GABA ({gamma}-aminobutyric acid). Crystal structures of the bacterial homologue, LeuT, in substrate-bound outward-occluded and competitive inhibitor-bound outward-facing states have advanced our mechanistic understanding of neurotransmitter sodium symporters but have left fundamental questions unanswered. Here we report crystal structures of LeuT mutants in complexes with conformation-specific antibody fragments in the outward-open and inward-open states. In the absence of substrate but in the presence of sodium the transporter is outward-open, illustrating how the binding of substrate closes themore » extracellular gate through local conformational changes: hinge-bending movements of the extracellular halves of transmembrane domains 1, 2 and 6, together with translation of extracellular loop 4. The inward-open conformation, by contrast, involves large-scale conformational changes, including a reorientation of transmembrane domains 1, 2, 5, 6 and 7, a marked hinge bending of transmembrane domain 1a and occlusion of the extracellular vestibule by extracellular loop 4. These changes close the extracellular gate, open an intracellular vestibule, and largely disrupt the two sodium sites, thus providing a mechanism by which ions and substrate are released to the cytoplasm. The new structures establish a structural framework for the mechanism of neurotransmitter sodium symporters and their modulation by therapeutic and illicit substances.« less
Design Aspects of the Rayleigh Convection Code
NASA Astrophysics Data System (ADS)
Featherstone, N. A.
2017-12-01
Understanding the long-term generation of planetary or stellar magnetic field requires complementary knowledge of the large-scale fluid dynamics pervading large fractions of the object's interior. Such large-scale motions are sensitive to the system's geometry which, in planets and stars, is spherical to a good approximation. As a result, computational models designed to study such systems often solve the MHD equations in spherical geometry, frequently employing a spectral approach involving spherical harmonics. We present computational and user-interface design aspects of one such modeling tool, the Rayleigh convection code, which is suitable for deployment on desktop and petascale-hpc architectures alike. In this poster, we will present an overview of this code's parallel design and its built-in diagnostics-output package. Rayleigh has been developed with NSF support through the Computational Infrastructure for Geodynamics and is expected to be released as open-source software in winter 2017/2018.
Ocean Research Enabled by Underwater Gliders.
Rudnick, Daniel L
2016-01-01
Underwater gliders are autonomous underwater vehicles that profile vertically by changing their buoyancy and use wings to move horizontally. Gliders are useful for sustained observation at relatively fine horizontal scales, especially to connect the coastal and open ocean. In this review, research topics are grouped by time and length scales. Large-scale topics addressed include the eastern and western boundary currents and the regional effects of climate variability. The accessibility of horizontal length scales of order 1 km allows investigation of mesoscale and submesoscale features such as fronts and eddies. Because the submesoscales dominate vertical fluxes in the ocean, gliders have found application in studies of biogeochemical processes. At the finest scales, gliders have been used to measure internal waves and turbulent dissipation. The review summarizes gliders' achievements to date and assesses their future in ocean observation.
McCammon, J. Andrew
2011-01-01
Chagas' disease, caused by the protozoan parasite Trypanosoma cruzi (T. cruzi), is a life-threatening illness affecting 11–18 million people. Currently available treatments are limited, with unacceptable efficacy and safety profiles. Recent studies have revealed an essential T. cruzi proline racemase enzyme (TcPR) as an attractive candidate for improved chemotherapeutic intervention. Conformational changes associated with substrate binding to TcPR are believed to expose critical residues that elicit a host mitogenic B-cell response, a process contributing to parasite persistence and immune system evasion. Characterization of the conformational states of TcPR requires access to long-time-scale motions that are currently inaccessible by standard molecular dynamics simulations. Here we describe advanced accelerated molecular dynamics that extend the effective simulation time and capture large-scale motions of functional relevance. Conservation and fragment mapping analyses identified potential conformational epitopes located in the vicinity of newly identified transient binding pockets. The newly identified open TcPR conformations revealed by this study along with knowledge of the closed to open interconversion mechanism advances our understanding of TcPR function. The results and the strategy adopted in this work constitute an important step toward the rationalization of the molecular basis behind the mitogenic B-cell response of TcPR and provide new insights for future structure-based drug discovery. PMID:22022240
de Oliveira, César Augusto F; Grant, Barry J; Zhou, Michelle; McCammon, J Andrew
2011-10-01
Chagas' disease, caused by the protozoan parasite Trypanosoma cruzi (T. cruzi), is a life-threatening illness affecting 11-18 million people. Currently available treatments are limited, with unacceptable efficacy and safety profiles. Recent studies have revealed an essential T. cruzi proline racemase enzyme (TcPR) as an attractive candidate for improved chemotherapeutic intervention. Conformational changes associated with substrate binding to TcPR are believed to expose critical residues that elicit a host mitogenic B-cell response, a process contributing to parasite persistence and immune system evasion. Characterization of the conformational states of TcPR requires access to long-time-scale motions that are currently inaccessible by standard molecular dynamics simulations. Here we describe advanced accelerated molecular dynamics that extend the effective simulation time and capture large-scale motions of functional relevance. Conservation and fragment mapping analyses identified potential conformational epitopes located in the vicinity of newly identified transient binding pockets. The newly identified open TcPR conformations revealed by this study along with knowledge of the closed to open interconversion mechanism advances our understanding of TcPR function. The results and the strategy adopted in this work constitute an important step toward the rationalization of the molecular basis behind the mitogenic B-cell response of TcPR and provide new insights for future structure-based drug discovery.
WeaVR: a self-contained and wearable immersive virtual environment simulation system.
Hodgson, Eric; Bachmann, Eric R; Vincent, David; Zmuda, Michael; Waller, David; Calusdian, James
2015-03-01
We describe WeaVR, a computer simulation system that takes virtual reality technology beyond specialized laboratories and research sites and makes it available in any open space, such as a gymnasium or a public park. Novel hardware and software systems enable HMD-based immersive virtual reality simulations to be conducted in any arbitrary location, with no external infrastructure and little-to-no setup or site preparation. The ability of the WeaVR system to provide realistic motion-tracked navigation for users, to improve the study of large-scale navigation, and to generate usable behavioral data is shown in three demonstrations. First, participants navigated through a full-scale virtual grocery store while physically situated in an open grass field. Trajectory data are presented for both normal tracking and for tracking during the use of redirected walking that constrained users to a predefined area. Second, users followed a straight path within a virtual world for distances of up to 2 km while walking naturally and being redirected to stay within the field, demonstrating the ability of the system to study large-scale navigation by simulating virtual worlds that are potentially unlimited in extent. Finally, the portability and pedagogical implications of this system were demonstrated by taking it to a regional high school for live use by a computer science class on their own school campus.
TomoMiner and TomoMinerCloud: A software platform for large-scale subtomogram structural analysis
Frazier, Zachary; Xu, Min; Alber, Frank
2017-01-01
SUMMARY Cryo-electron tomography (cryoET) captures the 3D electron density distribution of macromolecular complexes in close to native state. With the rapid advance of cryoET acquisition technologies, it is possible to generate large numbers (>100,000) of subtomograms, each containing a macromolecular complex. Often, these subtomograms represent a heterogeneous sample due to variations in structure and composition of a complex in situ form or because particles are a mixture of different complexes. In this case subtomograms must be classified. However, classification of large numbers of subtomograms is a time-intensive task and often a limiting bottleneck. This paper introduces an open source software platform, TomoMiner, for large-scale subtomogram classification, template matching, subtomogram averaging, and alignment. Its scalable and robust parallel processing allows efficient classification of tens to hundreds of thousands of subtomograms. Additionally, TomoMiner provides a pre-configured TomoMinerCloud computing service permitting users without sufficient computing resources instant access to TomoMiners high-performance features. PMID:28552576
Adaptive latitudinal variation in Common Blackbird Turdus merula nest characteristics
Mainwaring, Mark C; Deeming, D Charles; Jones, Chris I; Hartley, Ian R
2014-01-01
Nest construction is taxonomically widespread, yet our understanding of adaptive intraspecific variation in nest design remains poor. Nest characteristics are expected to vary adaptively in response to predictable variation in spring temperatures over large spatial scales, yet such variation in nest design remains largely overlooked, particularly amongst open-cup-nesting birds. Here, we systematically examined the effects of latitudinal variation in spring temperatures and precipitation on the morphology, volume, composition, and insulatory properties of open-cup-nesting Common Blackbirds’ Turdus merula nests to test the hypothesis that birds living in cooler environments at more northerly latitudes would build better insulated nests than conspecifics living in warmer environments at more southerly latitudes. As spring temperatures increased with decreasing latitude, the external diameter of nests decreased. However, as nest wall thickness also decreased, there was no variation in the diameter of the internal nest cups. Only the mass of dry grasses within nests decreased with warmer temperatures at lower latitudes. The insulatory properties of nests declined with warmer temperatures at lower latitudes and nests containing greater amounts of dry grasses had higher insulatory properties. The insulatory properties of nests decreased with warmer temperatures at lower latitudes, via changes in morphology (wall thickness) and composition (dry grasses). Meanwhile, spring precipitation did not vary with latitude, and none of the nest characteristics varied with spring precipitation. This suggests that Common Blackbirds nesting at higher latitudes were building nests with thicker walls in order to counteract the cooler temperatures. We have provided evidence that the nest construction behavior of open-cup-nesting birds systematically varies in response to large-scale spatial variation in spring temperatures. PMID:24683466
Adaptive latitudinal variation in Common Blackbird Turdus merula nest characteristics.
Mainwaring, Mark C; Deeming, D Charles; Jones, Chris I; Hartley, Ian R
2014-03-01
Nest construction is taxonomically widespread, yet our understanding of adaptive intraspecific variation in nest design remains poor. Nest characteristics are expected to vary adaptively in response to predictable variation in spring temperatures over large spatial scales, yet such variation in nest design remains largely overlooked, particularly amongst open-cup-nesting birds. Here, we systematically examined the effects of latitudinal variation in spring temperatures and precipitation on the morphology, volume, composition, and insulatory properties of open-cup-nesting Common Blackbirds' Turdus merula nests to test the hypothesis that birds living in cooler environments at more northerly latitudes would build better insulated nests than conspecifics living in warmer environments at more southerly latitudes. As spring temperatures increased with decreasing latitude, the external diameter of nests decreased. However, as nest wall thickness also decreased, there was no variation in the diameter of the internal nest cups. Only the mass of dry grasses within nests decreased with warmer temperatures at lower latitudes. The insulatory properties of nests declined with warmer temperatures at lower latitudes and nests containing greater amounts of dry grasses had higher insulatory properties. The insulatory properties of nests decreased with warmer temperatures at lower latitudes, via changes in morphology (wall thickness) and composition (dry grasses). Meanwhile, spring precipitation did not vary with latitude, and none of the nest characteristics varied with spring precipitation. This suggests that Common Blackbirds nesting at higher latitudes were building nests with thicker walls in order to counteract the cooler temperatures. We have provided evidence that the nest construction behavior of open-cup-nesting birds systematically varies in response to large-scale spatial variation in spring temperatures.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chapin, M.A.; Mahaffie, M.J.; Tiller, G.M.
1996-12-31
Economics of most deep-water development projects require large reservoir volumes to be drained with relatively few wells. The presence of reservoir compartments must therefore be detected and planned for in a pre-development stage. We have used 3-D seismic data to constrain large-scale, deterministic reservoir bodies in a 3-D architecture model of Pliocene-turbidite sands of the {open_quotes}E{close_quotes} or {open_quotes}Pink{close_quotes} reservoir, Prospect Mars, Mississippi Canyon Areas 763 and 807, Gulf of Mexico. Reservoir compartmentalization is influenced by stratigraphic shingling, which in turn is caused by low accommodation space predentin the upper portion of a ponded seismic sequence within a salt withdrawal mini-basin.more » The accumulation is limited by updip onlap onto a condensed section marl, and by lateral truncation by a large scale submarine erosion surface. Compartments were suggested by RFT pressure variations and by geochemical analysis of RFT fluid samples. A geological interpretation derived from high-resolution 3-D seismic and three wells was linked to 3-D architecture models through seismic inversion, resulting in a reservoir all available data. Distinguishing subtle stratigraphical shingles from faults was accomplished by detailed, loop-level mapping, and was important to characterize the different types of reservoir compartments. Seismic inversion was used to detune the seismic amplitude, adjust sandbody thickness, and update the rock properties. Recent development wells confirm the architectural style identified. This modeling project illustrates how high-quality seismic data and architecture models can be combined in a pre-development phase of a prospect, in order to optimize well placement.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chapin, M.A.; Mahaffie, M.J.; Tiller, G.M.
1996-01-01
Economics of most deep-water development projects require large reservoir volumes to be drained with relatively few wells. The presence of reservoir compartments must therefore be detected and planned for in a pre-development stage. We have used 3-D seismic data to constrain large-scale, deterministic reservoir bodies in a 3-D architecture model of Pliocene-turbidite sands of the [open quotes]E[close quotes] or [open quotes]Pink[close quotes] reservoir, Prospect Mars, Mississippi Canyon Areas 763 and 807, Gulf of Mexico. Reservoir compartmentalization is influenced by stratigraphic shingling, which in turn is caused by low accommodation space predentin the upper portion of a ponded seismic sequence withinmore » a salt withdrawal mini-basin. The accumulation is limited by updip onlap onto a condensed section marl, and by lateral truncation by a large scale submarine erosion surface. Compartments were suggested by RFT pressure variations and by geochemical analysis of RFT fluid samples. A geological interpretation derived from high-resolution 3-D seismic and three wells was linked to 3-D architecture models through seismic inversion, resulting in a reservoir all available data. Distinguishing subtle stratigraphical shingles from faults was accomplished by detailed, loop-level mapping, and was important to characterize the different types of reservoir compartments. Seismic inversion was used to detune the seismic amplitude, adjust sandbody thickness, and update the rock properties. Recent development wells confirm the architectural style identified. This modeling project illustrates how high-quality seismic data and architecture models can be combined in a pre-development phase of a prospect, in order to optimize well placement.« less
Access Control Management for SCADA Systems
NASA Astrophysics Data System (ADS)
Hong, Seng-Phil; Ahn, Gail-Joon; Xu, Wenjuan
The information technology revolution has transformed all aspects of our society including critical infrastructures and led a significant shift from their old and disparate business models based on proprietary and legacy environments to more open and consolidated ones. Supervisory Control and Data Acquisition (SCADA) systems have been widely used not only for industrial processes but also for some experimental facilities. Due to the nature of open environments, managing SCADA systems should meet various security requirements since system administrators need to deal with a large number of entities and functions involved in critical infrastructures. In this paper, we identify necessary access control requirements in SCADA systems and articulate access control policies for the simulated SCADA systems. We also attempt to analyze and realize those requirements and policies in the context of role-based access control that is suitable for simplifying administrative tasks in large scale enterprises.
Köhler, Gunther; Pérez, Raúl Gómez Trejo; Petersen, Claus Bo P; de La Cruz, Fausto R Mendez
2014-01-09
We describe the new species Anolis peucephilus sp. nov. from the Pacific versant of southern Mexico. Anolis peucephilus differs from all congeners by having a combination of (1) smooth ventral scales; (2) usually a patch of three greatly enlarged supraocular scales; (3) extremely short hind legs, longest toe of adpressed hind leg reaching to a point between levels of axilla and ear opening, ratio shank length/snout-vent length 0.18-0.21; (4) circumnasal usually in contact with first supralabial; and (5) a large yellowish orange dewlap in males and a very small to small white dewlap in females. In external morphology, A. peucephilus is most similar to A. omiltemanus from which it differs by having even shorter hind legs with the longest toe of adpressed hind leg reaching to a point between levels of axilla and ear opening (versus usually to ear opening, occasionally to slightly beyond ear opening or to a point between shoulder and ear opening in A. omiltemanus), a slightly larger dewlap in females, to 64 mm2 (versus to 41 mm2 in A. omiltemanus), the circumnasal usually in contact with the first supralabial (versus those scales separated by the presence of a subnasal in A. omiltemanus), and 4-6 internasal scales in the new species (versus usually 6-7 in A. omiltemanus). Furthermore, A. peucephilus differs from A. omiltemanus in hemipenial morphology (no finger-like processus on asulcate side in A. peucephilus versus such a processus present in A. omiltemanus). Also, in a preliminary molecular genetic analysis of the mitochondrial CO1 gene fragment, A. peucephilus has a genetic distance of 11.5% from A. omiltemanus. Anolis peucephilus was collected at night while the lizards were sleeping in pine trees, 2-10 m above the ground.
Li, Zhengqi; Kuang, Min; Zhang, Jia; Han, Yunfeng; Zhu, Qunyi; Yang, Lianjie; Kong, Weiguang
2010-02-01
Cold airflow experiments were conducted to investigate the aerodynamic field in a small-scale furnace of a down-fired pulverized-coal 300 MW(e) utility boiler arranged with direct flow split burners enriched by cyclones. By increasing the staged-air ratio, a deflected flow field appeared in the lower furnace; larger staged-air ratios produced larger deflections. Industrial-sized experiments on a full-scale boiler were also performed at different staged-air damper openings with measurements taken of gas temperatures in the burner region and near the right-side wall, wall heat fluxes, and gas components (O(2), CO, and NO(x)) in the near-wall region. Combustion was unstable at staged-air damper openings below 30%. For openings of 30% and 40%, late ignition of the pulverized coal developed and large differences arose in gas temperatures and heat fluxes between the regions near the front and rear walls. In conjunction, carbon content in the fly ash was high and boiler efficiency was low with high NO(x) emission above 1200 mg/m(3) (at 6% O(2) dry). For fully open dampers, differences in gas temperatures and heat fluxes, carbon in fly ash and NO(x) emission decreased yielding an increase in boiler efficiency. The optimal setting is fully open staged-air dampers.
NASA Astrophysics Data System (ADS)
Dednam, W.; Botha, A. E.
2015-01-01
Solvation of bio-molecules in water is severely affected by the presence of co-solvent within the hydration shell of the solute structure. Furthermore, since solute molecules can range from small molecules, such as methane, to very large protein structures, it is imperative to understand the detailed structure-function relationship on the microscopic level. For example, it is useful know the conformational transitions that occur in protein structures. Although such an understanding can be obtained through large-scale molecular dynamic simulations, it is often the case that such simulations would require excessively large simulation times. In this context, Kirkwood-Buff theory, which connects the microscopic pair-wise molecular distributions to global thermodynamic properties, together with the recently developed technique, called finite size scaling, may provide a better method to reduce system sizes, and hence also the computational times. In this paper, we present molecular dynamics trial simulations of biologically relevant low-concentration solvents, solvated by aqueous co-solvent solutions. In particular we compare two different methods of calculating the relevant Kirkwood-Buff integrals. The first (traditional) method computes running integrals over the radial distribution functions, which must be obtained from large system-size NVT or NpT simulations. The second, newer method, employs finite size scaling to obtain the Kirkwood-Buff integrals directly by counting the particle number fluctuations in small, open sub-volumes embedded within a larger reservoir that can be well approximated by a much smaller simulation cell. In agreement with previous studies, which made a similar comparison for aqueous co-solvent solutions, without the additional solvent, we conclude that the finite size scaling method is also applicable to the present case, since it can produce computationally more efficient results which are equivalent to the more costly radial distribution function method.
NASA Astrophysics Data System (ADS)
Chatterjee, Tanmoy; Peet, Yulia T.
2017-07-01
A large eddy simulation (LES) methodology coupled with near-wall modeling has been implemented in the current study for high Re neutral atmospheric boundary layer flows using an exponentially accurate spectral element method in an open-source research code Nek 5000. The effect of artificial length scales due to subgrid scale (SGS) and near wall modeling (NWM) on the scaling laws and structure of the inner and outer layer eddies is studied using varying SGS and NWM parameters in the spectral element framework. The study provides an understanding of the various length scales and dynamics of the eddies affected by the LES model and also the fundamental physics behind the inner and outer layer eddies which are responsible for the correct behavior of the mean statistics in accordance with the definition of equilibrium layers by Townsend. An economical and accurate LES model based on capturing the near wall coherent eddies has been designed, which is successful in eliminating the artificial length scale effects like the log-layer mismatch or the secondary peak generation in the streamwise variance.
eIFL (Electronic Information for Libraries): A Global Initiative of the Soros Foundations Network.
ERIC Educational Resources Information Center
Feret, Blazej; Kay, Michael
This paper presents the history, current status, and future development of eIFL (Electronic Information for Libraries Direct)--a large-scale project run by the Soros Foundations Network and the Open Society Institute. The project aims to provide libraries in developing countries with access to a menu of electronic information resources. In 1999,…
e-Tutor: A Multilingual Open Educational Resource for Faculty Development to Teach Online
ERIC Educational Resources Information Center
Rapp, Christian; Gülbahar, Yasemin; Adnan, Muge
2016-01-01
The situation in Ukraine poses severe problems to the higher education system and to students in Eastern Ukraine. Many students and academicians had been compelled to leave their university buildings and move westwards. Hence, they are forced to substitute face-to-face teaching with distance learning, often on a large scale, but within a short…
USDA-ARS?s Scientific Manuscript database
Aquarius is a combined passive/active L-band microwave instrument developed to map the ocean surface salinity field from space. The primary science objective of this mission is to monitor the seasonal and interannual variation of the large scale features of the surface salinity field in the open oc...
The Management Aspect of the e-Portfolio as an Assessment Tool: Sample of Anadolu University
ERIC Educational Resources Information Center
Ozgur, Aydin Ziya; Kaya, Secil
2011-01-01
This article intends to introduce an e-portfolio system to help mentors assess the teacher candidates' performances and products in a large scale open and distance learning teacher training program. The Pre-School Teacher Training Program (PSTTP) of Anadolu University is a completely distance program that helps around 12.000 students get the…
ERIC Educational Resources Information Center
Wall, Kate; Higgins, Steve; Remedios, Richard; Rafferty, Victoria; Tiplady, Lucy
2013-01-01
A key challenge of visual methodology is how to combine large-scale qualitative data sets with epistemologically acceptable and rigorous analysis techniques. The authors argue that a pragmatic approach drawing on ideas from mixed methods is helpful to open up the full potential of visual data. However, before one starts to "mix" the…
USDA-ARS?s Scientific Manuscript database
As a cool season crop, pea (Pisum sativum L.) can tolerate frost at the vegetative stage but has yield loss when freezing stress occurs at reproductive stage. Cold tolerance improvement of pea varieties is important for the stable yield and the expansion of winter pea planting area. Under the natura...
ERIC Educational Resources Information Center
Greene, Jay; Loveless, Tom; MacLeod, W. Bentley; Nechyba, Thomas; Peterson, Paul; Rosenthal, Meredith; Whitehurst, Grover
2010-01-01
Choice is most frequently realized within the public sector using the mechanisms of residence, magnet schools, and open enrollment systems, whereas the voucher-like systems applauded by choice advocates and feared by opponents are extremely rare. Further, the charter sector is neither large enough nor sufficiently prepared to go to scale to…
NASA Technical Reports Server (NTRS)
Fisk, L. A.
2005-01-01
The purpose of this grant was to develop a theoretical understanding of the processes by which open magnetic flux undergoes large-scale transport in the solar corona, and to use this understanding to develop a predictive model for the heliospheric magnetic field, the configuration for which is determined by such motions.
What makes computational open source software libraries successful?
NASA Astrophysics Data System (ADS)
Bangerth, Wolfgang; Heister, Timo
2013-01-01
Software is the backbone of scientific computing. Yet, while we regularly publish detailed accounts about the results of scientific software, and while there is a general sense of which numerical methods work well, our community is largely unaware of best practices in writing the large-scale, open source scientific software upon which our discipline rests. This is particularly apparent in the commonly held view that writing successful software packages is largely the result of simply ‘being a good programmer’ when in fact there are many other factors involved, for example the social skill of community building. In this paper, we consider what we have found to be the necessary ingredients for successful scientific software projects and, in particular, for software libraries upon which the vast majority of scientific codes are built today. In particular, we discuss the roles of code, documentation, communities, project management and licenses. We also briefly comment on the impact on academic careers of engaging in software projects.
Simulation of fatigue crack growth under large scale yielding conditions
NASA Astrophysics Data System (ADS)
Schweizer, Christoph; Seifert, Thomas; Riedel, Hermann
2010-07-01
A simple mechanism based model for fatigue crack growth assumes a linear correlation between the cyclic crack-tip opening displacement (ΔCTOD) and the crack growth increment (da/dN). The objective of this work is to compare analytical estimates of ΔCTOD with results of numerical calculations under large scale yielding conditions and to verify the physical basis of the model by comparing the predicted and the measured evolution of the crack length in a 10%-chromium-steel. The material is described by a rate independent cyclic plasticity model with power-law hardening and Masing behavior. During the tension-going part of the cycle, nodes at the crack-tip are released such that the crack growth increment corresponds approximately to the crack-tip opening. The finite element analysis performed in ABAQUS is continued for so many cycles until a stabilized value of ΔCTOD is reached. The analytical model contains an interpolation formula for the J-integral, which is generalized to account for cyclic loading and crack closure. Both simulated and estimated ΔCTOD are reasonably consistent. The predicted crack length evolution is found to be in good agreement with the behavior of microcracks observed in a 10%-chromium steel.
Porsa, Sina; Lin, Yi-Chung; Pandy, Marcus G
2016-08-01
The aim of this study was to compare the computational performances of two direct methods for solving large-scale, nonlinear, optimal control problems in human movement. Direct shooting and direct collocation were implemented on an 8-segment, 48-muscle model of the body (24 muscles on each side) to compute the optimal control solution for maximum-height jumping. Both algorithms were executed on a freely-available musculoskeletal modeling platform called OpenSim. Direct collocation converged to essentially the same optimal solution up to 249 times faster than direct shooting when the same initial guess was assumed (3.4 h of CPU time for direct collocation vs. 35.3 days for direct shooting). The model predictions were in good agreement with the time histories of joint angles, ground reaction forces and muscle activation patterns measured for subjects jumping to their maximum achievable heights. Both methods converged to essentially the same solution when started from the same initial guess, but computation time was sensitive to the initial guess assumed. Direct collocation demonstrates exceptional computational performance and is well suited to performing predictive simulations of movement using large-scale musculoskeletal models.
NASA Astrophysics Data System (ADS)
Amann, Florian; Gischig, Valentin; Evans, Keith; Doetsch, Joseph; Jalali, Reza; Valley, Benoît; Krietsch, Hannes; Dutler, Nathan; Villiger, Linus; Brixel, Bernard; Klepikova, Maria; Kittilä, Anniina; Madonna, Claudio; Wiemer, Stefan; Saar, Martin O.; Loew, Simon; Driesner, Thomas; Maurer, Hansruedi; Giardini, Domenico
2018-02-01
In this contribution, we present a review of scientific research results that address seismo-hydromechanically coupled processes relevant for the development of a sustainable heat exchanger in low-permeability crystalline rock and introduce the design of the In situ Stimulation and Circulation (ISC) experiment at the Grimsel Test Site dedicated to studying such processes under controlled conditions. The review shows that research on reservoir stimulation for deep geothermal energy exploitation has been largely based on laboratory observations, large-scale projects and numerical models. Observations of full-scale reservoir stimulations have yielded important results. However, the limited access to the reservoir and limitations in the control on the experimental conditions during deep reservoir stimulations is insufficient to resolve the details of the hydromechanical processes that would enhance process understanding in a way that aids future stimulation design. Small-scale laboratory experiments provide fundamental insights into various processes relevant for enhanced geothermal energy, but suffer from (1) difficulties and uncertainties in upscaling the results to the field scale and (2) relatively homogeneous material and stress conditions that lead to an oversimplistic fracture flow and/or hydraulic fracture propagation behavior that is not representative of a heterogeneous reservoir. Thus, there is a need for intermediate-scale hydraulic stimulation experiments with high experimental control that bridge the various scales and for which access to the target rock mass with a comprehensive monitoring system is possible. The ISC experiment is designed to address open research questions in a naturally fractured and faulted crystalline rock mass at the Grimsel Test Site (Switzerland). Two hydraulic injection phases were executed to enhance the permeability of the rock mass. During the injection phases the rock mass deformation across fractures and within intact rock, the pore pressure distribution and propagation, and the microseismic response were monitored at a high spatial and temporal resolution.
A Survey on Routing Protocols for Large-Scale Wireless Sensor Networks
Li, Changle; Zhang, Hanxiao; Hao, Binbin; Li, Jiandong
2011-01-01
With the advances in micro-electronics, wireless sensor devices have been made much smaller and more integrated, and large-scale wireless sensor networks (WSNs) based the cooperation among the significant amount of nodes have become a hot topic. “Large-scale” means mainly large area or high density of a network. Accordingly the routing protocols must scale well to the network scope extension and node density increases. A sensor node is normally energy-limited and cannot be recharged, and thus its energy consumption has a quite significant effect on the scalability of the protocol. To the best of our knowledge, currently the mainstream methods to solve the energy problem in large-scale WSNs are the hierarchical routing protocols. In a hierarchical routing protocol, all the nodes are divided into several groups with different assignment levels. The nodes within the high level are responsible for data aggregation and management work, and the low level nodes for sensing their surroundings and collecting information. The hierarchical routing protocols are proved to be more energy-efficient than flat ones in which all the nodes play the same role, especially in terms of the data aggregation and the flooding of the control packets. With focus on the hierarchical structure, in this paper we provide an insight into routing protocols designed specifically for large-scale WSNs. According to the different objectives, the protocols are generally classified based on different criteria such as control overhead reduction, energy consumption mitigation and energy balance. In order to gain a comprehensive understanding of each protocol, we highlight their innovative ideas, describe the underlying principles in detail and analyze their advantages and disadvantages. Moreover a comparison of each routing protocol is conducted to demonstrate the differences between the protocols in terms of message complexity, memory requirements, localization, data aggregation, clustering manner and other metrics. Finally some open issues in routing protocol design in large-scale wireless sensor networks and conclusions are proposed. PMID:22163808
Carbon dioxide and methane emissions from the scale model of open dairy lots.
Ding, Luyu; Cao, Wei; Shi, Zhengxiang; Li, Baoming; Wang, Chaoyuan; Zhang, Guoqiang; Kristensen, Simon
2016-07-01
To investigate the impacts of major factors on carbon loss via gaseous emissions, carbon dioxide (CO2) and methane (CH4) emissions from the ground of open dairy lots were tested by a scale model experiment at various air temperatures (15, 25, and 35 °C), surface velocities (0.4, 0.7, 1.0, and 1.2 m sec(-1)), and floor types (unpaved soil floor and brick-paved floor) in controlled laboratory conditions using the wind tunnel method. Generally, CO2 and CH4 emissions were significantly enhanced with the increase of air temperature and velocity (P < 0.05). Floor type had different effects on the CO2 and CH4 emissions, which were also affected by air temperature and soil characteristics of the floor. Although different patterns were observed on CH4 emission from the soil and brick floors at different air temperature-velocity combinations, statistical analysis showed no significant difference in CH4 emissions from different floors (P > 0.05). For CO2, similar emissions were found from the soil and brick floors at 15 and 25 °C, whereas higher rates were detected from the brick floor at 35 °C (P < 0.05). Results showed that CH4 emission from the scale model was exponentially related to CO2 flux, which might be helpful in CH4 emission estimation from manure management. Gaseous emissions from the open lots are largely dependent on outdoor climate, floor systems, and management practices, which are quite different from those indoors. This study assessed the effects of floor types and air velocities on CO2 and CH4 emissions from the open dairy lots at various temperatures by a wind tunnel. It provided some valuable information for decision-making and further studies on gaseous emissions from open lots.
Universality in voting behavior: an empirical analysis
Chatterjee, Arnab; Mitrović, Marija; Fortunato, Santo
2013-01-01
Election data represent a precious source of information to study human behavior at a large scale. In proportional elections with open lists, the number of votes received by a candidate, rescaled by the average performance of all competitors in the same party list, has the same distribution regardless of the country and the year of the election. Here we provide the first thorough assessment of this claim. We analyzed election datasets of 15 countries with proportional systems. We confirm that a class of nations with similar election rules fulfill the universality claim. Discrepancies from this trend in other countries with open-lists elections are always associated with peculiar differences in the election rules, which matter more than differences between countries and historical periods. Our analysis shows that the role of parties in the electoral performance of candidates is crucial: alternative scalings not taking into account party affiliations lead to poor results. PMID:23308342
NASA Astrophysics Data System (ADS)
Terminanto, A.; Swantoro, H. A.; Hidayanto, A. N.
2017-12-01
Enterprise Resource Planning (ERP) is an integrated information system to manage business processes of companies of various business scales. Because of the high cost of ERP investment, ERP implementation is usually done in large-scale enterprises, Due to the complexity of implementation problems, the success rate of ERP implementation is still low. Open Source System ERP becomes an alternative choice of ERP application to SME companies in terms of cost and customization. This study aims to identify characteristics and configure the implementation of OSS ERP Payroll module in KKPS (Employee Cooperative PT SRI) using OSS ERP Odoo and using ASAP method. This study is classified into case study research and action research. Implementation of OSS ERP Payroll module is done because the HR section of KKPS has not been integrated with other parts. The results of this study are the characteristics and configuration of OSS ERP payroll module in KKPS.
GPU Accelerated Browser for Neuroimaging Genomics.
Zigon, Bob; Li, Huang; Yao, Xiaohui; Fang, Shiaofen; Hasan, Mohammad Al; Yan, Jingwen; Moore, Jason H; Saykin, Andrew J; Shen, Li
2018-04-25
Neuroimaging genomics is an emerging field that provides exciting opportunities to understand the genetic basis of brain structure and function. The unprecedented scale and complexity of the imaging and genomics data, however, have presented critical computational bottlenecks. In this work we present our initial efforts towards building an interactive visual exploratory system for mining big data in neuroimaging genomics. A GPU accelerated browsing tool for neuroimaging genomics is created that implements the ANOVA algorithm for single nucleotide polymorphism (SNP) based analysis and the VEGAS algorithm for gene-based analysis, and executes them at interactive rates. The ANOVA algorithm is 110 times faster than the 4-core OpenMP version, while the VEGAS algorithm is 375 times faster than its 4-core OpenMP counter part. This approach lays a solid foundation for researchers to address the challenges of mining large-scale imaging genomics datasets via interactive visual exploration.
Pérez-Rodríguez, Gael; Glez-Peña, Daniel; Azevedo, Nuno F; Pereira, Maria Olívia; Fdez-Riverola, Florentino; Lourenço, Anália
2015-03-01
Biofilms are receiving increasing attention from the biomedical community. Biofilm-like growth within human body is considered one of the key microbial strategies to augment resistance and persistence during infectious processes. The Biofilms Experiment Workbench is a novel software workbench for the operation and analysis of biofilms experimental data. The goal is to promote the interchange and comparison of data among laboratories, providing systematic, harmonised and large-scale data computation. The workbench was developed with AIBench, an open-source Java desktop application framework for scientific software development in the domain of translational biomedicine. Implementation favours free and open-source third-parties, such as the R statistical package, and reaches for the Web services of the BiofOmics database to enable public experiment deposition. First, we summarise the novel, free, open, XML-based interchange format for encoding biofilms experimental data. Then, we describe the execution of common scenarios of operation with the new workbench, such as the creation of new experiments, the importation of data from Excel spreadsheets, the computation of analytical results, the on-demand and highly customised construction of Web publishable reports, and the comparison of results between laboratories. A considerable and varied amount of biofilms data is being generated, and there is a critical need to develop bioinformatics tools that expedite the interchange and comparison of microbiological and clinical results among laboratories. We propose a simple, open-source software infrastructure which is effective, extensible and easy to understand. The workbench is freely available for non-commercial use at http://sing.ei.uvigo.es/bew under LGPL license. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Nursing Services Delivery Theory: an open system approach
Meyer, Raquel M; O’Brien-Pallas, Linda L
2010-01-01
meyer r.m. & o’brien-pallas l.l. (2010)Nursing services delivery theory: an open system approach. Journal of Advanced Nursing66(12), 2828–2838. Aim This paper is a discussion of the derivation of the Nursing Services Delivery Theory from the application of open system theory to large-scale organizations. Background The underlying mechanisms by which staffing indicators influence outcomes remain under-theorized and unmeasured, resulting in a ‘black box’ that masks the nature and organization of nursing work. Theory linking nursing work, staffing, work environments, and outcomes in different settings is urgently needed to inform management decisions about the allocation of nurse staffing resources in organizations. Data sources A search of CINAHL and Business Source Premier for the years 1980–2008 was conducted using the following terms: theory, models, organization, organizational structure, management, administration, nursing units, and nursing. Seminal works were included. Discussion The healthcare organization is conceptualized as an open system characterized by energy transformation, a dynamic steady state, negative entropy, event cycles, negative feedback, differentiation, integration and coordination, and equifinality. The Nursing Services Delivery Theory proposes that input, throughput, and output factors interact dynamically to influence the global work demands placed on nursing work groups at the point of care in production subsystems. Implications for nursing The Nursing Services Delivery Theory can be applied to varied settings, cultures, and countries and supports the study of multi-level phenomena and cross-level effects. Conclusion The Nursing Services Delivery Theory gives a relational structure for reconciling disparate streams of research related to nursing work, staffing, and work environments. The theory can guide future research and the management of nursing services in large-scale healthcare organizations. PMID:20831573
Progress toward openness, transparency, and reproducibility in cognitive neuroscience.
Gilmore, Rick O; Diaz, Michele T; Wyble, Brad A; Yarkoni, Tal
2017-05-01
Accumulating evidence suggests that many findings in psychological science and cognitive neuroscience may prove difficult to reproduce; statistical power in brain imaging studies is low and has not improved recently; software errors in analysis tools are common and can go undetected for many years; and, a few large-scale studies notwithstanding, open sharing of data, code, and materials remain the rare exception. At the same time, there is a renewed focus on reproducibility, transparency, and openness as essential core values in cognitive neuroscience. The emergence and rapid growth of data archives, meta-analytic tools, software pipelines, and research groups devoted to improved methodology reflect this new sensibility. We review evidence that the field has begun to embrace new open research practices and illustrate how these can begin to address problems of reproducibility, statistical power, and transparency in ways that will ultimately accelerate discovery. © 2017 New York Academy of Sciences.
Perspective: Markov models for long-timescale biomolecular dynamics.
Schwantes, C R; McGibbon, R T; Pande, V S
2014-09-07
Molecular dynamics simulations have the potential to provide atomic-level detail and insight to important questions in chemical physics that cannot be observed in typical experiments. However, simply generating a long trajectory is insufficient, as researchers must be able to transform the data in a simulation trajectory into specific scientific insights. Although this analysis step has often been taken for granted, it deserves further attention as large-scale simulations become increasingly routine. In this perspective, we discuss the application of Markov models to the analysis of large-scale biomolecular simulations. We draw attention to recent improvements in the construction of these models as well as several important open issues. In addition, we highlight recent theoretical advances that pave the way for a new generation of models of molecular kinetics.
NASA Astrophysics Data System (ADS)
Karavosov, R. K.; Prozorov, A. G.
2012-01-01
We have investigated the spectra of pressure pulsations in the near field of the open working section of the wind tunnel with a vortex flow behind the tunnel blower formed like the flow behind the hydroturbine of a hydraulic power plant. We have made a comparison between the measurement data for pressure pulsations and the air stream velocity in tunnels of the above type and in tunnels in which a large-scale vortex structure behind the blower is not formed. It has been established that the large-scale vortex formation in the incompressible medium behind the blade system in the wind tunnel is a source of narrow-band acoustic radiation capable of exciting resonance self-oscillations in the tunnel channel.
NASA Technical Reports Server (NTRS)
Maxson, C. W.; Vaiana, G. S.
1977-01-01
In connection with high-quality solar soft X-ray images the 'quiet' features of the inner corona have been separated into two sharply different components, including the strongly reduced emission areas or coronal holes (CH) and the extended regions of looplike emission features or large-scale structures (LSS). Particular central meridian passage observations of the prominent CH1 on August 21, 1973, are selected for a quantitative study. Histogram photographic density distributions for full-disk images at other central meridian passages of CH 1 are also presented, and the techniques of converting low photographic density data to deposited energy are discussed, with particular emphasis on the problems associated with the CH data.
GenomeDiagram: a python package for the visualization of large-scale genomic data.
Pritchard, Leighton; White, Jennifer A; Birch, Paul R J; Toth, Ian K
2006-03-01
We present GenomeDiagram, a flexible, open-source Python module for the visualization of large-scale genomic, comparative genomic and other data with reference to a single chromosome or other biological sequence. GenomeDiagram may be used to generate publication-quality vector graphics, rastered images and in-line streamed graphics for webpages. The package integrates with datatypes from the BioPython project, and is available for Windows, Linux and Mac OS X systems. GenomeDiagram is freely available as source code (under GNU Public License) at http://bioinf.scri.ac.uk/lp/programs.html, and requires Python 2.3 or higher, and recent versions of the ReportLab and BioPython packages. A user manual, example code and images are available at http://bioinf.scri.ac.uk/lp/programs.html.
Nursing Services Delivery Theory: an open system approach.
Meyer, Raquel M; O'Brien-Pallas, Linda L
2010-12-01
This paper is a discussion of the derivation of the Nursing Services Delivery Theory from the application of open system theory to large-scale organizations. The underlying mechanisms by which staffing indicators influence outcomes remain under-theorized and unmeasured, resulting in a 'black box' that masks the nature and organization of nursing work. Theory linking nursing work, staffing, work environments, and outcomes in different settings is urgently needed to inform management decisions about the allocation of nurse staffing resources in organizations. A search of CINAHL and Business Source Premier for the years 1980-2008 was conducted using the following terms: theory, models, organization, organizational structure, management, administration, nursing units, and nursing. Seminal works were included. The healthcare organization is conceptualized as an open system characterized by energy transformation, a dynamic steady state, negative entropy, event cycles, negative feedback, differentiation, integration and coordination, and equifinality. The Nursing Services Delivery Theory proposes that input, throughput, and output factors interact dynamically to influence the global work demands placed on nursing work groups at the point of care in production subsystems. THE Nursing Services Delivery Theory can be applied to varied settings, cultures, and countries and supports the study of multi-level phenomena and cross-level effects. The Nursing Services Delivery Theory gives a relational structure for reconciling disparate streams of research related to nursing work, staffing, and work environments. The theory can guide future research and the management of nursing services in large-scale healthcare organizations. © 2010 Blackwell Publishing Ltd.
A survey of acoustic conditions in semi-open plan classrooms in the United Kingdom.
Greenland, Emma E; Shield, Bridget M
2011-09-01
This paper reports the results of a large scale, detailed acoustic survey of 42 open plan classrooms of varying design in the UK each of which contained between 2 and 14 teaching areas or classbases. The objective survey procedure, which was designed specifically for use in open plan classrooms, is described. The acoustic measurements relating to speech intelligibility within a classbase, including ambient noise level, intrusive noise level, speech to noise ratio, speech transmission index, and reverberation time, are presented. The effects on speech intelligibility of critical physical design variables, such as the number of classbases within an open plan unit and the selection of acoustic finishes for control of reverberation, are examined. This analysis enables limitations of open plan classrooms to be discussed and acoustic design guidelines to be developed to ensure good listening conditions. The types of teaching activity to provide adequate acoustic conditions, plus the speech intelligibility requirements of younger children, are also discussed. © 2011 Acoustical Society of America
Large-scale production of lentiviral vector in a closed system hollow fiber bioreactor
Sheu, Jonathan; Beltzer, Jim; Fury, Brian; Wilczek, Katarzyna; Tobin, Steve; Falconer, Danny; Nolta, Jan; Bauer, Gerhard
2015-01-01
Lentiviral vectors are widely used in the field of gene therapy as an effective method for permanent gene delivery. While current methods of producing small scale vector batches for research purposes depend largely on culture flasks, the emergence and popularity of lentiviral vectors in translational, preclinical and clinical research has demanded their production on a much larger scale, a task that can be difficult to manage with the numbers of producer cell culture flasks required for large volumes of vector. To generate a large scale, partially closed system method for the manufacturing of clinical grade lentiviral vector suitable for the generation of induced pluripotent stem cells (iPSCs), we developed a method employing a hollow fiber bioreactor traditionally used for cell expansion. We have demonstrated the growth, transfection, and vector-producing capability of 293T producer cells in this system. Vector particle RNA titers after subsequent vector concentration yielded values comparable to lentiviral iPSC induction vector batches produced using traditional culture methods in 225 cm2 flasks (T225s) and in 10-layer cell factories (CF10s), while yielding a volume nearly 145 times larger than the yield from a T225 flask and nearly three times larger than the yield from a CF10. Employing a closed system hollow fiber bioreactor for vector production offers the possibility of manufacturing large quantities of gene therapy vector while minimizing reagent usage, equipment footprint, and open system manipulation. PMID:26151065
NASA Astrophysics Data System (ADS)
Klingbeil, Knut; Lemarié, Florian; Debreu, Laurent; Burchard, Hans
2018-05-01
The state of the art of the numerics of hydrostatic structured-grid coastal ocean models is reviewed here. First, some fundamental differences in the hydrodynamics of the coastal ocean, such as the large surface elevation variation compared to the mean water depth, are contrasted against large scale ocean dynamics. Then the hydrodynamic equations as they are used in coastal ocean models as well as in large scale ocean models are presented, including parameterisations for turbulent transports. As steps towards discretisation, coordinate transformations and spatial discretisations based on a finite-volume approach are discussed with focus on the specific requirements for coastal ocean models. As in large scale ocean models, splitting of internal and external modes is essential also for coastal ocean models, but specific care is needed when drying & flooding of intertidal flats is included. As one obvious characteristic of coastal ocean models, open boundaries occur and need to be treated in a way that correct model forcing from outside is transmitted to the model domain without reflecting waves from the inside. Here, also new developments in two-way nesting are presented. Single processes such as internal inertia-gravity waves, advection and turbulence closure models are discussed with focus on the coastal scales. Some overview on existing hydrostatic structured-grid coastal ocean models is given, including their extensions towards non-hydrostatic models. Finally, an outlook on future perspectives is made.
Blanco Gonzalez, Enrique; Aritaki, Masato; Knutsen, Halvor; Taniguchi, Nobuhiko
2015-01-01
Large-scale hatchery releases are carried out for many marine fish species worldwide; nevertheless, the long-term effects of this practice on the genetic structure of natural populations remains unclear. The lack of knowledge is especially evident when independent stock enhancement programs are conducted simultaneously on the same species at different geographical locations, as occurs with red sea bream (Pagrus major, Temminck et Schlegel) in Japan. In this study, we examined the putative effects of intensive offspring releases on the genetic structure of red sea bream populations along the Japanese archipelago by genotyping 848 fish at fifteen microsatellite loci. Our results suggests weak but consistent patterns of genetic divergence (F(ST) = 0.002, p < 0.001). Red sea bream in Japan appeared spatially structured with several patches of distinct allelic composition, which corresponded to areas receiving an important influx of fish of hatchery origin, either released intentionally or from unintentional escapees from aquaculture operations. In addition to impacts upon local populations inhabiting semi-enclosed embayments, large-scale releases (either intentionally or from unintentional escapes) appeared also to have perturbed genetic structure in open areas. Hence, results of the present study suggest that independent large-scale marine stock enhancement programs conducted simultaneously on one species at different geographical locations may compromise native genetic structure and lead to patchy patterns in population genetic structure.
Wu, Hailong; Huo, Yuanzi; Zhang, Jianheng; Liu, Yuanyuan; Zhao, Yating; He, Peimin
2015-06-15
The bioremediation efficiency of China's largest scale Porphyra yezoensis cultivation for removing dissolved nutrients and controlling harmful algae was studied in the radial sandbanks waters of Jiangsu Province in the year 2012-2013. Mean nutrient concentration values in the P. yezoensis cultivation area were significantly lower than those in the non-cultivation area, especially during the cultivation season (p<0.05). Tissue nitrogen and phosphorus contents of seaweeds were 5.99-0.80% (dry weight (DW)) and 0.16-0.19% (DW), respectively. Production of P. yezoensis was 58950.87tons DW. Based on these values, 3688.15tons of tissue nitrogen and 105.61tons of tissue phosphorus were removed by harvesting P. yezoensis. The richness index of the red tide species Skeleton emacostatum declined from 0.32 to 0.05 during the P. yezoensis cultivation season. These results indicate that large-scale cultivation of P. yezoensis can be used to efficiently alleviate eutrophication and control harmful algae blooms in open sea. Copyright © 2015 Elsevier Ltd. All rights reserved.
Visualization of SV2A conformations in situ by the use of Protein Tomography
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lynch, Berkley A.; Matagne, Alain; Braennstroem, Annika
2008-10-31
The synaptic vesicle protein 2A (SV2A), the brain-binding site of the anti-epileptic drug levetiracetam (LEV), has been characterized by Protein Tomography{sup TM}. We identified two major conformations of SV2A in mouse brain tissue: first, a compact, funnel-structure with a pore-like opening towards the cytoplasm; second, a more open, V-shaped structure with a cleft-like opening towards the intravesicular space. The large differences between these conformations suggest a high degree of flexibility and support a valve-like mechanism consistent with the postulated transporter role of SV2A. These two conformations are represented both in samples treated with LEV, and in saline-treated samples, which indicatesmore » that LEV binding does not cause a large-scale conformational change of SV2A, or lock a specific conformational state of the protein. This study provides the first direct structural data on SV2A, and supports a transporter function suggested by sequence homology to MFS class of transporter proteins.« less
Fluctuations in the DNA double helix
NASA Astrophysics Data System (ADS)
Peyrard, M.; López, S. C.; Angelov, D.
2007-08-01
DNA is not the static entity suggested by the famous double helix structure. It shows large fluctuational openings, in which the bases, which contain the genetic code, are temporarily open. Therefore it is an interesting system to study the effect of nonlinearity on the physical properties of a system. A simple model for DNA, at a mesoscopic scale, can be investigated by computer simulation, in the same spirit as the original work of Fermi, Pasta and Ulam. These calculations raise fundamental questions in statistical physics because they show a temporary breaking of equipartition of energy, regions with large amplitude fluctuations being able to coexist with regions where the fluctuations are very small, even when the model is studied in the canonical ensemble. This phenomenon can be related to nonlinear excitations in the model. The ability of the model to describe the actual properties of DNA is discussed by comparing theoretical and experimental results for the probability that base pairs open an a given temperature in specific DNA sequences. These studies give us indications on the proper description of the effect of the sequence in the mesoscopic model.
NASA Astrophysics Data System (ADS)
Sandvig Mariegaard, Jesper; Huiban, Méven Robin; Tornfeldt Sørensen, Jacob; Andersson, Henrik
2017-04-01
Determining the optimal domain size and associated position of open boundaries in local high-resolution downscaling ocean models is often difficult. As an important input data set for downscaling ocean modelling, the European Copernicus Marine Environment Monitoring Service (CMEMS) provides baroclinic initial and boundary conditions for local ocean models. Tidal dynamics is often neglected in CMEMS services at large scale but tides are generally crucial for coastal ocean dynamics. To address this need, tides can be superposed via Flather (1976) boundary conditions and the combined flow downscaled using unstructured mesh. The surge component is also only partially represented in selected CMEMS products and must be modelled inside the domain and modelled independently and superposed if the domain becomes too small to model the effect in the downscaling model. The tide and surge components can generally be improved by assimilating water level from tide gauge and altimetry data. An intrinsic part of the problem is to find the limitations of local scale data assimilation and the requirement for consistency between the larger scale ocean models and the local scale assimilation methodologies. This contribution investigates the impact of domain size and associated positions of open boundaries with and without data assimilation of water level. We have used the baroclinic ocean model, MIKE 3 FM, and its newly re-factored built-in data assimilation package. We consider boundary conditions of salinity, temperature, water level and depth varying currents from the Global CMEMS 1/4 degree resolution model from 2011, where in situ ADCP velocity data is available for validation. We apply data assimilation of in-situ tide gauge water levels and along track altimetry surface elevation data from selected satellites. The MIKE 3 FM data assimilation model which use the Ensemble Kalman filter have recently been parallelized with MPI allowing for much larger applications running on HPC. The success of the downscaling is to a large degree determined by the ability to realistically describe and dynamically model the errors on the open boundaries. Three different sizes of downscaling model domains in the Northern North Sea have been examined and two different strategies for modelling the uncertainties on the open Flather boundaries are investigated. The combined downscaling and local data assimilation skill is assessed and the impact on recommended domain size is compared to pure downscaling.
SUSY’s Ladder: Reframing sequestering at Large Volume
Reece, Matthew; Xue, Wei
2016-04-07
Theories with approximate no-scale structure, such as the Large Volume Scenario, have a distinctive hierarchy of multiple mass scales in between TeV gaugino masses and the Planck scale, which we call SUSY's Ladder. This is a particular realization of Split Supersymmetry in which the same small parameter suppresses gaugino masses relative to scalar soft masses, scalar soft masses relative to the gravitino mass, and the UV cutoff or string scale relative to the Planck scale. This scenario has many phenomenologically interesting properties, and can avoid dangers including the gravitino problem, flavor problems, and the moduli-induced LSP problem that plague othermore » supersymmetric theories. We study SUSY's Ladder using a superspace formalism that makes the mysterious cancelations in previous computations manifest. This opens the possibility of a consistent effective field theory understanding of the phenomenology of these scenarios, based on power-counting in the small ratio of string to Planck scales. We also show that four-dimensional theories with approximate no-scale structure enforced by a single volume modulus arise only from two special higher-dimensional theories: five-dimensional supergravity and ten-dimensional type IIB supergravity. As a result, this gives a phenomenological argument in favor of ten dimensional ultraviolet physics which is different from standard arguments based on the consistency of superstring theory.« less
MultispeQ Beta: a tool for large-scale plant phenotyping connected to the open PhotosynQ network
Austic, Greg; Zegarac, Robert; Osei-Bonsu, Isaac; Hoh, Donghee; Chilvers, Martin I.; Roth, Mitchell G.; Bi, Kevin; TerAvest, Dan; Weebadde, Prabode; Kramer, David M.
2016-01-01
Large-scale high-throughput plant phenotyping (sometimes called phenomics) is becoming increasingly important in plant biology and agriculture and is essential to cutting-edge plant breeding and management approaches needed to meet the food and fuel needs for the next century. Currently, the application of these approaches is severely limited by the availability of appropriate instrumentation and by the ability to communicate experimental protocols, results and analyses. To address these issues, we have developed a low-cost, yet sophisticated open-source scientific instrument designed to enable communities of researchers, plant breeders, educators, farmers and citizen scientists to collect high-quality field data on a large scale. The MultispeQ provides measurements in the field or laboratory of both, environmental conditions (light intensity and quality, temperature, humidity, CO2 levels, time and location) and useful plant phenotypes, including photosynthetic parameters—photosystem II quantum yield (ΦII), non-photochemical exciton quenching (NPQ), photosystem II photoinhibition, light-driven proton translocation and thylakoid proton motive force, regulation of the chloroplast ATP synthase and potentially many others—and leaf chlorophyll and other pigments. Plant phenotype data are transmitted from the MultispeQ to mobile devices, laptops or desktop computers together with key metadata that gets saved to the PhotosynQ platform (https://photosynq.org) and provides a suite of web-based tools for sharing, visualization, filtering, dissemination and analyses. We present validation experiments, comparing MultispeQ results with established platforms, and show that it can be usefully deployed in both laboratory and field settings. We present evidence that MultispeQ can be used by communities of researchers to rapidly measure, store and analyse multiple environmental and plant properties, allowing for deeper understanding of the complex interactions between plants and their environment. PMID:27853580
MultispeQ Beta: a tool for large-scale plant phenotyping connected to the open PhotosynQ network
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kuhlgert, Sebastian; Austic, Greg; Zegarac, Robert
Large-scale high-throughput plant phenotyping (sometimes called phenomics) is becoming increasingly important in plant biology and agriculture and is essential to cutting-edge plant breeding and management approaches needed to meet the food and fuel needs for the next century. Currently, the application of these approaches is severely limited by the availability of appropriate instrumentation and by the ability to communicate experimental protocols, results and analyses. To address these issues, we have developed a low-cost, yet sophisticated open-source scientific instrument designed to enable communities of researchers, plant breeders, educators, farmers and citizen scientists to collect high-quality field data on a large scale.more » The MultispeQ provides measurements in the field or laboratory of both, environmental conditions (light intensity and quality, temperature, humidity, CO 2 levels, time and location) and useful plant phenotypes, including photosynthetic parameters—photosystem II quantum yield (Φ II), non-photochemical exciton quenching (NPQ), photosystem II photoinhibition, light-driven proton translocation and thylakoid proton motive force, regulation of the chloroplast ATP synthase and potentially many others—and leaf chlorophyll and other pigments. Plant phenotype data are transmitted from the MultispeQ to mobile devices, laptops or desktop computers together with key metadata that gets saved to the PhotosynQ platform (https://photosynq.org) and provides a suite of web-based tools for sharing, visualization, filtering, dissemination and analyses. We present validation experiments, comparing MultispeQ results with established platforms, and show that it can be usefully deployed in both laboratory and field settings. We present evidence that MultispeQ can be used by communities of researchers to rapidly measure, store and analyse multiple environmental and plant properties, allowing for deeper understanding of the complex interactions between plants and their environment.« less
MultispeQ Beta: a tool for large-scale plant phenotyping connected to the open PhotosynQ network
Kuhlgert, Sebastian; Austic, Greg; Zegarac, Robert; ...
2016-10-26
Large-scale high-throughput plant phenotyping (sometimes called phenomics) is becoming increasingly important in plant biology and agriculture and is essential to cutting-edge plant breeding and management approaches needed to meet the food and fuel needs for the next century. Currently, the application of these approaches is severely limited by the availability of appropriate instrumentation and by the ability to communicate experimental protocols, results and analyses. To address these issues, we have developed a low-cost, yet sophisticated open-source scientific instrument designed to enable communities of researchers, plant breeders, educators, farmers and citizen scientists to collect high-quality field data on a large scale.more » The MultispeQ provides measurements in the field or laboratory of both, environmental conditions (light intensity and quality, temperature, humidity, CO 2 levels, time and location) and useful plant phenotypes, including photosynthetic parameters—photosystem II quantum yield (Φ II), non-photochemical exciton quenching (NPQ), photosystem II photoinhibition, light-driven proton translocation and thylakoid proton motive force, regulation of the chloroplast ATP synthase and potentially many others—and leaf chlorophyll and other pigments. Plant phenotype data are transmitted from the MultispeQ to mobile devices, laptops or desktop computers together with key metadata that gets saved to the PhotosynQ platform (https://photosynq.org) and provides a suite of web-based tools for sharing, visualization, filtering, dissemination and analyses. We present validation experiments, comparing MultispeQ results with established platforms, and show that it can be usefully deployed in both laboratory and field settings. We present evidence that MultispeQ can be used by communities of researchers to rapidly measure, store and analyse multiple environmental and plant properties, allowing for deeper understanding of the complex interactions between plants and their environment.« less
Large-Scale Event Extraction from Literature with Multi-Level Gene Normalization
Wei, Chih-Hsuan; Hakala, Kai; Pyysalo, Sampo; Ananiadou, Sophia; Kao, Hung-Yu; Lu, Zhiyong; Salakoski, Tapio; Van de Peer, Yves; Ginter, Filip
2013-01-01
Text mining for the life sciences aims to aid database curation, knowledge summarization and information retrieval through the automated processing of biomedical texts. To provide comprehensive coverage and enable full integration with existing biomolecular database records, it is crucial that text mining tools scale up to millions of articles and that their analyses can be unambiguously linked to information recorded in resources such as UniProt, KEGG, BioGRID and NCBI databases. In this study, we investigate how fully automated text mining of complex biomolecular events can be augmented with a normalization strategy that identifies biological concepts in text, mapping them to identifiers at varying levels of granularity, ranging from canonicalized symbols to unique gene and proteins and broad gene families. To this end, we have combined two state-of-the-art text mining components, previously evaluated on two community-wide challenges, and have extended and improved upon these methods by exploiting their complementary nature. Using these systems, we perform normalization and event extraction to create a large-scale resource that is publicly available, unique in semantic scope, and covers all 21.9 million PubMed abstracts and 460 thousand PubMed Central open access full-text articles. This dataset contains 40 million biomolecular events involving 76 million gene/protein mentions, linked to 122 thousand distinct genes from 5032 species across the full taxonomic tree. Detailed evaluations and analyses reveal promising results for application of this data in database and pathway curation efforts. The main software components used in this study are released under an open-source license. Further, the resulting dataset is freely accessible through a novel API, providing programmatic and customized access (http://www.evexdb.org/api/v001/). Finally, to allow for large-scale bioinformatic analyses, the entire resource is available for bulk download from http://evexdb.org/download/, under the Creative Commons – Attribution – Share Alike (CC BY-SA) license. PMID:23613707
SNMG: a social-level norm-based methodology for macro-governing service collaboration processes
NASA Astrophysics Data System (ADS)
Gao, Ji; Lv, Hexin; Jin, Zhiyong; Xu, Ping
2017-08-01
In order to adapt to the accelerative open tendency of collaborations between enterprises, this paper proposes a Social-level Norm-based methodology for Macro-Governing service collaboration processes, called SNMG, to regulate and control the social-level visible macro-behaviors of the social individuals participating in collaborations. SNMG not only can remove effectively the uncontrollability hindrance confronted with by open social activities, but also enables across-management-domain collaborations to be implemented by uniting the centralized controls of social individuals for respective social activities. Therefore, this paper provides a brand-new system construction mode to promote the development and large-scale deployment of service collaborations.
Effect of buoyancy on fuel containment in an open-cycle gas-core nuclear rocket engine.
NASA Technical Reports Server (NTRS)
Putre, H. A.
1971-01-01
Analysis aimed at determining the scaling laws for the buoyancy effect on fuel containment in an open-cycle gas-core nuclear rocket engine, so conducted that experimental conditions can be related to engine conditions. The fuel volume fraction in a short coaxial flow cavity is calculated with a programmed numerical solution of the steady Navier-Stokes equations for isothermal, variable density fluid mixing. A dimensionless parameter B, called the Buoyancy number, was found to correlate the fuel volume fraction for large accelerations and various density ratios. This parameter has the value B = 0 for zero acceleration, and B = 350 for typical engine conditions.
Parallel Adjective High-Order CFD Simulations Characterizing SOFIA Cavity Acoustics
NASA Technical Reports Server (NTRS)
Barad, Michael F.; Brehm, Christoph; Kiris, Cetin C.; Biswas, Rupak
2016-01-01
This paper presents large-scale MPI-parallel computational uid dynamics simulations for the Stratospheric Observatory for Infrared Astronomy (SOFIA). SOFIA is an airborne, 2.5-meter infrared telescope mounted in an open cavity in the aft fuselage of a Boeing 747SP. These simulations focus on how the unsteady ow eld inside and over the cavity interferes with the optical path and mounting structure of the telescope. A temporally fourth-order accurate Runge-Kutta, and spatially fth-order accurate WENO- 5Z scheme was used to perform implicit large eddy simulations. An immersed boundary method provides automated gridding for complex geometries and natural coupling to a block-structured Cartesian adaptive mesh re nement framework. Strong scaling studies using NASA's Pleiades supercomputer with up to 32k CPU cores and 4 billion compu- tational cells shows excellent scaling. Dynamic load balancing based on execution time on individual AMR blocks addresses irregular numerical cost associated with blocks con- taining boundaries. Limits to scaling beyond 32k cores are identi ed, and targeted code optimizations are discussed.
Parallel Adaptive High-Order CFD Simulations Characterizing SOFIA Cavitiy Acoustics
NASA Technical Reports Server (NTRS)
Barad, Michael F.; Brehm, Christoph; Kiris, Cetin C.; Biswas, Rupak
2015-01-01
This paper presents large-scale MPI-parallel computational uid dynamics simulations for the Stratospheric Observatory for Infrared Astronomy (SOFIA). SOFIA is an airborne, 2.5-meter infrared telescope mounted in an open cavity in the aft fuselage of a Boeing 747SP. These simulations focus on how the unsteady ow eld inside and over the cavity interferes with the optical path and mounting structure of the telescope. A tempo- rally fourth-order accurate Runge-Kutta, and a spatially fth-order accurate WENO-5Z scheme were used to perform implicit large eddy simulations. An immersed boundary method provides automated gridding for complex geometries and natural coupling to a block-structured Cartesian adaptive mesh re nement framework. Strong scaling studies using NASA's Pleiades supercomputer with up to 32k CPU cores and 4 billion compu- tational cells shows excellent scaling. Dynamic load balancing based on execution time on individual AMR blocks addresses irregular numerical cost associated with blocks con- taining boundaries. Limits to scaling beyond 32k cores are identi ed, and targeted code optimizations are discussed.
Adjusting to Social Change - A Multi-Level Analysis in Three Cultures
2013-08-01
including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed , and completing and reviewing the...presence is often associated with the large-scale movement of civilian populations, and who need to better understand the Distribution A: Approved for...valuing openness to change (self-direction, stimulation and sometimes hedonism values) with valuing conservation (conformity, tradition and security
Conservation easements in the Adirondack Park of New York state
Chad P. Dawson; Steven Bick; Peter D' Luhosch; Matthew Nowak; Diane Kuehn
2015-01-01
The use of conservation easements to keep private lands undeveloped and protect open space and large scale landscapes has grown rapidly. The New York State Adirondack Park includes 2.5 million acres (1 million ha) of state owned land and 3 million acres (1.2 million ha) of private lands; over 781,000 acres (316,194 ha) of these private lands were under publicly held...
ERIC Educational Resources Information Center
Gogia, Laura Park
2016-01-01
Virginia Commonwealth University (VCU) is implementing a large scale exploration of digital pedagogies, including connected learning and open education, in an effort to promote digital fluency and integrative thinking among students. The purpose of this study was to develop a classroom assessment toolkit for faculty who wish to document student…
Statistical Literacy in Data Revolution Era: Building Blocks and Instructional Dilemmas
ERIC Educational Resources Information Center
Prodromou, Theodosia; Dunne, Tim
2017-01-01
The data revolution has given citizens access to enormous large-scale open databases. In order to take into account the full complexity of data, we have to change the way we think in terms of the nature of data and its availability, the ways in which it is displayed and used, and the skills that are required for its interpretation. Substantial…
Mathematical and Computational Challenges in Population Biology and Ecosystems Science
NASA Technical Reports Server (NTRS)
Levin, Simon A.; Grenfell, Bryan; Hastings, Alan; Perelson, Alan S.
1997-01-01
Mathematical and computational approaches provide powerful tools in the study of problems in population biology and ecosystems science. The subject has a rich history intertwined with the development of statistics and dynamical systems theory, but recent analytical advances, coupled with the enhanced potential of high-speed computation, have opened up new vistas and presented new challenges. Key challenges involve ways to deal with the collective dynamics of heterogeneous ensembles of individuals, and to scale from small spatial regions to large ones. The central issues-understanding how detail at one scale makes its signature felt at other scales, and how to relate phenomena across scales-cut across scientific disciplines and go to the heart of algorithmic development of approaches to high-speed computation. Examples are given from ecology, genetics, epidemiology, and immunology.
Ibrahim, Khaled Z.; Madduri, Kamesh; Williams, Samuel; ...
2013-07-18
The Gyrokinetic Toroidal Code (GTC) uses the particle-in-cell method to efficiently simulate plasma microturbulence. This paper presents novel analysis and optimization techniques to enhance the performance of GTC on large-scale machines. We introduce cell access analysis to better manage locality vs. synchronization tradeoffs on CPU and GPU-based architectures. Finally, our optimized hybrid parallel implementation of GTC uses MPI, OpenMP, and NVIDIA CUDA, achieves up to a 2× speedup over the reference Fortran version on multiple parallel systems, and scales efficiently to tens of thousands of cores.
Visualization and analysis of flow structures in an open cavity
NASA Astrophysics Data System (ADS)
Liu, Jun; Cai, Jinsheng; Yang, Dangguo; Wu, Junqiang; Wang, Xiansheng
2018-05-01
A numerical study is performed on the supersonic flow over an open cavity at Mach number of 1.5. A newly developed visualization method is employed to visualize the complicated flow structures, which provide an insight into major flow physics. Four types of shock/compressive waves which existed in experimental schlieren are observed in numerical visualization results. Furthermore, other flow structures such as multi-scale vortices are also obtained in the numerical results. And a new type of shocklet which is beneath large vortices is found. The shocklet beneath the vortex originates from leading edge, then, is strengthened by successive interactions between feedback compressive waves and its attached vortex. Finally, it collides against the trailing surface and generates a large number of feedback compressive waves and intensive pressure fluctuations. It is suggested that the shocklets beneath vortex play an important role of cavity self-sustained oscillation.
Harispe, Sébastien; Ranwez, Sylvie; Janaqi, Stefan; Montmain, Jacky
2014-03-01
The semantic measures library and toolkit are robust open-source and easy to use software solutions dedicated to semantic measures. They can be used for large-scale computations and analyses of semantic similarities between terms/concepts defined in terminologies and ontologies. The comparison of entities (e.g. genes) annotated by concepts is also supported. A large collection of measures is available. Not limited to a specific application context, the library and the toolkit can be used with various controlled vocabularies and ontology specifications (e.g. Open Biomedical Ontology, Resource Description Framework). The project targets both designers and practitioners of semantic measures providing a JAVA library, as well as a command-line tool that can be used on personal computers or computer clusters. Downloads, documentation, tutorials, evaluation and support are available at http://www.semantic-measures-library.org.
Structure and organization of Stratocumulus fields: A network approach
NASA Astrophysics Data System (ADS)
Glassmeier, Franziska; Feingold, Graham
2017-04-01
The representation of Stratocumulus (Sc) clouds and their radiative impact is one of the large challenges for global climate models. Aerosol-cloud-precipitation interactions greatly contribute to this challenge by influencing the morphology of Sc fields: In the absence of rain, Sc are arranged in a relatively regular pattern of cloudy cells separated by cloud-free rings of down welling air ('closed cells'). Raining cloud fields, in contrast, exhibit an oscillating pattern of cloudy rings surrounding cloud free cells of negatively buoyant air caused by sedimentation and evaporation of rain ('open cells'). Surprisingly, these regular structures of open and closed cellular Sc fields and their potential for the development of new parameterizations have hardly been explored. In this contribution, we approach the organization of Sc from the perspective of a 2-dimensional random network. We find that cellular networks derived from LES simulations of open- and closed-cell Sc cases are almost indistinguishable and share the following features: (i) The distributions of nearest neighbors, or cell degree, are centered at six. This corresponds to approximately hexagonal cloud cells and is a direct mathematical consequence (Euler formula) of the triple junctions featured by Sc organization. (ii) The degree of individual cells is found to be proportional to the normalized size of the cells. This means that cell arrangement is independent of the typical cell size. (iii) Reflecting the continuously renewing dynamics of Sc fields, large (high-degree) cells tend to be neighbored by small (low-degree) cells and vice versa. These macroscopic network properties emerge independent of the state of the Sc field because the different processes governing the evolution of closed as compared to open cells correspond to topologically equivalent network dynamics. By developing a heuristic model, we show that open and closed cell dynamics can both be mimicked by versions of cell division and cell disappearance and are biased towards the expansion of smaller cells. As a conclusion of our network analysis, Sc organization can be characterized by a typical length scale and a scale-independent cell arrangement. While the typical length scale emerges from the full complexity of aerosol-cloud-precipitation-radiation interactions, cell arrangement is independent of cloud processes and its evolution could be parameterized based on our heuristic model.
Wan, Shixiang; Zou, Quan
2017-01-01
Multiple sequence alignment (MSA) plays a key role in biological sequence analyses, especially in phylogenetic tree construction. Extreme increase in next-generation sequencing results in shortage of efficient ultra-large biological sequence alignment approaches for coping with different sequence types. Distributed and parallel computing represents a crucial technique for accelerating ultra-large (e.g. files more than 1 GB) sequence analyses. Based on HAlign and Spark distributed computing system, we implement a highly cost-efficient and time-efficient HAlign-II tool to address ultra-large multiple biological sequence alignment and phylogenetic tree construction. The experiments in the DNA and protein large scale data sets, which are more than 1GB files, showed that HAlign II could save time and space. It outperformed the current software tools. HAlign-II can efficiently carry out MSA and construct phylogenetic trees with ultra-large numbers of biological sequences. HAlign-II shows extremely high memory efficiency and scales well with increases in computing resource. THAlign-II provides a user-friendly web server based on our distributed computing infrastructure. HAlign-II with open-source codes and datasets was established at http://lab.malab.cn/soft/halign.
Polignano, Francesco M; Quyn, Aaron J; de Figueiredo, Rodrigo S M; Henderson, Nikola A; Kulli, Christoph; Tait, Iain S
2008-12-01
Reduction in hospital stay, blood loss, postoperative pain and complications are common findings after laparoscopic liver resection, suggesting that the laparoscopic approach may be a suitable alternative to open surgery. Some concerns have been raised regarding cost effectiveness of this procedure and potential implications of its large-scale application. Our aim has been to determine cost effectiveness of laparoscopic liver surgery by a case-matched, case-control, intention-to-treat analysis of its costs and short-term clinical outcomes compared with open surgery. Laparoscopic liver segmentectomies and bisegmentectomies performed at Ninewells Hospital and Medical School between 2005 and 2007 were considered. Resections involving more than two Couinaud segments, or involving any synchronous procedure, were excluded. An operation-magnitude-matched control group was identified amongst open liver resections performed between 2004 and 2007. Hospital costs were obtained from the Scottish Health Service Costs Book (ISD Scotland) and average national costs were calculated. Cost of theatre time, disposable surgical devices, hospital stay, and high-dependency unit (HDU) and intensive care unit (ICU) usage were the main endpoints for comparison. Secondary endpoints were morbidity and mortality. Statistical analysis was performed with Student's t-test, chi(2) and Fisher exact test as most appropriate. Twenty-five laparoscopic liver resections were considered, including atypical resection, segmentectomy and bisegmentectomy, and they were compared to 25 matching open resections. The two groups were homogeneous by age, sex, coexistent morbidity, magnitude of resection, prevalence of liver cirrhosis and indications. Operative time (p < 0.03), blood loss (p < 0.0001), Pringle manoeuvre (p < 0.03), hospital stay (p < 0.003) and postoperative complications (p < 0.002) were significantly reduced in the laparoscopic group. Overall hospital cost was significantly lower in the laparoscopic group by an average of 2,571 pounds sterling (p < 0.04). Laparoscopic liver segmentectomy and bisegmentectomy are feasible, safe and cost effective compared to similar open resections. Large-scale application of laparoscopic liver surgery could translate into significant savings to hospitals and health care programmes.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gershenzon, Naum I.; Soltanian, Mohamad Reza; Ritzi, Robert W.
Understanding multi-phase fluid flow and transport processes within aquifers, candidate reservoirs for CO 2 sequestration, and petroleum reservoirs requires understanding a diverse set of geologic properties of the aquifer or reservoir, over a wide range of spatial and temporal scales. We focus on multiphase flow dynamics with wetting (e.g., water) and non-wetting (e.g., gas or oil) fluids, with one invading another. This problem is of general interest in a number of fields and is illustrated here by considering the sweep efficiency of oil during a waterflood. Using a relatively fine-resolution grid throughout a relatively large domain in these simulations andmore » probing the results with advanced scientific visualization tools (Reservoir Visualization Analysis [RVA]/ ParaView software) promote a better understanding of how smaller-scale features affect the aggregate behavior at larger scales. We studied the effects on oil-sweep efficiency of the proportion, hierarchical organization, and connectivity of high-permeability open-framework conglomerate (OFC) cross-sets within the multi-scale stratal architecture found in fluvial deposits. We further analyzed oil production rate, water breakthrough time, and spatial and temporal distribution of residual oil saturation. As expected, the effective permeability of the reservoir exhibits large-scale anisotropy created by the organization of OFC cross-sets within unit bars, and the organization of unit bars within compound- bars. As a result, oil-sweep efficiency critically depends on the direction of the pressure gradient. However, contrary to expectations, the total amount of trapped oil due to the effect of capillary trapping does not depend on the magnitude of the pressure gradient within the examined range. Hence the pressure difference between production and injection wells does not affect sweep efficiency; although the spatial distribution of oil remaining in the reservoir depends on this value. Whether or not clusters of connected OFC span the domain affects only the absolute rate of oil production—not sweep efficiency.« less
Gershenzon, Naum I.; Soltanian, Mohamad Reza; Ritzi, Robert W.; ...
2015-10-23
Understanding multi-phase fluid flow and transport processes within aquifers, candidate reservoirs for CO 2 sequestration, and petroleum reservoirs requires understanding a diverse set of geologic properties of the aquifer or reservoir, over a wide range of spatial and temporal scales. We focus on multiphase flow dynamics with wetting (e.g., water) and non-wetting (e.g., gas or oil) fluids, with one invading another. This problem is of general interest in a number of fields and is illustrated here by considering the sweep efficiency of oil during a waterflood. Using a relatively fine-resolution grid throughout a relatively large domain in these simulations andmore » probing the results with advanced scientific visualization tools (Reservoir Visualization Analysis [RVA]/ ParaView software) promote a better understanding of how smaller-scale features affect the aggregate behavior at larger scales. We studied the effects on oil-sweep efficiency of the proportion, hierarchical organization, and connectivity of high-permeability open-framework conglomerate (OFC) cross-sets within the multi-scale stratal architecture found in fluvial deposits. We further analyzed oil production rate, water breakthrough time, and spatial and temporal distribution of residual oil saturation. As expected, the effective permeability of the reservoir exhibits large-scale anisotropy created by the organization of OFC cross-sets within unit bars, and the organization of unit bars within compound- bars. As a result, oil-sweep efficiency critically depends on the direction of the pressure gradient. However, contrary to expectations, the total amount of trapped oil due to the effect of capillary trapping does not depend on the magnitude of the pressure gradient within the examined range. Hence the pressure difference between production and injection wells does not affect sweep efficiency; although the spatial distribution of oil remaining in the reservoir depends on this value. Whether or not clusters of connected OFC span the domain affects only the absolute rate of oil production—not sweep efficiency.« less
Spatial scale of similarity as an indicator of metacommunity stability in exploited marine systems.
Shackell, Nancy L; Fisher, Jonathan A D; Frank, Kenneth T; Lawton, Peter
2012-01-01
The spatial scale of similarity among fish communities is characteristically large in temperate marine systems: connectivity is enhanced by high rates of dispersal during the larval/juvenile stages and the increased mobility of large-bodied fish. A larger spatial scale of similarity (low beta diversity) is advantageous in heavily exploited systems because locally depleted populations are more likely to be "rescued" by neighboring areas. We explored whether the spatial scale of similarity changed from 1970 to 2006 due to overfishing of dominant, large-bodied groundfish across a 300 000-km2 region of the Northwest Atlantic. Annually, similarities among communities decayed slowly with increasing geographic distance in this open system, but through time the decorrelation distance declined by 33%, concomitant with widespread reductions in biomass, body size, and community evenness. The decline in connectivity stemmed from an erosion of community similarity among local subregions separated by distances as small as 100 km. Larger fish, of the same species, contribute proportionally more viable offspring, so observed body size reductions will have affected maternal output. The cumulative effect of nonlinear maternal influences on egg/larval quality may have compromised the spatial scale of effective larval dispersal, which may account for the delayed recovery of certain member species. Our study adds strong support for using the spatial scale of similarity as an indicator of metacommunity stability both to understand the spatial impacts of exploitation and to refine how spatial structure is used in management plans.
The cosmological principle is not in the sky
NASA Astrophysics Data System (ADS)
Park, Chan-Gyung; Hyun, Hwasu; Noh, Hyerim; Hwang, Jai-chan
2017-08-01
The homogeneity of matter distribution at large scales, known as the cosmological principle, is a central assumption in the standard cosmological model. The case is testable though, thus no longer needs to be a principle. Here we perform a test for spatial homogeneity using the Sloan Digital Sky Survey Luminous Red Galaxies (LRG) sample by counting galaxies within a specified volume with the radius scale varying up to 300 h-1 Mpc. We directly confront the large-scale structure data with the definition of spatial homogeneity by comparing the averages and dispersions of galaxy number counts with allowed ranges of the random distribution with homogeneity. The LRG sample shows significantly larger dispersions of number counts than the random catalogues up to 300 h-1 Mpc scale, and even the average is located far outside the range allowed in the random distribution; the deviations are statistically impossible to be realized in the random distribution. This implies that the cosmological principle does not hold even at such large scales. The same analysis of mock galaxies derived from the N-body simulation, however, suggests that the LRG sample is consistent with the current paradigm of cosmology, thus the simulation is also not homogeneous in that scale. We conclude that the cosmological principle is neither in the observed sky nor demanded to be there by the standard cosmological world model. This reveals the nature of the cosmological principle adopted in the modern cosmology paradigm, and opens a new field of research in theoretical cosmology.
Hadrava, Jiří; Albrecht, Tomáš; Tryjanowski, Piotr
2018-01-01
Birds sitting or feeding on live large African herbivorous mammals are a visible, yet quite neglected, type of commensalistic–mutualistic association. Here, we investigate general patterns in such relationships at large spatial and taxonomic scales. To obtain large-scale data, an extensive internet-based search for photos was carried out on Google Images. To characterize patterns of the structural organization of commensalistic–mutualistic associations between African birds and herbivorous mammals, we used a network analysis approach. We then employed phylogenetically-informed comparative analysis to explore whether features of bird visitation of mammals, i.e., their mean number, mass and species richness per mammal species, are shaped by a combination of host mammal (body mass and herd size) and environmental (habitat openness) characteristics. We found that the association web structure was only weakly nested for commensalistic as well as for mutualistic birds (oxpeckers Buphagus spp.) and African mammals. Moreover, except for oxpeckers, nestedness did not differ significantly from a null model indicating that birds do not prefer mammal species which are visited by a large number of bird species. In oxpeckers, however, a nested structure suggests a non-random assignment of birds to their mammal hosts. We also identified some new or rare associations between birds and mammals, but we failed to find several previously described associations. Furthermore, we found that mammal body mass positively influenced the number and mass of birds observed sitting on them in the full set of species (i.e., taking oxpeckers together with other bird species). We also found a positive correlation between mammal body mass and mass of non-oxpecker species as well as oxpeckers. Mammal herd size was associated with a higher mass of birds in the full set of species as well as in non-oxpecker species, and mammal species living in larger herds also attracted more bird species in the full set of species. Habitat openness influenced the mass of birds sitting on mammals as well as the number of species recorded sitting on mammals in the full set of species. In non-oxpecker species habitat openness was correlated with the bird number, mass and species richness. Our results provide evidence that patterns of bird–mammal associations can be linked to mammal and environmental characteristics and highlight the potential role of information technologies and new media in further studies of ecology and evolution. However, further study is needed to get a proper insight into the biological and methodological processes underlying the observed patterns. PMID:29576981
Communication: Polymer entanglement dynamics: Role of attractive interactions
Grest, Gary S.
2016-10-10
The coupled dynamics of entangled polymers, which span broad time and length scales, govern their unique viscoelastic properties. To follow chain mobility by numerical simulations from the intermediate Rouse and reptation regimes to the late time diffusive regime, highly coarse grained models with purely repulsive interactions between monomers are widely used since they are computationally the most efficient. In this paper, using large scale molecular dynamics simulations, the effect of including the attractive interaction between monomers on the dynamics of entangled polymer melts is explored for the first time over a wide temperature range. Attractive interactions have little effect onmore » the local packing for all temperatures T and on the chain mobility for T higher than about twice the glass transition T g. Finally, these results, across a broad range of molecular weight, show that to study the dynamics of entangled polymer melts, the interactions can be treated as pure repulsive, confirming a posteriori the validity of previous studies and opening the way to new large scale numerical simulations.« less
Hybrid multiphoton volumetric functional imaging of large-scale bioengineered neuronal networks
NASA Astrophysics Data System (ADS)
Dana, Hod; Marom, Anat; Paluch, Shir; Dvorkin, Roman; Brosh, Inbar; Shoham, Shy
2014-06-01
Planar neural networks and interfaces serve as versatile in vitro models of central nervous system physiology, but adaptations of related methods to three dimensions (3D) have met with limited success. Here, we demonstrate for the first time volumetric functional imaging in a bioengineered neural tissue growing in a transparent hydrogel with cortical cellular and synaptic densities, by introducing complementary new developments in nonlinear microscopy and neural tissue engineering. Our system uses a novel hybrid multiphoton microscope design combining a 3D scanning-line temporal-focusing subsystem and a conventional laser-scanning multiphoton microscope to provide functional and structural volumetric imaging capabilities: dense microscopic 3D sampling at tens of volumes per second of structures with mm-scale dimensions containing a network of over 1,000 developing cells with complex spontaneous activity patterns. These developments open new opportunities for large-scale neuronal interfacing and for applications of 3D engineered networks ranging from basic neuroscience to the screening of neuroactive substances.
Anagnostou, Paolo; Dominici, Valentina; Battaggia, Cinzia; Pagani, Luca; Vilar, Miguel; Wells, R. Spencer; Pettener, Davide; Sarno, Stefania; Boattini, Alessio; Francalacci, Paolo; Colonna, Vincenza; Vona, Giuseppe; Calò, Carla; Destro Bisol, Giovanni; Tofanelli, Sergio
2017-01-01
Human populations are often dichotomized into “isolated” and “open” categories using cultural and/or geographical barriers to gene flow as differential criteria. Although widespread, the use of these alternative categories could obscure further heterogeneity due to inter-population differences in effective size, growth rate, and timing or amount of gene flow. We compared intra and inter-population variation measures combining novel and literature data relative to 87,818 autosomal SNPs in 14 open populations and 10 geographic and/or linguistic European isolates. Patterns of intra-population diversity were found to vary considerably more among isolates, probably due to differential levels of drift and inbreeding. The relatively large effective size estimated for some population isolates challenges the generalized view that they originate from small founding groups. Principal component scores based on measures of intra-population variation of isolated and open populations were found to be distributed along a continuum, with an area of intersection between the two groups. Patterns of inter-population diversity were even closer, as we were able to detect some differences between population groups only for a few multidimensional scaling dimensions. Therefore, different lines of evidence suggest that dichotomizing human populations into open and isolated groups fails to capture the actual relations among their genomic features. PMID:28145502
Modeling the Hydrologic Effects of Large-Scale Green Infrastructure Projects with GIS
NASA Astrophysics Data System (ADS)
Bado, R. A.; Fekete, B. M.; Khanbilvardi, R.
2015-12-01
Impervious surfaces in urban areas generate excess runoff, which in turn causes flooding, combined sewer overflows, and degradation of adjacent surface waters. Municipal environmental protection agencies have shown a growing interest in mitigating these effects with 'green' infrastructure practices that partially restore the perviousness and water holding capacity of urban centers. Assessment of the performance of current and future green infrastructure projects is hindered by the lack of adequate hydrological modeling tools; conventional techniques fail to account for the complex flow pathways of urban environments, and detailed analyses are difficult to prepare for the very large domains in which green infrastructure projects are implemented. Currently, no standard toolset exists that can rapidly and conveniently predict runoff, consequent inundations, and sewer overflows at a city-wide scale. We demonstrate how streamlined modeling techniques can be used with open-source GIS software to efficiently model runoff in large urban catchments. Hydraulic parameters and flow paths through city blocks, roadways, and sewer drains are automatically generated from GIS layers, and ultimately urban flow simulations can be executed for a variety of rainfall conditions. With this methodology, users can understand the implications of large-scale land use changes and green/gray storm water retention systems on hydraulic loading, peak flow rates, and runoff volumes.
Self-assembly of three-dimensional open structures using patchy colloidal particles.
Rocklin, D Zeb; Mao, Xiaoming
2014-10-14
Open structures can display a number of unusual properties, including a negative Poisson's ratio, negative thermal expansion, and holographic elasticity, and have many interesting applications in engineering. However, it is a grand challenge to self-assemble open structures at the colloidal scale, where short-range interactions and low coordination number can leave them mechanically unstable. In this paper we discuss the self-assembly of three-dimensional open structures using triblock Janus particles, which have two large attractive patches that can form multiple bonds, separated by a band with purely hard-sphere repulsion. Such surface patterning leads to open structures that are stabilized by orientational entropy (in an order-by-disorder effect) and selected over close-packed structures by vibrational entropy. For different patch sizes the particles can form into either tetrahedral or octahedral structural motifs which then compose open lattices, including the pyrochlore, the hexagonal tetrastack and the perovskite lattices. Using an analytic theory, we examine the phase diagrams of these possible open and close-packed structures for triblock Janus particles and characterize the mechanical properties of these structures. Our theory leads to rational designs of particles for the self-assembly of three-dimensional colloidal structures that are possible using current experimental techniques.
Wiedinmyer, Christine; Yokelson, Robert J; Gullett, Brian K
2014-08-19
The open burning of waste, whether at individual residences, businesses, or dump sites, is a large source of air pollutants. These emissions, however, are not included in many current emission inventories used for chemistry and climate modeling applications. This paper presents the first comprehensive and consistent estimates of the global emissions of greenhouse gases, particulate matter, reactive trace gases, and toxic compounds from open waste burning. Global emissions of CO2 from open waste burning are relatively small compared to total anthropogenic CO2; however, regional CO2 emissions, particularly in many developing countries in Asia and Africa, are substantial. Further, emissions of reactive trace gases and particulate matter from open waste burning are more significant on regional scales. For example, the emissions of PM10 from open domestic waste burning in China is equivalent to 22% of China's total reported anthropogenic PM10 emissions. The results of the emissions model presented here suggest that emissions of many air pollutants are significantly underestimated in current inventories because open waste burning is not included, consistent with studies that compare model results with available observations.
The Privacy and Security Implications of Open Data in Healthcare.
Kobayashi, Shinji; Kane, Thomas B; Paton, Chris
2018-04-22
The International Medical Informatics Association (IMIA) Open Source Working Group (OSWG) initiated a group discussion to discuss current privacy and security issues in the open data movement in the healthcare domain from the perspective of the OSWG membership. Working group members independently reviewed the recent academic and grey literature and sampled a number of current large-scale open data projects to inform the working group discussion. This paper presents an overview of open data repositories and a series of short case reports to highlight relevant issues present in the recent literature concerning the adoption of open approaches to sharing healthcare datasets. Important themes that emerged included data standardisation, the inter-connected nature of the open source and open data movements, and how publishing open data can impact on the ethics, security, and privacy of informatics projects. The open data and open source movements in healthcare share many common philosophies and approaches including developing international collaborations across multiple organisations and domains of expertise. Both movements aim to reduce the costs of advancing scientific research and improving healthcare provision for people around the world by adopting open intellectual property licence agreements and codes of practice. Implications of the increased adoption of open data in healthcare include the need to balance the security and privacy challenges of opening data sources with the potential benefits of open data for improving research and healthcare delivery. Georg Thieme Verlag KG Stuttgart.
GATECloud.net: a platform for large-scale, open-source text processing on the cloud.
Tablan, Valentin; Roberts, Ian; Cunningham, Hamish; Bontcheva, Kalina
2013-01-28
Cloud computing is increasingly being regarded as a key enabler of the 'democratization of science', because on-demand, highly scalable cloud computing facilities enable researchers anywhere to carry out data-intensive experiments. In the context of natural language processing (NLP), algorithms tend to be complex, which makes their parallelization and deployment on cloud platforms a non-trivial task. This study presents a new, unique, cloud-based platform for large-scale NLP research--GATECloud. net. It enables researchers to carry out data-intensive NLP experiments by harnessing the vast, on-demand compute power of the Amazon cloud. Important infrastructural issues are dealt with by the platform, completely transparently for the researcher: load balancing, efficient data upload and storage, deployment on the virtual machines, security and fault tolerance. We also include a cost-benefit analysis and usage evaluation.
Ecologic and Geographic Distribution of Filovirus Disease
Bauer, John T.; Mills, James N.
2004-01-01
We used ecologic niche modeling of outbreaks and sporadic cases of filovirus-associated hemorrhagic fever (HF) to provide a large-scale perspective on the geographic and ecologic distributions of Ebola and Marburg viruses. We predicted that filovirus would occur across the Afrotropics: Ebola HF in the humid rain forests of central and western Africa, and Marburg HF in the drier and more open areas of central and eastern Africa. Most of the predicted geographic extent of Ebola HF has been observed; Marburg HF has the potential to occur farther south and east. Ecologic conditions appropriate for Ebola HF are also present in Southeast Asia and the Philippines, where Ebola Reston is hypothesized to be distributed. This first large-scale ecologic analysis provides a framework for a more informed search for taxa that could constitute the natural reservoir for this virus family. PMID:15078595
NASA Technical Reports Server (NTRS)
Kashlinsky, A.
1992-01-01
This study presents a method for obtaining the true rms peculiar flow in the universe on scales up to 100-120/h Mpc using APM data as an input assuming only that peculiar motions are caused by peculiar gravity. The comparison to the local (Great Attractor) flow is expected to give clear information on the density parameter, Omega, and the local bias parameter, b. The observed peculiar flows in the Great Attractor region are found to be in better agreement with the open (Omega = 0.1) universe in which light traces mass (b = 1) than with a flat (Omega = 1) universe unless the bias parameter is unrealistically large (b is not less than 4). Constraints on Omega from a comparison of the APM and PV samples are discussed.
NASA Astrophysics Data System (ADS)
Schuite, Jonathan; Longuevergne, Laurent; Bour, Olivier; Boudin, Frédérick; Durand, Stéphane; Lavenant, Nicolas
2015-12-01
Fractured aquifers which bear valuable water resources are often difficult to characterize with classical hydrogeological tools due to their intrinsic heterogeneities. Here we implement ground surface deformation tools (tiltmetry and optical leveling) to monitor groundwater pressure changes induced by a classical hydraulic test at the Ploemeur observatory. By jointly analyzing complementary time constraining data (tilt) and spatially constraining data (vertical displacement), our results strongly suggest that the use of these surface deformation observations allows for estimating storativity and structural properties (dip, root depth, and lateral extension) of a large hydraulically active fracture, in good agreement with previous studies. Hence, we demonstrate that ground surface deformation is a useful addition to traditional hydrogeological techniques and opens possibilities for characterizing important large-scale properties of fractured aquifers with short-term well tests as a controlled forcing.
Coherent structures: Comments on mechanisms
NASA Technical Reports Server (NTRS)
Hunt, J. C. R.
1987-01-01
There is now overwhelming evidence that in most turbulent flows there exist regions moving with the flow where the velocity and vorticity have a characteristic structure. These regions are called coherent structures because within them the large-scale distributions of velocity and/or vorticity remain coherent even as these structures move through the flow and interact with other structures. Since the flow enters and leaves the bounding surfaces of these structures, a useful definition for coherent structures is that they are open volumes with distinctive large-scale vorticity distributions. Possible fruitful directions for the study of the dynamics of coherent structures are suggested. Most coherent structures research to data was concentrated on measurement and kinematical analysis; there is now a welcome move to examine the dynamics of coherent structures, by a variety of different methods. A few of them will be described.
Graduating general surgery resident operative confidence: perspective from a national survey.
Fonseca, Annabelle L; Reddy, Vikram; Longo, Walter E; Gusberg, Richard J
2014-08-01
General surgical training has changed significantly over the last decade with work hour restrictions, increasing subspecialization, the expanding use of minimally invasive techniques, and nonoperative management for solid organ trauma. Given these changes, this study was undertaken to assess the confidence of graduating general surgery residents in performing open surgical operations and to determine factors associated with increased confidence. A survey was developed and sent to general surgery residents nationally. We queried them regarding demographics and program characteristics, asked them to rate their confidence (rated 1-5 on a Likert scale) in performing open surgical procedures and compared those who indicated confidence with those who did not. We received 653 responses from the fifth year (postgraduate year 5) surgical residents: 69% male, 68% from university programs, and 51% from programs affiliated with a Veterans Affairs hospital; 22% from small programs, 34% from medium programs, and 44% from large programs. Anticipated postresidency operative confidence was 72%. More than 25% of residents reported a lack of confidence in performing eight of the 13 operations they were queried about. Training at a university program, a large program, dedicated research years, future fellowship plans, and training at a program that performed a large percentage of operations laparoscopically was associated with decreased confidence in performing a number of open surgical procedures. Increased surgical volume was associated with increased operative confidence. Confidence in performing open surgery also varied regionally. Graduating surgical residents indicated a significant lack of confidence in performing a variety of open surgical procedures. This decreased confidence was associated with age, operative volume as well as type, and location of training program. Analyzing and addressing this confidence deficit merits further study. Copyright © 2014 Elsevier Inc. All rights reserved.
Large-scale exact diagonalizations reveal low-momentum scales of nuclei
NASA Astrophysics Data System (ADS)
Forssén, C.; Carlsson, B. D.; Johansson, H. T.; Sääf, D.; Bansal, A.; Hagen, G.; Papenbrock, T.
2018-03-01
Ab initio methods aim to solve the nuclear many-body problem with controlled approximations. Virtually exact numerical solutions for realistic interactions can only be obtained for certain special cases such as few-nucleon systems. Here we extend the reach of exact diagonalization methods to handle model spaces with dimension exceeding 1010 on a single compute node. This allows us to perform no-core shell model (NCSM) calculations for 6Li in model spaces up to Nmax=22 and to reveal the 4He+d halo structure of this nucleus. Still, the use of a finite harmonic-oscillator basis implies truncations in both infrared (IR) and ultraviolet (UV) length scales. These truncations impose finite-size corrections on observables computed in this basis. We perform IR extrapolations of energies and radii computed in the NCSM and with the coupled-cluster method at several fixed UV cutoffs. It is shown that this strategy enables information gain also from data that is not fully UV converged. IR extrapolations improve the accuracy of relevant bound-state observables for a range of UV cutoffs, thus making them profitable tools. We relate the momentum scale that governs the exponential IR convergence to the threshold energy for the first open decay channel. Using large-scale NCSM calculations we numerically verify this small-momentum scale of finite nuclei.
Quantum fluctuations and CMB anisotropies in one-bubble open inflation models
NASA Astrophysics Data System (ADS)
Yamamoto, Kazuhiro; Sasaki, Misao; Tanaka, Takahiro
1996-10-01
We first develop a method to calculate a complete set of mode functions that describe the quantum fluctuations generated in one-bubble open inflation models. We consider two classes of models. One is a single scalar field model proposed by Bucher, Goldhaber, and Turok and by us as an example of the open inflation scenario, and the other is a two-field model such as the ``supernatural'' inflation proposed by Linde and Mezhlumian. In both cases we assume the difference in the vacuum energy density between inside and outside the bubble is negligible. There are two kinds of mode functions. One kind has the usual continuous spectrum and the other has a discrete spectrum with characteristic wavelengths exceeding the spatial curvature scale. The latter can be further divided into two classes in terms of its origin. One is called the de Sitter supercurvature mode, which arises due to the global spacetime structure of de Sitter space, and the other is due to fluctuations of the bubble wall. We calculate the spectrum of quantum fluctuations in these models and evaluate the resulting large angular scale CMB anisotropies. We find there are ranges of model parameters that are consistent with observed CMB anisotropies.
Pilot study of large-scale production of mutant pigs by ENU mutagenesis.
Hai, Tang; Cao, Chunwei; Shang, Haitao; Guo, Weiwei; Mu, Yanshuang; Yang, Shulin; Zhang, Ying; Zheng, Qiantao; Zhang, Tao; Wang, Xianlong; Liu, Yu; Kong, Qingran; Li, Kui; Wang, Dayu; Qi, Meng; Hong, Qianlong; Zhang, Rui; Wang, Xiupeng; Jia, Qitao; Wang, Xiao; Qin, Guosong; Li, Yongshun; Luo, Ailing; Jin, Weiwu; Yao, Jing; Huang, Jiaojiao; Zhang, Hongyong; Li, Menghua; Xie, Xiangmo; Zheng, Xuejuan; Guo, Kenan; Wang, Qinghua; Zhang, Shibin; Li, Liang; Xie, Fei; Zhang, Yu; Weng, Xiaogang; Yin, Zhi; Hu, Kui; Cong, Yimei; Zheng, Peng; Zou, Hailong; Xin, Leilei; Xia, Jihan; Ruan, Jinxue; Li, Hegang; Zhao, Weiming; Yuan, Jing; Liu, Zizhan; Gu, Weiwang; Li, Ming; Wang, Yong; Wang, Hongmei; Yang, Shiming; Liu, Zhonghua; Wei, Hong; Zhao, Jianguo; Zhou, Qi; Meng, Anming
2017-06-22
N-ethyl-N-nitrosourea (ENU) mutagenesis is a powerful tool to generate mutants on a large scale efficiently, and to discover genes with novel functions at the whole-genome level in Caenorhabditis elegans, flies, zebrafish and mice, but it has never been tried in large model animals. We describe a successful systematic three-generation ENU mutagenesis screening in pigs with the establishment of the Chinese Swine Mutagenesis Consortium. A total of 6,770 G1 and 6,800 G3 pigs were screened, 36 dominant and 91 recessive novel pig families with various phenotypes were established. The causative mutations in 10 mutant families were further mapped. As examples, the mutation of SOX10 (R109W) in pig causes inner ear malfunctions and mimics human Mondini dysplasia, and upregulated expression of FBXO32 is associated with congenital splay legs. This study demonstrates the feasibility of artificial random mutagenesis in pigs and opens an avenue for generating a reservoir of mutants for agricultural production and biomedical research.
Large-Eddy Simulation of Internal Flow through Human Vocal Folds
NASA Astrophysics Data System (ADS)
Lasota, Martin; Šidlof, Petr
2018-06-01
The phonatory process occurs when air is expelled from the lungs through the glottis and the pressure drop causes flow-induced oscillations of the vocal folds. The flow fields created in phonation are highly unsteady and the coherent vortex structures are also generated. For accuracy it is essential to compute on humanlike computational domain and appropriate mathematical model. The work deals with numerical simulation of air flow within the space between plicae vocales and plicae vestibulares. In addition to the dynamic width of the rima glottidis, where the sound is generated, there are lateral ventriculus laryngis and sacculus laryngis included in the computational domain as well. The paper presents the results from OpenFOAM which are obtained with a large-eddy simulation using second-order finite volume discretization of incompressible Navier-Stokes equations. Large-eddy simulations with different subgrid scale models are executed on structured mesh. In these cases are used only the subgrid scale models which model turbulence via turbulent viscosity and Boussinesq approximation in subglottal and supraglottal area in larynx.
Dispersion in Fractures with Ramified Dissolution Patterns
NASA Astrophysics Data System (ADS)
Xu, Le; Marks, Benjy; Toussaint, Renaud; Flekkøy, Eirik G.; Måløy, Knut J.
2018-04-01
The injection of a reactive fluid into an open fracture may modify the fracture surface locally and create a ramified structure around the injection point. This structure will have a significant impact on the dispersion of the injected fluid due to increased permeability, which will introduce large velocity fluctuations into the fluid. Here, we have injected a fluorescent tracer fluid into a transparent artificial fracture with such a ramified structure. The transparency of the model makes it possible to follow the detailed dispersion of the tracer concentration. The experiments have been compared to two dimensional (2D) computer simulations which include both convective motion and molecular diffusion. A comparison was also performed between the dispersion from an initially ramified dissolution structure and the dispersion from an initially circular region. A significant difference was seen both at small and large length scales. At large length scales, the persistence of the anisotropy of the concentration distribution far from the ramified structure is discussed with reference to some theoretical considerations and comparison with simulations.
Solar Cycle variations in Earth's open flux content measured by the SuperDARN radar network
NASA Astrophysics Data System (ADS)
Imber, S. M.; Milan, S. E.; Lester, M.
2013-09-01
We present a long term study, from 1996 - 2012, of the latitude of the Heppner-Maynard Boundary (HMB) determined using the northern hemisphere SuperDARN radars. The HMB represents the equatorward extent of ionospheric convection and is here used as a proxy for the amount of open flux in the polar cap. The mean HMB latitude (measured at midnight) is found to be at 64 degrees during the entire period, with secondary peaks at lower latitudes during the solar maximum of 2003, and at higher latitudes during the recent extreme solar minimum of 2008-2011. We associate these large scale statistical variations in open flux content with solar cycle variations in the solar wind parameters leading to changes in the intensity of the coupling between the solar wind and the magnetosphere.
The effects of the stellar wind and orbital motion on the jets of high-mass microquasars
NASA Astrophysics Data System (ADS)
Bosch-Ramon, V.; Barkov, M. V.
2016-05-01
Context. High-mass microquasar jets propagate under the effect of the wind from the companion star, and the orbital motion of the binary system. The stellar wind and the orbit may be dominant factors determining the jet properties beyond the binary scales. Aims: This is an analytical study, performed to characterise the effects of the stellar wind and the orbital motion on the jet properties. Methods: Accounting for the wind thrust transferred to the jet, we derive analytical estimates to characterise the jet evolution under the impact of the stellar wind. We include the Coriolis force effect, induced by orbital motion and enhanced by the wind's presence. Large-scale evolution of the jet is sketched, accounting for wind-to-jet thrust transfer, total energy conservation, and wind-jet flow mixing. Results: If the angle of the wind-induced jet bending is larger than its half-opening angle, the following is expected: (I) a strong recollimation shock; (II) bending against orbital motion, caused by Coriolis forces and enhanced by the wind presence; and (III) non-ballistic helical propagation further away. Even if disrupted, the jet can re-accelerate due to ambient pressure gradients, but wind entrainment can weaken this acceleration. On large scales, the opening angle of the helical structure is determined by the wind-jet thrust relation, and the wind-loaded jet flow can be rather slow. Conclusions: The impact of stellar winds on high-mass microquasar jets can yield non-ballistic helical jet trajectories, jet partial disruption and wind mixing, shocks, and possibly non-thermal emission. Among other observational diagnostics, such as radiation variability at any band, the radio morphology on milliarcsecond scales can be informative on the wind-jet interaction.
Košulič, Ondřej; Michalko, Radek; Hula, Vladimír
2016-01-01
Traditional woodland management created a mosaic of differently aged patches providing favorable conditions for a variety of arthropods. After abandonment of historical ownership patterns and traditional management and the deliberate transformation to high forest after World War II, large forest areas became darker and more homogeneous. This had significant negative consequences for biodiversity. An important question is whether even small-scale habitat structures maintained by different levels of canopy openness in abandoned coppiced forest may constitute conditions suitable for forest as well as open habitat specialists. We investigated the effect of canopy openness in former traditionally coppiced woodlands on the species richness, functional diversity, activity density, conservation value, and degree of rareness of epigeic spiders. In each of the eight studied locations, 60-m-long transect was established consisting of five pitfall traps placed at regular 15 m intervals along the gradient. Spiders were collected from May to July 2012. We recorded 90 spider species, including high proportions of xeric specialists (40%) and red-listed threatened species (26%). The peaks of conservation indicators, as well as spider community abundance, were shifted toward more open canopies. On the other hand, functional diversity peaked at more closed canopies followed by a rapid decrease with increasing canopy openness. Species richness was highest in the middle of the canopy openness gradient, suggesting an ecotone effect. Ordinations revealed that species of conservation concern tended to be associated with sparse and partly opened canopy. The results show that the various components of biodiversity peaked at different levels of canopy openness. Therefore, the restoration and suitable forest management of such conditions will retain important diversification of habitats in formerly coppiced oak forest stands. We indicate that permanent presence of small-scale improvements could be suitable conservation tools to prevent the general decline of woodland biodiversity in the intensified landscape of Central Europe. PMID:26845431
Bai, Ling; Mai, Van Cuong; Lim, Yun; Hou, Shuai; Möhwald, Helmuth; Duan, Hongwei
2018-03-01
Structural colors originating from interaction of light with intricately arranged micro-/nanostructures have stimulated considerable interest because of their inherent photostability and energy efficiency. In particular, noniridescent structural color with wide viewing angle has been receiving increasing attention recently. However, no method is yet available for rapid and large-scale fabrication of full-spectrum structural color patterns with wide viewing angles. Here, infiltration-driven nonequilibrium assembly of colloidal particles on liquid-permeable and particle-excluding substrates is demonstrated to direct the particles to form amorphous colloidal arrays (ACAs) within milliseconds. The infiltration-assisted (IFAST) colloidal assembly opens new possibilities for rapid manufacture of noniridescent structural colors of ACAs and straightforward structural color mixing. Full-spectrum noniridescent structural colors are successfully produced by mixing primary structural colors of red, blue, and yellow using a commercial office inkjet printer. Rapid fabrication of large-scale structural color patterns with sophisticated color combination/layout by IFAST printing is realized. The IFAST technology is versatile for developing structural color patterns with wide viewing angles, as colloidal particles, inks, and substrates are flexibly designable for diverse applications. © 2018 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Chang, Xueli; Du, Siliang; Li, Yingying; Fang, Shenghui
2018-01-01
Large size high resolution (HR) satellite image matching is a challenging task due to local distortion, repetitive structures, intensity changes and low efficiency. In this paper, a novel matching approach is proposed for the large size HR satellite image registration, which is based on coarse-to-fine strategy and geometric scale-invariant feature transform (SIFT). In the coarse matching step, a robust matching method scale restrict (SR) SIFT is implemented at low resolution level. The matching results provide geometric constraints which are then used to guide block division and geometric SIFT in the fine matching step. The block matching method can overcome the memory problem. In geometric SIFT, with area constraints, it is beneficial for validating the candidate matches and decreasing searching complexity. To further improve the matching efficiency, the proposed matching method is parallelized using OpenMP. Finally, the sensing image is rectified to the coordinate of reference image via Triangulated Irregular Network (TIN) transformation. Experiments are designed to test the performance of the proposed matching method. The experimental results show that the proposed method can decrease the matching time and increase the number of matching points while maintaining high registration accuracy. PMID:29702589
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dutta, Tanushree
Large-scale assemblies of people in a confined space can exert significant impacts on the local air chemistry due to human emissions of volatile organics. Variations of air-quality in such small scale can be studied by quantifying fingerprint volatile organic compounds (VOCs) such as acetone, toluene, and isoprene produced during concerts, movie screenings, and sport events (like the Olympics and the World Cup). This review summarizes the extent of VOC accumulation resulting from a large population in a confined area or in a small open area during sporting and other recreational activities. Apart from VOCs emitted directly from human bodies (e.g.,more » perspiration and exhaled breath), those released indirectly from other related sources (e.g., smoking, waste disposal, discharge of food-waste, and use of personal-care products) are also discussed. Although direct and indirect emissions of VOCs from human may constitute <1% of the global atmospheric VOCs budget, unique spatiotemporal variations in VOCs species within a confined space can have unforeseen impacts on the local atmosphere to lead to acute human exposure to harmful pollutants.« less
The future of emissions trading in light of the acid rain experience
DOE Office of Scientific and Technical Information (OSTI.GOV)
McLean, B.J.; Rico, R.
1995-12-31
The idea of emissions trading was developed more than two decades ago by environmental economists eager to provide new ideas for how to improve the efficiency of environmental protection. However, early emissions trading efforts were built on the historical {open_quotes}command and control{close_quotes} infrastructure which has dominated U.S. environmental protection until today. The {open_quotes}command and control{close_quotes} model initially had advantages that were of a very pragmatic character: it assured large pollution reductions in a time when large, cheap reductions were available and necessary; and it did not require a sophisticated government infrastructure. Within the last five years, large-scale emission trading programsmore » have been successfully designed and started that are fundamentally different from the earlier efforts, creating a new paradigm for environmental control just when our understanding of environmental problems is changing as well. The purpose of this paper is to focus on the largest national-scale program--the Acid Rain Program--and from that experience, forecast when emission trading programs may be headed based on our understanding of the factors currently influencing environmental management. The first section of this paper will briefly review the history of emissions trading programs, followed by a summary of the features of the Acid Rain Program, highlighting those features that distinguish it from previous efforts. The last section addresses the opportunities for emissions trading (and its probable future directions).« less
Large-size porous ZnO flakes with superior gas-sensing performance
NASA Astrophysics Data System (ADS)
Wen, Wei; Wu, Jin-Ming; Wang, Yu-De
2012-06-01
A simple top-down route is developed to fabricate large size porous ZnO flakes via solution combustion synthesis followed by a subsequent calcination in air, which is template-free and can be easily enlarged to an industrial scale. The achieved porous ZnO flakes, which are tens to hundreds of micrometers in flat and tens of nanometers in thickness, exhibit high response for detecting acetone and ethanol, because the unique two-dimensional architecture shortens effectively the gas diffusion distance and provides highly accessible open channels and active surfaces for the target gas.
Scaling Up Coordinate Descent Algorithms for Large ℓ1 Regularization Problems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Scherrer, Chad; Halappanavar, Mahantesh; Tewari, Ambuj
2012-07-03
We present a generic framework for parallel coordinate descent (CD) algorithms that has as special cases the original sequential algorithms of Cyclic CD and Stochastic CD, as well as the recent parallel Shotgun algorithm of Bradley et al. We introduce two novel parallel algorithms that are also special cases---Thread-Greedy CD and Coloring-Based CD---and give performance measurements for an OpenMP implementation of these.
Mary A. Arthur; Beth A. Blankenship; Angela Schörgendorfer; David L. Loftis; Heather D. Alexander
2015-01-01
Without large scale disturbances to alter forest structure and open the canopy, historically oak-dominated forests of the central and Appalachian hardwood regions of eastern North America are shifting to dominance by shade-tolerant, âmesophyticâ species. In response, prescribed fire is applied with increasing frequency and spatial extent to decrease non-oak species and...
Brian G. Tavernia; Mark D. Nelson; Michael E. Goerndt; Brian F. Walters; Chris Toney
2013-01-01
Large-scale and long-term habitat management plans are needed to maintain the diversity of habitat classes required by wildlife species. Planning efforts would benefit from assessments of potential climate and land-use change effects on habitats. We assessed climate and land-use driven changes in areas of closed- and open-canopy forest across the Northeast and Midwest...
Fujinuma, Junichi; Harrison, Rhett D
2012-01-01
Edge-effects greatly extend the area of tropical forests degraded through human activities. At Pasoh, Peninsular Malaysia, it has been suggested that soil disturbance by highly abundant wild pigs (Sus scrofa), which feed in adjacent Oil Palm plantations, may have mediated the invasion of Clidemia hirta (Melastomataceae) into the diverse tropical lowland rain forest. To investigate this hypothesis, we established three 1 km transects from the forest/Oil Palm plantation boundary into the forest interior. We recorded the distribution of soil disturbance by wild pigs, C. hirta abundance, and environmental variables. These data were analyzed using a hierarchical Bayesian model that incorporated spatial auto-correlation in the environmental variables. As predicted, soil disturbance by wild pigs declined with distance from forest edge and C. hirta abundance was correlated with the level of soil disturbance. Importantly there was no effect of distance on C. hirta abundance, after controlling for the effect of soil disturbance. Clidemia hirta abundance was also correlated with the presence of canopy openings, but there was no significant association between the occurrence of canopy openings and distance from the edge. Increased levels of soil disturbance and C. hirta abundance were still detectable approximately 1 km from the edge, demonstrating the potential for exceptionally large-scale animal mediated edge effects.
Fujinuma, Junichi; Harrison, Rhett D.
2012-01-01
Edge-effects greatly extend the area of tropical forests degraded through human activities. At Pasoh, Peninsular Malaysia, it has been suggested that soil disturbance by highly abundant wild pigs (Sus scrofa), which feed in adjacent Oil Palm plantations, may have mediated the invasion of Clidemia hirta (Melastomataceae) into the diverse tropical lowland rain forest. To investigate this hypothesis, we established three 1 km transects from the forest/Oil Palm plantation boundary into the forest interior. We recorded the distribution of soil disturbance by wild pigs, C. hirta abundance, and environmental variables. These data were analyzed using a hierarchical Bayesian model that incorporated spatial auto-correlation in the environmental variables. As predicted, soil disturbance by wild pigs declined with distance from forest edge and C. hirta abundance was correlated with the level of soil disturbance. Importantly there was no effect of distance on C. hirta abundance, after controlling for the effect of soil disturbance. Clidemia hirta abundance was also correlated with the presence of canopy openings, but there was no significant association between the occurrence of canopy openings and distance from the edge. Increased levels of soil disturbance and C. hirta abundance were still detectable approximately 1 km from the edge, demonstrating the potential for exceptionally large-scale animal mediated edge effects. PMID:22615977
NASA Astrophysics Data System (ADS)
Sakaizawa, Ryosuke; Kawai, Takaya; Sato, Toru; Oyama, Hiroyuki; Tsumune, Daisuke; Tsubono, Takaki; Goto, Koichi
2018-03-01
The target seas of tidal-current models are usually semi-closed bays, minimally affected by ocean currents. For these models, tidal currents are simulated in computational domains with a spatial scale of a couple hundred kilometers or less, by setting tidal elevations at their open boundaries. However, when ocean currents cannot be ignored in the sea areas of interest, such as in open seas near coastlines, it is necessary to include ocean-current effects in these tidal-current models. In this study, we developed a numerical method to analyze tidal currents near coasts by incorporating pre-calculated ocean-current velocities. First, a large regional-scale simulation with a spatial scale of several thousand kilometers was conducted and temporal changes in the ocean-current velocity at each grid point were stored. Next, the spatially and temporally interpolated ocean-current velocity was incorporated as forcing into the cross terms of the convection term of a tidal-current model having computational domains with spatial scales of hundreds of kilometers or less. Then, we applied this method to the diffusion of dissolved CO2 in a sea area off Tomakomai, Japan, and compared the numerical results and measurements to validate the proposed method.
NASA Astrophysics Data System (ADS)
Folsom, C. P.; Bouvier, J.; Petit, P.; Lèbre, A.; Amard, L.; Palacios, A.; Morin, J.; Donati, J.-F.; Vidotto, A. A.
2018-03-01
There is a large change in surface rotation rates of sun-like stars on the pre-main sequence and early main sequence. Since these stars have dynamo-driven magnetic fields, this implies a strong evolution of their magnetic properties over this time period. The spin-down of these stars is controlled by interactions between stellar and magnetic fields, thus magnetic evolution in turn plays an important role in rotational evolution. We present here the second part of a study investigating the evolution of large-scale surface magnetic fields in this critical time period. We observed stars in open clusters and stellar associations with known ages between 120 and 650 Myr, and used spectropolarimetry and Zeeman Doppler Imaging to characterize their large-scale magnetic field strength and geometry. We report 15 stars with magnetic detections here. These stars have masses from 0.8 to 0.95 M⊙, rotation periods from 0.326 to 10.6 d, and we find large-scale magnetic field strengths from 8.5 to 195 G with a wide range of geometries. We find a clear trend towards decreasing magnetic field strength with age, and a power law decrease in magnetic field strength with Rossby number. There is some tentative evidence for saturation of the large-scale magnetic field strength at Rossby numbers below 0.1, although the saturation point is not yet well defined. Comparing to younger classical T Tauri stars, we support the hypothesis that differences in internal structure produce large differences in observed magnetic fields, however for weak-lined T Tauri stars this is less clear.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Ruo-Yu; Rieger, F. M.; Aharonian, F. A., E-mail: ruoyu@mpi-hd.mpg.de, E-mail: frank.rieger@mpi-hd.mpg.de, E-mail: aharon@mpi-hd.mpg.de
The origin of the extended X-ray emission in the large-scale jets of active galactic nuclei (AGNs) poses challenges to conventional models of acceleration and emission. Although electron synchrotron radiation is considered the most feasible radiation mechanism, the formation of the continuous large-scale X-ray structure remains an open issue. As astrophysical jets are expected to exhibit some turbulence and shearing motion, we here investigate the potential of shearing flows to facilitate an extended acceleration of particles and evaluate its impact on the resultant particle distribution. Our treatment incorporates systematic shear and stochastic second-order Fermi effects. We show that for typical parametersmore » applicable to large-scale AGN jets, stochastic second-order Fermi acceleration, which always accompanies shear particle acceleration, can play an important role in facilitating the whole process of particle energization. We study the time-dependent evolution of the resultant particle distribution in the presence of second-order Fermi acceleration, shear acceleration, and synchrotron losses using a simple Fokker–Planck approach and provide illustrations for the possible emergence of a complex (multicomponent) particle energy distribution with different spectral branches. We present examples for typical parameters applicable to large-scale AGN jets, indicating the relevance of the underlying processes for understanding the extended X-ray emission and the origin of ultrahigh-energy cosmic rays.« less
From Pleistocene to Holocene: the prehistory of southwest Asia in evolutionary context.
Watkins, Trevor
2017-08-14
In this paper I seek to show how cultural niche construction theory offers the potential to extend the human evolutionary story beyond the Pleistocene, through the Neolithic, towards the kind of very large-scale societies in which we live today. The study of the human past has been compartmentalised, each compartment using different analytical vocabularies, so that their accounts are written in mutually incompatible languages. In recent years social, cognitive and cultural evolutionary theories, building on a growing body of archaeological evidence, have made substantial sense of the social and cultural evolution of the genus Homo. However, specialists in this field of studies have found it difficult to extend their kind of analysis into the Holocene human world. Within southwest Asia the three or four millennia of the Neolithic period at the beginning of the Holocene represents a pivotal point, which saw the transformation of human society in the emergence of the first large-scale, permanent communities, the domestication of plants and animals, and the establishment of effective farming economies. Following the Neolithic, the pace of human social, economic and cultural evolution continued to increase. By 5000 years ago, in parts of southwest Asia and northeast Africa there were very large-scale urban societies, and the first large-scale states (kingdoms). An extension of cultural niche construction theory enables us to extend the evolutionary narrative of the Pleistocene into the Holocene, opening the way to developing a single, long-term, evolutionary account of human history.
Business Development Corporation, Inc.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jasek, S.
1995-12-31
Business Development Corporation, Inc., is a company specializing in opportunity seeking and business development activities in the {open_quotes}new{close_quotes} post communist Central and Eastern Europe, with particular emphasis on the Republics of Poland and Slovakia. The company currently focuses its expertise on strategic investing and business development between Central Europe and the United States of America. In Poland and Slovakia, the company specializes in developing large scale energy and environmental {open_quotes}infrastructure{close_quotes} development projects on the federal, state, and local level. In addition, the company assists large state owned industries in the transformation and privatization process. Business Development Corporation has assisted andmore » continues to assist in projects of national importance. The staff of experts advise numerous large Polish and Slovak companies, most owned or in the process of privatization, on matters of restructuring, finance, capital structure, strategic parternships or investors, mergers, acquisitions and joint ventures with U.S. based firms. The company also assists and advises on a variety of environmental and energy matters in the public and private sector.« less
Measuring Self-Regulation in Self-Paced Open and Distance Learning Environments
ERIC Educational Resources Information Center
Kocdar, Serpil; Karadeniz, Abdulkadir; Bozkurt, Aras; Buyuk, Koksal
2018-01-01
Previous studies have described many scales for measuring self-regulation; however, no scale has been developed specifically for self-paced open and distance learning environments. Therefore, the aim of this study is to develop a scale for determining the self-regulated learning skills of distance learners in selfpaced open and distance learning…
Duchi, Diego; Gryte, Kristofer; Robb, Nicole C; Morichaud, Zakia; Sheppard, Carol; Wigneshweraraj, Sivaramesh
2018-01-01
Abstract Transcription initiation is a major step in gene regulation for all organisms. In bacteria, the promoter DNA is first recognized by RNA polymerase (RNAP) to yield an initial closed complex. This complex subsequently undergoes conformational changes resulting in DNA strand separation to form a transcription bubble and an RNAP-promoter open complex; however, the series and sequence of conformational changes, and the factors that influence them are unclear. To address the conformational landscape and transitions in transcription initiation, we applied single-molecule Förster resonance energy transfer (smFRET) on immobilized Escherichia coli transcription open complexes. Our results revealed the existence of two stable states within RNAP–DNA complexes in which the promoter DNA appears to adopt closed and partially open conformations, and we observed large-scale transitions in which the transcription bubble fluctuated between open and closed states; these transitions, which occur roughly on the 0.1 s timescale, are distinct from the millisecond-timescale dynamics previously observed within diffusing open complexes. Mutational studies indicated that the σ70 region 3.2 of the RNAP significantly affected the bubble dynamics. Our results have implications for many steps of transcription initiation, and support a bend-load-open model for the sequence of transitions leading to bubble opening during open complex formation. PMID:29177430
Bellamy, Chloe; Altringham, John
2015-01-01
Conservation increasingly operates at the landscape scale. For this to be effective, we need landscape scale information on species distributions and the environmental factors that underpin them. Species records are becoming increasingly available via data centres and online portals, but they are often patchy and biased. We demonstrate how such data can yield useful habitat suitability models, using bat roost records as an example. We analysed the effects of environmental variables at eight spatial scales (500 m - 6 km) on roost selection by eight bat species (Pipistrellus pipistrellus, P. pygmaeus, Nyctalus noctula, Myotis mystacinus, M. brandtii, M. nattereri, M. daubentonii, and Plecotus auritus) using the presence-only modelling software MaxEnt. Modelling was carried out on a selection of 418 data centre roost records from the Lake District National Park, UK. Target group pseudoabsences were selected to reduce the impact of sampling bias. Multi-scale models, combining variables measured at their best performing spatial scales, were used to predict roosting habitat suitability, yielding models with useful predictive abilities. Small areas of deciduous woodland consistently increased roosting habitat suitability, but other habitat associations varied between species and scales. Pipistrellus were positively related to built environments at small scales, and depended on large-scale woodland availability. The other, more specialist, species were highly sensitive to human-altered landscapes, avoiding even small rural towns. The strength of many relationships at large scales suggests that bats are sensitive to habitat modifications far from the roost itself. The fine resolution, large extent maps will aid targeted decision-making by conservationists and planners. We have made available an ArcGIS toolbox that automates the production of multi-scale variables, to facilitate the application of our methods to other taxa and locations. Habitat suitability modelling has the potential to become a standard tool for supporting landscape-scale decision-making as relevant data and open source, user-friendly, and peer-reviewed software become widely available.
Plastic debris in the open ocean
Cózar, Andrés; Echevarría, Fidel; González-Gordillo, J. Ignacio; Irigoien, Xabier; Úbeda, Bárbara; Hernández-León, Santiago; Palma, Álvaro T.; Navarro, Sandra; García-de-Lomas, Juan; Ruiz, Andrea; Fernández-de-Puelles, María L.; Duarte, Carlos M.
2014-01-01
There is a rising concern regarding the accumulation of floating plastic debris in the open ocean. However, the magnitude and the fate of this pollution are still open questions. Using data from the Malaspina 2010 circumnavigation, regional surveys, and previously published reports, we show a worldwide distribution of plastic on the surface of the open ocean, mostly accumulating in the convergence zones of each of the five subtropical gyres with comparable density. However, the global load of plastic on the open ocean surface was estimated to be on the order of tens of thousands of tons, far less than expected. Our observations of the size distribution of floating plastic debris point at important size-selective sinks removing millimeter-sized fragments of floating plastic on a large scale. This sink may involve a combination of fast nano-fragmentation of the microplastic into particles of microns or smaller, their transference to the ocean interior by food webs and ballasting processes, and processes yet to be discovered. Resolving the fate of the missing plastic debris is of fundamental importance to determine the nature and significance of the impacts of plastic pollution in the ocean. PMID:24982135
Plastic debris in the open ocean.
Cózar, Andrés; Echevarría, Fidel; González-Gordillo, J Ignacio; Irigoien, Xabier; Ubeda, Bárbara; Hernández-León, Santiago; Palma, Alvaro T; Navarro, Sandra; García-de-Lomas, Juan; Ruiz, Andrea; Fernández-de-Puelles, María L; Duarte, Carlos M
2014-07-15
There is a rising concern regarding the accumulation of floating plastic debris in the open ocean. However, the magnitude and the fate of this pollution are still open questions. Using data from the Malaspina 2010 circumnavigation, regional surveys, and previously published reports, we show a worldwide distribution of plastic on the surface of the open ocean, mostly accumulating in the convergence zones of each of the five subtropical gyres with comparable density. However, the global load of plastic on the open ocean surface was estimated to be on the order of tens of thousands of tons, far less than expected. Our observations of the size distribution of floating plastic debris point at important size-selective sinks removing millimeter-sized fragments of floating plastic on a large scale. This sink may involve a combination of fast nano-fragmentation of the microplastic into particles of microns or smaller, their transference to the ocean interior by food webs and ballasting processes, and processes yet to be discovered. Resolving the fate of the missing plastic debris is of fundamental importance to determine the nature and significance of the impacts of plastic pollution in the ocean.
NASA Astrophysics Data System (ADS)
Jana, Biman; Adkar, Bharat V.; Biswas, Rajib; Bagchi, Biman
2011-01-01
The catalytic conversion of adenosine triphosphate (ATP) and adenosine monophosphate (AMP) to adenosine diphosphate (ADP) by adenylate kinase (ADK) involves large amplitude, ligand induced domain motions, involving the opening and the closing of ATP binding domain (LID) and AMP binding domain (NMP) domains, during the repeated catalytic cycle. We discover and analyze an interesting dynamical coupling between the motion of the two domains during the opening, using large scale atomistic molecular dynamics trajectory analysis, covariance analysis, and multidimensional free energy calculations with explicit water. Initially, the LID domain must open by a certain amount before the NMP domain can begin to open. Dynamical correlation map shows interesting cross-peak between LID and NMP domain which suggests the presence of correlated motion between them. This is also reflected in our calculated two-dimensional free energy surface contour diagram which has an interesting elliptic shape, revealing a strong correlation between the opening of the LID domain and that of the NMP domain. Our free energy surface of the LID domain motion is rugged due to interaction with water and the signature of ruggedness is evident in the observed root mean square deviation variation and its fluctuation time correlation functions. We develop a correlated dynamical disorder-type theoretical model to explain the observed dynamic coupling between the motion of the two domains in ADK. Our model correctly reproduces several features of the cross-correlation observed in simulations.
Evaluation of a cognitive psychophysiological model for management of tic disorders: an open trial.
O'Connor, Kieron; Lavoie, Marc; Blanchet, Pierre; St-Pierre-Delorme, Marie-Ève
2016-07-01
Tic disorders, in particular chronic tic disorder and Tourette syndrome, affect about 1% of the population. The current treatment of choice is pharmacological or behavioural, addressing tics or the premonitory urges preceding tic onset. The current study reports an open trial evaluating the effectiveness of a cognitive psychophysiological treatment addressing Tourette-specific sensorimotor activation processes rather than the tic. Forty-nine people with Tourette syndrome and 36 people with chronic tics completed 10 weeks of individual cognitive psychophysiological therapy. Outcome measures included two tic severity scales and psychosocial measures. Post-treatment both groups had significantly improved on the tic scales with strong effect sizes across tic locations and complex and simple tics, maintained at 6-month follow-up with further change in perfectionism and self-esteem. The cognitive psychophysiological approach targeting underlying sensorimotor processes rather than tics in Tourette's and chronic tic disorder reduced symptoms with a large effect size. © The Royal College of Psychiatrists 2016.
Gillani, Nabeel; Yasseri, Taha; Eynon, Rebecca; Hjorth, Isis
2014-09-23
Massive Open Online Courses (MOOCs) bring together a global crowd of thousands of learners for several weeks or months. In theory, the openness and scale of MOOCs can promote iterative dialogue that facilitates group cognition and knowledge construction. Using data from two successive instances of a popular business strategy MOOC, we filter observed communication patterns to arrive at the "significant" interaction networks between learners and use complex network analysis to explore the vulnerability and information diffusion potential of the discussion forums. We find that different discussion topics and pedagogical practices promote varying levels of 1) "significant" peer-to-peer engagement, 2) participant inclusiveness in dialogue, and ultimately, 3) modularity, which impacts information diffusion to prevent a truly "global" exchange of knowledge and learning. These results indicate the structural limitations of large-scale crowd-based learning and highlight the different ways that learners in MOOCs leverage, and learn within, social contexts. We conclude by exploring how these insights may inspire new developments in online education.
Moseley, Michael William; Allerman, Andrew A.; Crawford, Mary H.; ...
2014-08-04
Electrical current transport through leakage paths in AlGaN-based deep ultraviolet (DUV) lightemitting diodes (LEDs) and their effect on LED performance are investigated. Open-core threading dislocations, or nanopipes, are found to conduct current through nominally insulating Al0.7Ga0.3N layers and limit the performance of DUV-LEDs. A defect-sensitive phosphoric acid etch reveals these opencore threading dislocations in the form of large, micron-scale hexagonal etch pits visible with optical microscopy, while closed-core screw-, edge-, and mixed-type threading dislocations are represented by smaller and more numerous nanometer-scale pits visible by atomic-force microscopy. The electrical and optical performances of DUV-LEDs fabricated on similar Si-doped Al0.7Ga0.3N templatesmore » are found to have a strong correlation to the density of these nanopipes, despite their small fraction (<0.1% in this study) of the total density of threading dislocations.« less
Gillani, Nabeel; Yasseri, Taha; Eynon, Rebecca; Hjorth, Isis
2014-01-01
Massive Open Online Courses (MOOCs) bring together a global crowd of thousands of learners for several weeks or months. In theory, the openness and scale of MOOCs can promote iterative dialogue that facilitates group cognition and knowledge construction. Using data from two successive instances of a popular business strategy MOOC, we filter observed communication patterns to arrive at the “significant” interaction networks between learners and use complex network analysis to explore the vulnerability and information diffusion potential of the discussion forums. We find that different discussion topics and pedagogical practices promote varying levels of 1) “significant” peer-to-peer engagement, 2) participant inclusiveness in dialogue, and ultimately, 3) modularity, which impacts information diffusion to prevent a truly “global” exchange of knowledge and learning. These results indicate the structural limitations of large-scale crowd-based learning and highlight the different ways that learners in MOOCs leverage, and learn within, social contexts. We conclude by exploring how these insights may inspire new developments in online education. PMID:25244925
NASA Astrophysics Data System (ADS)
Gillani, Nabeel; Yasseri, Taha; Eynon, Rebecca; Hjorth, Isis
2014-09-01
Massive Open Online Courses (MOOCs) bring together a global crowd of thousands of learners for several weeks or months. In theory, the openness and scale of MOOCs can promote iterative dialogue that facilitates group cognition and knowledge construction. Using data from two successive instances of a popular business strategy MOOC, we filter observed communication patterns to arrive at the ``significant'' interaction networks between learners and use complex network analysis to explore the vulnerability and information diffusion potential of the discussion forums. We find that different discussion topics and pedagogical practices promote varying levels of 1) ``significant'' peer-to-peer engagement, 2) participant inclusiveness in dialogue, and ultimately, 3) modularity, which impacts information diffusion to prevent a truly ``global'' exchange of knowledge and learning. These results indicate the structural limitations of large-scale crowd-based learning and highlight the different ways that learners in MOOCs leverage, and learn within, social contexts. We conclude by exploring how these insights may inspire new developments in online education.
He, Qiaoning; Yang, Haijian; Hu, Chunxiang
2016-10-01
Cultivation modes of autotrophic microalgae for biodiesel production utilizing open raceway pond were analyzed in this study. Five before screened good microalgae were tested their lipid productivity and biodiesel quality again in outdoor 1000L ORP. Then, Chlorella sp. L1 and Monoraphidium dybowskii Y2 were selected due to their stronger environmental adaptability, higher lipid productivity and better biodiesel properties. Further scale up cultivation for two species with batch and semi-continuous culture was conducted. In 40,000L ORP, higher lipid productivity (5.15 versus 4.06gm(-2)d(-1) for Chlorella sp. L1, 5.35 versus 3.00gm(-2)d(-1) for M. dybowskii Y2) was achieved in semi-continuous mode. Moreover, the financial costs of 14.18$gal(-1) and 13.31$gal(-1) for crude biodiesel in two microalgae with semi-continuous mode were more economically feasible for commercial production on large scale outdoors. Copyright © 2016 Elsevier Ltd. All rights reserved.
Open inflation in the landscape
NASA Astrophysics Data System (ADS)
Yamauchi, Daisuke; Linde, Andrei; Naruko, Atsushi; Sasaki, Misao; Tanaka, Takahiro
2011-08-01
The open inflation scenario is attracting a renewed interest in the context of the string landscape. Since there are a large number of metastable de Sitter vacua in the string landscape, tunneling transitions to lower metastable vacua through the bubble nucleation occur quite naturally, which leads to a natural realization of open inflation. Although the deviation of Ω0 from unity is small by the observational bound, we argue that the effect of this small deviation on the large-angle CMB anisotropies can be significant for tensor-type perturbation in the open inflation scenario. We consider the situation in which there is a large hierarchy between the energy scale of the quantum tunneling and that of the slow-roll inflation in the nucleated bubble. If the potential just after tunneling is steep enough, a rapid-roll phase appears before the slow-roll inflation. In this case the power spectrum is basically determined by the Hubble rate during the slow-roll inflation. On the other hand, if such a rapid-roll phase is absent, the power spectrum keeps the memory of the high energy density there in the large angular components. Furthermore, the amplitude of large angular components can be enhanced due to the effects of the wall fluctuation mode if the bubble wall tension is small. Therefore, although even the dominant quadrupole component is suppressed by the factor (1-Ω0)2, one can construct some models in which the deviation of Ω0 from unity is large enough to produce measurable effects. We also consider a more general class of models, where the false vacuum decay may occur due to Hawking-Moss tunneling, as well as the models involving more than one scalar field. We discuss scalar perturbations in these models and point out that a large set of such models is already ruled out by observational data, unless there was a very long stage of slow-roll inflation after the tunneling. These results show that observational data allow us to test various assumptions concerning the structure of the string theory potentials and the duration of the last stage of inflation.
NASA Astrophysics Data System (ADS)
Langouët, Loïc; Daire, Marie-Yvane
2009-12-01
The present-day maritime landscape of Western France forms the geographical framework for a recent research project dedicated to the archaeological study of ancient fish-traps, combining regional-scale and site-scale investigations. Based on the compilation and exploitation of a large unpublished dataset including more than 550 sites, a preliminary synthetic study allows us to present some examples of synchronic and thematic approaches, and propose a morphological classification of the weirs. These encouraging first results open up new perspectives on fish-trap chronology closely linked to wider studies on Holocene sea-level changes.
Sybil--efficient constraint-based modelling in R.
Gelius-Dietrich, Gabriel; Desouki, Abdelmoneim Amer; Fritzemeier, Claus Jonathan; Lercher, Martin J
2013-11-13
Constraint-based analyses of metabolic networks are widely used to simulate the properties of genome-scale metabolic networks. Publicly available implementations tend to be slow, impeding large scale analyses such as the genome-wide computation of pairwise gene knock-outs, or the automated search for model improvements. Furthermore, available implementations cannot easily be extended or adapted by users. Here, we present sybil, an open source software library for constraint-based analyses in R; R is a free, platform-independent environment for statistical computing and graphics that is widely used in bioinformatics. Among other functions, sybil currently provides efficient methods for flux-balance analysis (FBA), MOMA, and ROOM that are about ten times faster than previous implementations when calculating the effect of whole-genome single gene deletions in silico on a complete E. coli metabolic model. Due to the object-oriented architecture of sybil, users can easily build analysis pipelines in R or even implement their own constraint-based algorithms. Based on its highly efficient communication with different mathematical optimisation programs, sybil facilitates the exploration of high-dimensional optimisation problems on small time scales. Sybil and all its dependencies are open source. Sybil and its documentation are available for download from the comprehensive R archive network (CRAN).
Yarkoni, Tal
2012-01-01
Traditional pre-publication peer review of scientific output is a slow, inefficient, and unreliable process. Efforts to replace or supplement traditional evaluation models with open evaluation platforms that leverage advances in information technology are slowly gaining traction, but remain in the early stages of design and implementation. Here I discuss a number of considerations relevant to the development of such platforms. I focus particular attention on three core elements that next-generation evaluation platforms should strive to emphasize, including (1) open and transparent access to accumulated evaluation data, (2) personalized and highly customizable performance metrics, and (3) appropriate short-term incentivization of the userbase. Because all of these elements have already been successfully implemented on a large scale in hundreds of existing social web applications, I argue that development of new scientific evaluation platforms should proceed largely by adapting existing techniques rather than engineering entirely new evaluation mechanisms. Successful implementation of open evaluation platforms has the potential to substantially advance both the pace and the quality of scientific publication and evaluation, and the scientific community has a vested interest in shifting toward such models as soon as possible. PMID:23060783
PASSIM--an open source software system for managing information in biomedical studies.
Viksna, Juris; Celms, Edgars; Opmanis, Martins; Podnieks, Karlis; Rucevskis, Peteris; Zarins, Andris; Barrett, Amy; Neogi, Sudeshna Guha; Krestyaninova, Maria; McCarthy, Mark I; Brazma, Alvis; Sarkans, Ugis
2007-02-09
One of the crucial aspects of day-to-day laboratory information management is collection, storage and retrieval of information about research subjects and biomedical samples. An efficient link between sample data and experiment results is absolutely imperative for a successful outcome of a biomedical study. Currently available software solutions are largely limited to large-scale, expensive commercial Laboratory Information Management Systems (LIMS). Acquiring such LIMS indeed can bring laboratory information management to a higher level, but often implies sufficient investment of time, effort and funds, which are not always available. There is a clear need for lightweight open source systems for patient and sample information management. We present a web-based tool for submission, management and retrieval of sample and research subject data. The system secures confidentiality by separating anonymized sample information from individuals' records. It is simple and generic, and can be customised for various biomedical studies. Information can be both entered and accessed using the same web interface. User groups and their privileges can be defined. The system is open-source and is supplied with an on-line tutorial and necessary documentation. It has proven to be successful in a large international collaborative project. The presented system closes the gap between the need and the availability of lightweight software solutions for managing information in biomedical studies involving human research subjects.
MilxXplore: a web-based system to explore large imaging datasets.
Bourgeat, P; Dore, V; Villemagne, V L; Rowe, C C; Salvado, O; Fripp, J
2013-01-01
As large-scale medical imaging studies are becoming more common, there is an increasing reliance on automated software to extract quantitative information from these images. As the size of the cohorts keeps increasing with large studies, there is a also a need for tools that allow results from automated image processing and analysis to be presented in a way that enables fast and efficient quality checking, tagging and reporting on cases in which automatic processing failed or was problematic. MilxXplore is an open source visualization platform, which provides an interface to navigate and explore imaging data in a web browser, giving the end user the opportunity to perform quality control and reporting in a user friendly, collaborative and efficient way. Compared to existing software solutions that often provide an overview of the results at the subject's level, MilxXplore pools the results of individual subjects and time points together, allowing easy and efficient navigation and browsing through the different acquisitions of a subject over time, and comparing the results against the rest of the population. MilxXplore is fast, flexible and allows remote quality checks of processed imaging data, facilitating data sharing and collaboration across multiple locations, and can be easily integrated into a cloud computing pipeline. With the growing trend of open data and open science, such a tool will become increasingly important to share and publish results of imaging analysis.
NASA Astrophysics Data System (ADS)
Izett, Jonathan G.; Fennel, Katja
2018-02-01
Rivers deliver large amounts of fresh water, nutrients, and other terrestrially derived materials to the coastal ocean. Where inputs accumulate on the shelf, harmful effects such as hypoxia and eutrophication can result. In contrast, where export to the open ocean is efficient riverine inputs contribute to global biogeochemical budgets. Assessing the fate of riverine inputs is difficult on a global scale. Global ocean models are generally too coarse to resolve the relatively small scale features of river plumes. High-resolution regional models have been developed for individual river plume systems, but it is impractical to apply this approach globally to all rivers. Recently, generalized parameterizations have been proposed to estimate the export of riverine fresh water to the open ocean (Izett & Fennel, 2018, https://doi.org/10.1002/2017GB005667; Sharples et al., 2017, https://doi.org/10.1002/2016GB005483). Here the relationships of Izett and Fennel, https://doi.org/10.1002/2017GB005667 are used to derive global estimates of open-ocean export of fresh water and dissolved inorganic silicate, dissolved organic carbon, and dissolved organic and inorganic phosphorus and nitrogen. We estimate that only 15-53% of riverine fresh water reaches the open ocean directly in river plumes; nutrient export is even less efficient because of processing on continental shelves. Due to geographic differences in riverine nutrient delivery, dissolved silicate is the most efficiently exported to the open ocean (7-56.7%), while dissolved inorganic nitrogen is the least efficiently exported (2.8-44.3%). These results are consistent with previous estimates and provide a simple way to parameterize export to the open ocean in global models.
High evolutionary potential of marine zooplankton
Peijnenburg, Katja T C A; Goetze, Erica
2013-01-01
Abstract Open ocean zooplankton often have been viewed as slowly evolving species that have limited capacity to respond adaptively to changing ocean conditions. Hence, attention has focused on the ecological responses of zooplankton to current global change, including range shifts and changing phenology. Here, we argue that zooplankton also are well poised for evolutionary responses to global change. We present theoretical arguments that suggest plankton species may respond rapidly to selection on mildly beneficial mutations due to exceptionally large population size, and consider the circumstantial evidence that supports our inference that selection may be particularly important for these species. We also review all primary population genetic studies of open ocean zooplankton and show that genetic isolation can be achieved at the scale of gyre systems in open ocean habitats (100s to 1000s of km). Furthermore, population genetic structure often varies across planktonic taxa, and appears to be linked to the particular ecological requirements of the organism. In combination, these characteristics should facilitate adaptive evolution to distinct oceanographic habitats in the plankton. We conclude that marine zooplankton may be capable of rapid evolutionary as well as ecological responses to changing ocean conditions, and discuss the implications of this view. We further suggest two priority areas for future research to test our hypothesis of high evolutionary potential in open ocean zooplankton, which will require (1) assessing how pervasive selection is in driving population divergence and (2) rigorously quantifying the spatial and temporal scales of population differentiation in the open ocean. Recent attention has focused on the ecological responses of open ocean zooplankton to current global change, including range shifts and changing phenology. Here, we argue that marine zooplankton also are well poised for evolutionary responses to global change. PMID:24567838
Raduha, S.; Butler, D.; Mozley, P. S.; ...
2016-06-18
Here, we examined the potential impact on CO 2 transport of zones of deformation bands in reservoir rock that transition to opening-mode fractures within overlying caprock. Sedimentological and petrophysical measurements were collected along an approximately 5 m × 5 m outcrop of the Slick Rock and Earthy Members of the Entrada Sandstone on the eastern flank of the San Rafael Swell, Utah, USA. Measured deformation band permeability (2 mD) within the reservoir facies is about three orders of magnitude lower than the host sandstone. Average permeability of the caprock facies (0.0005 mD) is about seven orders of magnitude lower thanmore » the host sandstone. Aperture-based permeability estimates of the opening-mode caprock fractures are high (3.3 × 10 7 mD). High-resolution CO 2–H 2O transport models incorporate these permeability data at the millimeter scale. We then varied fault properties at the reservoir/caprock interface between open fractures and deformation bands as part of a sensitivity study. Numerical modeling results suggest that zones of deformation bands within the reservoir strongly compartmentalize reservoir pressures largely blocking lateral, cross-fault flow of supercritical CO 2. Significant vertical CO 2 transport into the caprock occurred in some scenarios along opening-mode fractures. The magnitude of this vertical CO 2 transport depends on the small-scale geometry of the contact between the opening-mode fracture and the zone of deformation bands, as well as the degree to which fractures penetrate caprock. Finally, the presence of relatively permeable units within the caprock allows storage of significant volumes of CO 2, particularly when the fracture network does not extend all the way through the caprock.« less
Hiding an elephant: heavy sterile neutrino with large mixing angle does not contradict cosmology
NASA Astrophysics Data System (ADS)
Bezrukov, F.; Chudaykin, A.; Gorbunov, D.
2017-06-01
We study a model of a keV-scale sterile neutrino with a relatively large mixing with the Standard Model sector. Usual considerations predict active generation of such particles in the early Universe, which leads to constraints from the total Dark Matter density and absence of X-ray signal from sterile neutrino decay. These bounds together may deem any attempt of creation of the keV scale sterile neutrino in the laboratory unfeasible. We argue that for models with a hidden sector coupled to the sterile neutrino these bounds can be evaded, opening new perspectives for the direct studies at neutrino experiments such as Troitsk ν-mass and KATRIN. We estimate the generation of sterile neutrinos in scenarios with the hidden sector dynamics keeping the sterile neutrinos either massless or superheavy in the early Universe. In both cases the generation by oscillations from active neutrinos in plasma is suppressed.
SureChEMBL: a large-scale, chemically annotated patent document database.
Papadatos, George; Davies, Mark; Dedman, Nathan; Chambers, Jon; Gaulton, Anna; Siddle, James; Koks, Richard; Irvine, Sean A; Pettersson, Joe; Goncharoff, Nicko; Hersey, Anne; Overington, John P
2016-01-04
SureChEMBL is a publicly available large-scale resource containing compounds extracted from the full text, images and attachments of patent documents. The data are extracted from the patent literature according to an automated text and image-mining pipeline on a daily basis. SureChEMBL provides access to a previously unavailable, open and timely set of annotated compound-patent associations, complemented with sophisticated combined structure and keyword-based search capabilities against the compound repository and patent document corpus; given the wealth of knowledge hidden in patent documents, analysis of SureChEMBL data has immediate applications in drug discovery, medicinal chemistry and other commercial areas of chemical science. Currently, the database contains 17 million compounds extracted from 14 million patent documents. Access is available through a dedicated web-based interface and data downloads at: https://www.surechembl.org/. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.
Teaching mathematics online in the European Area of Higher Education: an instructor's point of view
NASA Astrophysics Data System (ADS)
Juan, Angel A.; Steegmann, Cristina; Huertas, Antonia; Martinez, M. Jesus; Simosa, J.
2011-03-01
This article first discusses how information technologies are changing the way knowledge is delivered at universities worldwide. Then, the article reviews some of the most popular learning management systems available today and some of the most useful online resources in the areas of Mathematics and Statistics. After that, some long-term experiences regarding the teaching of online courses in those areas at the Open University of Catalonia are discussed. Finally, the article presents the results of a large-scale survey performed in Spain that aims to reflect instructors' opinions and feelings about potential benefits and challenges of teaching mathematics online, as well as the role of emergent technologies in the context of the European Area of Higher Education. Therefore, this article contributes to the existing literature as an additional reference point, one based on our long-term experience in a large-scale online environment, for discussions involving mathematical e-learning.
Non-classical photon correlation in a two-dimensional photonic lattice.
Gao, Jun; Qiao, Lu-Feng; Lin, Xiao-Feng; Jiao, Zhi-Qiang; Feng, Zhen; Zhou, Zheng; Gao, Zhen-Wei; Xu, Xiao-Yun; Chen, Yuan; Tang, Hao; Jin, Xian-Min
2016-06-13
Quantum interference and quantum correlation, as two main features of quantum optics, play an essential role in quantum information applications, such as multi-particle quantum walk and boson sampling. While many experimental demonstrations have been done in one-dimensional waveguide arrays, it remains unexplored in higher dimensions due to tight requirement of manipulating and detecting photons in large-scale. Here, we experimentally observe non-classical correlation of two identical photons in a fully coupled two-dimensional structure, i.e. photonic lattice manufactured by three-dimensional femtosecond laser writing. Photon interference consists of 36 Hong-Ou-Mandel interference and 9 bunching. The overlap between measured and simulated distribution is up to 0.890 ± 0.001. Clear photon correlation is observed in the two-dimensional photonic lattice. Combining with controllably engineered disorder, our results open new perspectives towards large-scale implementation of quantum simulation on integrated photonic chips.
SureChEMBL: a large-scale, chemically annotated patent document database
Papadatos, George; Davies, Mark; Dedman, Nathan; Chambers, Jon; Gaulton, Anna; Siddle, James; Koks, Richard; Irvine, Sean A.; Pettersson, Joe; Goncharoff, Nicko; Hersey, Anne; Overington, John P.
2016-01-01
SureChEMBL is a publicly available large-scale resource containing compounds extracted from the full text, images and attachments of patent documents. The data are extracted from the patent literature according to an automated text and image-mining pipeline on a daily basis. SureChEMBL provides access to a previously unavailable, open and timely set of annotated compound-patent associations, complemented with sophisticated combined structure and keyword-based search capabilities against the compound repository and patent document corpus; given the wealth of knowledge hidden in patent documents, analysis of SureChEMBL data has immediate applications in drug discovery, medicinal chemistry and other commercial areas of chemical science. Currently, the database contains 17 million compounds extracted from 14 million patent documents. Access is available through a dedicated web-based interface and data downloads at: https://www.surechembl.org/. PMID:26582922
Towards AI-powered personalization in MOOC learning
NASA Astrophysics Data System (ADS)
Yu, Han; Miao, Chunyan; Leung, Cyril; White, Timothy John
2017-12-01
Massive Open Online Courses (MOOCs) represent a form of large-scale learning that is changing the landscape of higher education. In this paper, we offer a perspective on how advances in artificial intelligence (AI) may enhance learning and research on MOOCs. We focus on emerging AI techniques including how knowledge representation tools can enable students to adjust the sequence of learning to fit their own needs; how optimization techniques can efficiently match community teaching assistants to MOOC mediation tasks to offer personal attention to learners; and how virtual learning companions with human traits such as curiosity and emotions can enhance learning experience on a large scale. These new capabilities will also bring opportunities for educational researchers to analyse students' learning skills and uncover points along learning paths where students with different backgrounds may require different help. Ethical considerations related to the application of AI in MOOC education research are also discussed.
NASA Astrophysics Data System (ADS)
Vigouroux, G.; Destouni, G.; Chen, Y.; Bring, A.; Jönsson, A.; Cvetkovic, V.
2017-12-01
Coastal areas link human-driven conditions on land with open sea conditions, and include crucial and vulnerable ecosystems that provide a variety of ecosystem services. Eutrophication is a common problem that is not least observed in the Baltic Sea, where coastal water quality is influenced both by land-based nutrient loading and by partly eutrophic open sea conditions. Robust and adaptive management of coastal systems is essential and necessitates integration of large scale catchment-coastal-marine systems as well as consideration of anthropogenic drivers and impacts, and climate change. To address this coastal challenge, relevant methodological approaches are required for characterization of coupled land, local coastal, and open sea conditions under an adaptive management framework for water quality. In this paper we present a new general and scalable dynamic characterization approach, developed for and applied to the Baltic Sea and its coastal areas. A simple carbon-based water quality model is implemented, dividing the Baltic Sea into main management basins that are linked to corresponding hydrological catchments on land, as well as to each other though aggregated three-dimensional marine hydrodynamics. Relevant hydrodynamic variables and associated water quality results have been validated on the Baltic Sea scale and show good accordance with available observation data and other modelling approaches. Based on its scalability, this methodology is further used on coastal zone scale to investigate the effects of hydrodynamic, hydro-climatic and nutrient load drivers on water quality and management implications for coastal areas in the Baltic Sea.
Code of Federal Regulations, 2013 CFR
2013-10-01
... Dimensions Using All-Bar Cuts (Triangular Cuts); Large Frame TED Escape Opening; Minimum Dimensions Using All-Bar Cuts and Leading Edge Cut; Large Frame TED Escape Opening; Minimum Dimensions Using All-Points...—Large Frame TED Escape Opening; Minimum Dimensions Using All-Bar Cuts (Triangular Cuts); Large Frame TED...
Code of Federal Regulations, 2012 CFR
2012-10-01
... Dimensions Using All-Bar Cuts (Triangular Cuts); Large Frame TED Escape Opening; Minimum Dimensions Using All-Bar Cuts and Leading Edge Cut; Large Frame TED Escape Opening; Minimum Dimensions Using All-Points...—Large Frame TED Escape Opening; Minimum Dimensions Using All-Bar Cuts (Triangular Cuts); Large Frame TED...
Code of Federal Regulations, 2014 CFR
2014-10-01
... Dimensions Using All-Bar Cuts (Triangular Cuts); Large Frame TED Escape Opening; Minimum Dimensions Using All-Bar Cuts and Leading Edge Cut; Large Frame TED Escape Opening; Minimum Dimensions Using All-Points...—Large Frame TED Escape Opening; Minimum Dimensions Using All-Bar Cuts (Triangular Cuts); Large Frame TED...
Recombinant DNA production of spider silk proteins
Tokareva, Olena; Michalczechen-Lacerda, Valquíria A; Rech, Elíbio L; Kaplan, David L
2013-01-01
Spider dragline silk is considered to be the toughest biopolymer on Earth due to an extraordinary combination of strength and elasticity. Moreover, silks are biocompatible and biodegradable protein-based materials. Recent advances in genetic engineering make it possible to produce recombinant silks in heterologous hosts, opening up opportunities for large-scale production of recombinant silks for various biomedical and material science applications. We review the current strategies to produce recombinant spider silks. PMID:24119078
Network models of biology, whether curated or derived from large-scale data analysis, are critical tools in the understanding of cancer mechanisms and in the design and personalization of therapies. The NDEx Project (Network Data Exchange) will create, deploy, and maintain an open-source, web-based software platform and public website to enable scientists, organizations, and software applications to share, store, manipulate, and publish biological networks.
2015-05-01
decisions on the fly in an online retail environment. Tech. rep., Working Paper, Massachusetts Institute of Technology, Boston, MA. Arneson, Broderick , Ryan...Hayward, Philip Henderson. 2009. MoHex wins Hex tournament. International Computer Games Association Journal 32 114–116. Arneson, Broderick , Ryan B...Combina- torial Search. Enzenberger, Markus, Martin Muller, Broderick Arneson, Richard Segal. 2010. Fuego—an open-source framework for board games and
ERIC Educational Resources Information Center
Pete, Judith; Mulder, Fred; Neto, Jose Dutra Oliveira
2017-01-01
In order to obtain a fair "OER picture" for the Global South a large-scale study has been carried out for a series of countries, including Kenya. In this paper we report on the Kenya study, run at four universities that have been selected with randomly sampled students and lecturers. Empirical data have been generated by the use of a…
Battlespace Awareness: Heterogeneous Sensor Maps of Large Scale, Complex Environments
2017-06-13
reference frames enable a system designer to describe the position of any sensor or platform at any point of time. This section introduces the...analysis to evaluate the quality of reconstructions created by our algorithms. CloudCompare is an open-source tool designed for this purpose [65]. In...structure of the data. The data term seeks to keep the proposed solution (u) similar to the originally observed values ( f ). A systems designer must
Reverse-phase protein arrays (RPPA) represent a powerful functional proteomic approach to elucidate cancer-related molecular mechanisms and to develop novel cancer therapies. To facilitate community-based investigation of the large-scale protein expression data generated by this platform, we have developed a user-friendly, open-access bioinformatic resource, The Cancer Proteome Atlas (TCPA, http://tcpaportal.org), which contains two separate web applications.
Multi-scale virtual view on the precessing jet SS433
NASA Astrophysics Data System (ADS)
Monceau-Baroux, R.; Porth, O.; Meliani, Z.; Keppens, R.
2014-07-01
Observations of SS433 infer how an X-ray binary gives rise to a corkscrew patterned relativistic jet. XRB SS433 is well known on a large range of scales for wich we realize 3D simulation and radio mappings. For our study we use relativistic hydrodynamic in special relativity using a relativistic effective polytropic index. We use parameters extracted from observations to impose thermodynamical conditions of the ISM and jet. We follow the kinetic and thermal energy content, of the various ISM and jet regions. Our simulation follows simultaneously the evolution of the population of electrons which are accelerated by the jet. The evolving spectrum of these electrons, together with an assumed equipartition between dynamic and magnetic pressure, gives input for estimating the radio emission from our simulation. Ray tracing according to a direction of sight then realizes radio mappings of our data. Single snapshots are realised to compare with VLA observation as in Roberts et al. 2008. A radio movie is realised to compare with the 41 days movie made with the VLBA instrument. Finaly a larger scale simulation explore the discrepancy of opening angle between 10 and 20 degree between the large scale observation of SS433 and its close in observation.
BlazeDEM3D-GPU A Large Scale DEM simulation code for GPUs
NASA Astrophysics Data System (ADS)
Govender, Nicolin; Wilke, Daniel; Pizette, Patrick; Khinast, Johannes
2017-06-01
Accurately predicting the dynamics of particulate materials is of importance to numerous scientific and industrial areas with applications ranging across particle scales from powder flow to ore crushing. Computational discrete element simulations is a viable option to aid in the understanding of particulate dynamics and design of devices such as mixers, silos and ball mills, as laboratory scale tests comes at a significant cost. However, the computational time required to simulate an industrial scale simulation which consists of tens of millions of particles can take months to complete on large CPU clusters, making the Discrete Element Method (DEM) unfeasible for industrial applications. Simulations are therefore typically restricted to tens of thousands of particles with highly detailed particle shapes or a few million of particles with often oversimplified particle shapes. However, a number of applications require accurate representation of the particle shape to capture the macroscopic behaviour of the particulate system. In this paper we give an overview of the recent extensions to the open source GPU based DEM code, BlazeDEM3D-GPU, that can simulate millions of polyhedra and tens of millions of spheres on a desktop computer with a single or multiple GPUs.
2016-01-01
The development of new antimalarial compounds remains a pivotal part of the strategy for malaria elimination. Recent large-scale phenotypic screens have provided a wealth of potential starting points for hit-to-lead campaigns. One such public set is explored, employing an open source research mechanism in which all data and ideas were shared in real time, anyone was able to participate, and patents were not sought. One chemical subseries was found to exhibit oral activity but contained a labile ester that could not be replaced without loss of activity, and the original hit exhibited remarkable sensitivity to minor structural change. A second subseries displayed high potency, including activity within gametocyte and liver stage assays, but at the cost of low solubility. As an open source research project, unexplored avenues are clearly identified and may be explored further by the community; new findings may be cumulatively added to the present work. PMID:27800551
Aerodynamic effects by cooling flows within engine room of a car model
NASA Astrophysics Data System (ADS)
Sawaguchi, T.; Takakura, Y.
2017-10-01
The purpose of this research is to clarify the change of characteristics of aerodynamic drag and lift of a car by the engine loading system (engine arrangement) and the air inlet system (opening area and position) with and without a radiator in wind-tunnel experiments. A simplified car model with 1/5 scale is generated with reproduction of the engine room covered with the transparent acryl externals for visualization. In the wind-tunnel experiments, the moving-belt ground board is adopted to include ground effects with force measurements by use of load cells. The flows are visualized by the smoke method. As results, with enlargement of the opening area, the drag increased overall although depending largely on the engine loading system and the inlet opening position, the front lift increased and the rear left decreased; the effect of the radiator was to relieve the change of the drag and lift.
The beaming of subhalo accretion
NASA Astrophysics Data System (ADS)
Libeskind, Noam I.
2016-10-01
We examine the infall pattern of subhaloes onto hosts in the context of the large-scale structure. We find that the infall pattern is essentially driven by the shear tensor of the ambient velocity field. Dark matter subhaloes are preferentially accreted along the principal axis of the shear tensor which corresponds to the direction of weakest collapse. We examine the dependence of this preferential infall on subhalo mass, host halo mass and redshift. Although strongest for the most massive hosts and the most massive subhaloes at high redshift, the preferential infall of subhaloes is effectively universal in the sense that its always aligned with the axis of weakest collapse of the velocity shear tensor. It is the same shear tensor that dictates the structure of the cosmic web and hence the shear field emerges as the key factor that governs the local anisotropic pattern of structure formation. Since the small (sub-Mpc) scale is strongly correlated with the mid-range (~ 10 Mpc) scale - a scale accessible by current surveys of peculiar velocities - it follows that findings presented here open a new window into the relation between the observed large scale structure unveiled by current surveys of peculiar velocities and the preferential infall direction of the Local Group. This may shed light on the unexpected alignments of dwarf galaxies seen in the Local Group.
Rainbow: a tool for large-scale whole-genome sequencing data analysis using cloud computing.
Zhao, Shanrong; Prenger, Kurt; Smith, Lance; Messina, Thomas; Fan, Hongtao; Jaeger, Edward; Stephens, Susan
2013-06-27
Technical improvements have decreased sequencing costs and, as a result, the size and number of genomic datasets have increased rapidly. Because of the lower cost, large amounts of sequence data are now being produced by small to midsize research groups. Crossbow is a software tool that can detect single nucleotide polymorphisms (SNPs) in whole-genome sequencing (WGS) data from a single subject; however, Crossbow has a number of limitations when applied to multiple subjects from large-scale WGS projects. The data storage and CPU resources that are required for large-scale whole genome sequencing data analyses are too large for many core facilities and individual laboratories to provide. To help meet these challenges, we have developed Rainbow, a cloud-based software package that can assist in the automation of large-scale WGS data analyses. Here, we evaluated the performance of Rainbow by analyzing 44 different whole-genome-sequenced subjects. Rainbow has the capacity to process genomic data from more than 500 subjects in two weeks using cloud computing provided by the Amazon Web Service. The time includes the import and export of the data using Amazon Import/Export service. The average cost of processing a single sample in the cloud was less than 120 US dollars. Compared with Crossbow, the main improvements incorporated into Rainbow include the ability: (1) to handle BAM as well as FASTQ input files; (2) to split large sequence files for better load balance downstream; (3) to log the running metrics in data processing and monitoring multiple Amazon Elastic Compute Cloud (EC2) instances; and (4) to merge SOAPsnp outputs for multiple individuals into a single file to facilitate downstream genome-wide association studies. Rainbow is a scalable, cost-effective, and open-source tool for large-scale WGS data analysis. For human WGS data sequenced by either the Illumina HiSeq 2000 or HiSeq 2500 platforms, Rainbow can be used straight out of the box. Rainbow is available for third-party implementation and use, and can be downloaded from http://s3.amazonaws.com/jnj_rainbow/index.html.
Rainbow: a tool for large-scale whole-genome sequencing data analysis using cloud computing
2013-01-01
Background Technical improvements have decreased sequencing costs and, as a result, the size and number of genomic datasets have increased rapidly. Because of the lower cost, large amounts of sequence data are now being produced by small to midsize research groups. Crossbow is a software tool that can detect single nucleotide polymorphisms (SNPs) in whole-genome sequencing (WGS) data from a single subject; however, Crossbow has a number of limitations when applied to multiple subjects from large-scale WGS projects. The data storage and CPU resources that are required for large-scale whole genome sequencing data analyses are too large for many core facilities and individual laboratories to provide. To help meet these challenges, we have developed Rainbow, a cloud-based software package that can assist in the automation of large-scale WGS data analyses. Results Here, we evaluated the performance of Rainbow by analyzing 44 different whole-genome-sequenced subjects. Rainbow has the capacity to process genomic data from more than 500 subjects in two weeks using cloud computing provided by the Amazon Web Service. The time includes the import and export of the data using Amazon Import/Export service. The average cost of processing a single sample in the cloud was less than 120 US dollars. Compared with Crossbow, the main improvements incorporated into Rainbow include the ability: (1) to handle BAM as well as FASTQ input files; (2) to split large sequence files for better load balance downstream; (3) to log the running metrics in data processing and monitoring multiple Amazon Elastic Compute Cloud (EC2) instances; and (4) to merge SOAPsnp outputs for multiple individuals into a single file to facilitate downstream genome-wide association studies. Conclusions Rainbow is a scalable, cost-effective, and open-source tool for large-scale WGS data analysis. For human WGS data sequenced by either the Illumina HiSeq 2000 or HiSeq 2500 platforms, Rainbow can be used straight out of the box. Rainbow is available for third-party implementation and use, and can be downloaded from http://s3.amazonaws.com/jnj_rainbow/index.html. PMID:23802613
Seqcrawler: biological data indexing and browsing platform.
Sallou, Olivier; Bretaudeau, Anthony; Roult, Aurelien
2012-07-24
Seqcrawler takes its roots in software like SRS or Lucegene. It provides an indexing platform to ease the search of data and meta-data in biological banks and it can scale to face the current flow of data. While many biological bank search tools are available on the Internet, mainly provided by large organizations to search their data, there is a lack of free and open source solutions to browse one's own set of data with a flexible query system and able to scale from a single computer to a cloud system. A personal index platform will help labs and bioinformaticians to search their meta-data but also to build a larger information system with custom subsets of data. The software is scalable from a single computer to a cloud-based infrastructure. It has been successfully tested in a private cloud with 3 index shards (pieces of index) hosting ~400 millions of sequence information (whole GenBank, UniProt, PDB and others) for a total size of 600 GB in a fault tolerant architecture (high-availability). It has also been successfully integrated with software to add extra meta-data from blast results to enhance users' result analysis. Seqcrawler provides a complete open source search and store solution for labs or platforms needing to manage large amount of data/meta-data with a flexible and customizable web interface. All components (search engine, visualization and data storage), though independent, share a common and coherent data system that can be queried with a simple HTTP interface. The solution scales easily and can also provide a high availability infrastructure.
Sediment depositions upstream of open check dams: new elements from small scale models
NASA Astrophysics Data System (ADS)
Piton, Guillaume; Le Guern, Jules; Carbonari, Costanza; Recking, Alain
2015-04-01
Torrent hazard mitigation remains a big issue in mountainous regions. In steep slope streams and especially in their fan part, torrential floods mainly result from abrupt and massive sediment deposits. To curtail such phenomenon, soil conservation measures as well as torrent control works have been undertaken for decades. Since the 1950s, open check dams complete other structural and non-structural measures in watershed scale mitigation plans1. They are often built to trap sediments near the fan apexes. The development of earthmoving machinery after the WWII facilitated the dredging operations of open check dams. Hundreds of these structures have thus been built for 60 years. Their design evolved with the improving comprehension of torrential hydraulics and sediment transport; however this kind of structure has a general tendency to trap most of the sediments supplied by the headwaters. Secondary effects as channel incision downstream of the traps often followed an open check dam creation. This sediment starvation trend tends to propagate to the main valley rivers and to disrupt past geomorphic equilibriums. Taking it into account and to diminish useless dredging operation, a better selectivity of sediment trapping must be sought in open check dams, i.e. optimal open check dams would trap sediments during dangerous floods and flush them during normal small floods. An accurate description of the hydraulic and deposition processes that occur in sediment traps is needed to optimize existing structures and to design best-adjusted new structures. A literature review2 showed that if design criteria exist for the structure itself, little information is available on the dynamic of the sediment depositions upstream of open check dams, i.e. what are the geomorphic patterns that occur during the deposition?, what are the relevant friction laws and sediment transport formula that better describe massive depositions in sediment traps?, what are the range of Froude and Shields numbers that the flows tend to adopt? New small scale model experiments have been undertaken focusing on depositions processes and their related hydraulics. Accurate photogrammetric measurements allowed us to better describe the deposition processes3. Large Scale Particle Image Velocimetry (LS-PIV) was performed to determine surface velocity fields in highly active channels with low grain submersion4. We will present preliminary results of our experiments showing the new elements we observed in massive deposit dynamics. REFERENCES 1.Armanini, A., Dellagiacoma, F. & Ferrari, L. From the check dam to the development of functional check dams. Fluvial Hydraulics of Mountain Regions 37, 331-344 (1991). 2.Piton, G. & Recking, A. Design of sediment traps with open check dams: a review, part I: hydraulic and deposition processes. (Accepted by the) Journal of Hydraulic Engineering 1-23 (2015). 3.Le Guern, J. Ms Thesis: Modélisation physique des plages de depot : analyse de la dynamique de remplissage.(2014) . 4.Carbonari, C. Ms Thesis: Small scale experiments of deposition processes occuring in sediment traps, LS-PIV measurments and geomorphological descriptions. (in preparation).
NASA Astrophysics Data System (ADS)
McClain, Bobbi J.; Porter, William F.
2000-11-01
Satellite imagery is a useful tool for large-scale habitat analysis; however, its limitations need to be tested. We tested these limitations by varying the methods of a habitat evaluation for white-tailed deer ( Odocoileus virginianus) in the Adirondack Park, New York, USA, utilizing harvest data to create and validate the assessment models. We used two classified images, one with a large minimum mapping unit but high accuracy and one with no minimum mapping unit but slightly lower accuracy, to test the sensitivity of the evaluation to these differences. We tested the utility of two methods of assessment, habitat suitability index modeling, and pattern recognition modeling. We varied the scale at which the models were applied by using five separate sizes of analysis windows. Results showed that the presence of a large minimum mapping unit eliminates important details of the habitat. Window size is relatively unimportant if the data are averaged to a large resolution (i.e., township), but if the data are used at the smaller resolution, then the window size is an important consideration. In the Adirondacks, the proportion of hardwood and softwood in an area is most important to the spatial dynamics of deer populations. The low occurrence of open area in all parts of the park either limits the effect of this cover type on the population or limits our ability to detect the effect. The arrangement and interspersion of cover types were not significant to deer populations.
Large-scale inverse model analyses employing fast randomized data reduction
NASA Astrophysics Data System (ADS)
Lin, Youzuo; Le, Ellen B.; O'Malley, Daniel; Vesselinov, Velimir V.; Bui-Thanh, Tan
2017-08-01
When the number of observations is large, it is computationally challenging to apply classical inverse modeling techniques. We have developed a new computationally efficient technique for solving inverse problems with a large number of observations (e.g., on the order of 107 or greater). Our method, which we call the randomized geostatistical approach (RGA), is built upon the principal component geostatistical approach (PCGA). We employ a data reduction technique combined with the PCGA to improve the computational efficiency and reduce the memory usage. Specifically, we employ a randomized numerical linear algebra technique based on a so-called "sketching" matrix to effectively reduce the dimension of the observations without losing the information content needed for the inverse analysis. In this way, the computational and memory costs for RGA scale with the information content rather than the size of the calibration data. Our algorithm is coded in Julia and implemented in the MADS open-source high-performance computational framework (http://mads.lanl.gov). We apply our new inverse modeling method to invert for a synthetic transmissivity field. Compared to a standard geostatistical approach (GA), our method is more efficient when the number of observations is large. Most importantly, our method is capable of solving larger inverse problems than the standard GA and PCGA approaches. Therefore, our new model inversion method is a powerful tool for solving large-scale inverse problems. The method can be applied in any field and is not limited to hydrogeological applications such as the characterization of aquifer heterogeneity.
Parallel algorithm for multiscale atomistic/continuum simulations using LAMMPS
NASA Astrophysics Data System (ADS)
Pavia, F.; Curtin, W. A.
2015-07-01
Deformation and fracture processes in engineering materials often require simultaneous descriptions over a range of length and time scales, with each scale using a different computational technique. Here we present a high-performance parallel 3D computing framework for executing large multiscale studies that couple an atomic domain, modeled using molecular dynamics and a continuum domain, modeled using explicit finite elements. We use the robust Coupled Atomistic/Discrete-Dislocation (CADD) displacement-coupling method, but without the transfer of dislocations between atoms and continuum. The main purpose of the work is to provide a multiscale implementation within an existing large-scale parallel molecular dynamics code (LAMMPS) that enables use of all the tools associated with this popular open-source code, while extending CADD-type coupling to 3D. Validation of the implementation includes the demonstration of (i) stability in finite-temperature dynamics using Langevin dynamics, (ii) elimination of wave reflections due to large dynamic events occurring in the MD region and (iii) the absence of spurious forces acting on dislocations due to the MD/FE coupling, for dislocations further than 10 Å from the coupling boundary. A first non-trivial example application of dislocation glide and bowing around obstacles is shown, for dislocation lengths of ∼50 nm using fewer than 1 000 000 atoms but reproducing results of extremely large atomistic simulations at much lower computational cost.
Ecosystem variability in the offshore northeastern Chukchi Sea
NASA Astrophysics Data System (ADS)
Blanchard, Arny L.; Day, Robert H.; Gall, Adrian E.; Aerts, Lisanne A. M.; Delarue, Julien; Dobbins, Elizabeth L.; Hopcroft, Russell R.; Questel, Jennifer M.; Weingartner, Thomas J.; Wisdom, Sheyna S.
2017-12-01
Understanding influences of cumulative effects from multiple stressors in marine ecosystems requires an understanding of the sources for and scales of variability. A multidisciplinary ecosystem study in the offshore northeastern Chukchi Sea during 2008-2013 investigated the variability of the study area's two adjacent sub-ecosystems: a pelagic system influenced by interannual and/or seasonal temporal variation at large, oceanographic (regional) scales, and a benthic-associated system more influenced by small-scale spatial variations. Variability in zooplankton communities reflected interannual oceanographic differences in waters advected northward from the Bering Sea, whereas variation in benthic communities was associated with seafloor and bottom-water characteristics. Variations in the planktivorous seabird community were correlated with prey distributions, whereas interaction effects in ANOVA for walruses were related to declines of sea-ice. Long-term shifts in seabird distributions were also related to changes in sea-ice distributions that led to more open water. Although characteristics of the lower trophic-level animals within sub-ecosystems result from oceanographic variations and interactions with seafloor topography, distributions of apex predators were related to sea-ice as a feeding platform (walruses) or to its absence (i.e., open water) for feeding (seabirds). The stability of prey resources appears to be a key factor in mediating predator interactions with other ocean characteristics. Seabirds reliant on highly-variable zooplankton prey show long-term changes as open water increases, whereas walruses taking benthic prey in biomass hotspots respond to sea-ice changes in the short-term. A better understanding of how variability scales up from prey to predators and how prey resource stability (including how critical prey respond to environmental changes over space and time) might be altered by climate and anthropogenic stressors is essential to predicting the future state of both the Chukchi and other arctic systems.
OpenSHS: Open Smart Home Simulator.
Alshammari, Nasser; Alshammari, Talal; Sedky, Mohamed; Champion, Justin; Bauer, Carolin
2017-05-02
This paper develops a new hybrid, open-source, cross-platform 3D smart home simulator, OpenSHS, for dataset generation. OpenSHS offers an opportunity for researchers in the field of the Internet of Things (IoT) and machine learning to test and evaluate their models. Following a hybrid approach, OpenSHS combines advantages from both interactive and model-based approaches. This approach reduces the time and efforts required to generate simulated smart home datasets. We have designed a replication algorithm for extending and expanding a dataset. A small sample dataset produced, by OpenSHS, can be extended without affecting the logical order of the events. The replication provides a solution for generating large representative smart home datasets. We have built an extensible library of smart devices that facilitates the simulation of current and future smart home environments. Our tool divides the dataset generation process into three distinct phases: first design: the researcher designs the initial virtual environment by building the home, importing smart devices and creating contexts; second, simulation: the participant simulates his/her context-specific events; and third, aggregation: the researcher applies the replication algorithm to generate the final dataset. We conducted a study to assess the ease of use of our tool on the System Usability Scale (SUS).
OpenSHS: Open Smart Home Simulator
Alshammari, Nasser; Alshammari, Talal; Sedky, Mohamed; Champion, Justin; Bauer, Carolin
2017-01-01
This paper develops a new hybrid, open-source, cross-platform 3D smart home simulator, OpenSHS, for dataset generation. OpenSHS offers an opportunity for researchers in the field of the Internet of Things (IoT) and machine learning to test and evaluate their models. Following a hybrid approach, OpenSHS combines advantages from both interactive and model-based approaches. This approach reduces the time and efforts required to generate simulated smart home datasets. We have designed a replication algorithm for extending and expanding a dataset. A small sample dataset produced, by OpenSHS, can be extended without affecting the logical order of the events. The replication provides a solution for generating large representative smart home datasets. We have built an extensible library of smart devices that facilitates the simulation of current and future smart home environments. Our tool divides the dataset generation process into three distinct phases: first design: the researcher designs the initial virtual environment by building the home, importing smart devices and creating contexts; second, simulation: the participant simulates his/her context-specific events; and third, aggregation: the researcher applies the replication algorithm to generate the final dataset. We conducted a study to assess the ease of use of our tool on the System Usability Scale (SUS). PMID:28468330
A fracture criterion for widespread cracking in thin-sheet aluminum alloys
NASA Technical Reports Server (NTRS)
Newman, J. C., Jr.; Dawicke, D. S.; Sutton, M. A.; Bigelow, C. A.
1993-01-01
An elastic-plastic finite-element analysis was used with a critical crack-tip-opening angle (CTOA) fracture criterion to model stable crack growth in thin-sheet 2024-T3 aluminum alloy panels with single and multiple-site damage (MSD) cracks. Comparisons were made between critical angles determined from the analyses and those measured with photographic methods. Calculated load against crack extension and load against crack-tip displacement on single crack specimens agreed well with test data even for large-scale plastic deformations. The analyses were also able to predict the stable tearing behavior of large lead cracks in the presence of stably tearing MSD cracks. Small MSD cracks significantly reduced the residual strength for large lead cracks.
NASA Technical Reports Server (NTRS)
Johnson, H. T.; Linley, L. J.; Mansfield, J. A.
1982-01-01
A series of large-scale JP-4 fuel pool fire tests was conducted to refine existing mathematical models of large fires. Seven tests were conducted to make chemical concentration and temperature measurements in 7.5 and 15 meter-diameter pool fires. Measurements were made at heights of 0.7, 1.4, 2.9, 5.7, 11.4, and 21.3 meters above the fires. Temperatures were measured at up to 50 locations each second during the fires. Chemistry samples were taken at up to 23 locations within the fires and analyzed for combustion chemistry and soot concentration. Temperature and combustion chemistry profiles obtained during two 7.5 meter-diameter and two 15 meter-diameter fires are included.
Flood events across the North Atlantic region - past development and future perspectives
NASA Astrophysics Data System (ADS)
Matti, Bettina; Dieppois, Bastien; Lawler, Damian; Dahlke, Helen E.; Lyon, Steve W.
2016-04-01
Flood events have a large impact on humans, both socially and economically. An increase in winter and spring flooding across much of northern Europe in recent years opened up the question of changing underlying hydro-climatic drivers of flood events. Predicting the manifestation of such changes is difficult due to the natural variability and fluctuations in northern hydrological systems caused by large-scale atmospheric circulations, especially under altered climate conditions. Improving knowledge on the complexity of these hydrological systems and their interactions with climate is essential to be able to determine drivers of flood events and to predict changes in these drivers under altered climate conditions. This is particularly true for the North Atlantic region where both physical catchment properties and large-scale atmospheric circulations have a profound influence on floods. This study explores changes in streamflow across North Atlantic region catchments. An emphasis is placed on high-flow events, namely the timing and magnitude of past flood events, and selected flood percentiles were tested for stationarity by applying a flood frequency analysis. The issue of non-stationarity of flood return periods is important when linking streamflow to large-scale atmospheric circulations. Natural fluctuations in these circulations are found to have a strong influence on the outcome causing natural variability in streamflow records. Long time series and a multi-temporal approach allows for determining drivers of floods and linking streamflow to large-scale atmospheric circulations. Exploring changes in selected hydrological signatures consistency was found across much of the North Atlantic region suggesting a shift in flow regime. The lack of an overall regional pattern suggests that how catchments respond to changes in climatic drivers is strongly influenced by their physical characteristics. A better understanding of hydrological response to climate drivers is essential for example for forecasting purposes.
Non-Born-Oppenheimer self-consistent field calculations with cubic scaling
NASA Astrophysics Data System (ADS)
Moncada, Félix; Posada, Edwin; Flores-Moreno, Roberto; Reyes, Andrés
2012-05-01
An efficient nuclear molecular orbital methodology is presented. This approach combines an auxiliary density functional theory for electrons (ADFT) and a localized Hartree product (LHP) representation for the nuclear wave function. A series of test calculations conducted on small molecules exposed that energy and geometry errors introduced by the use of ADFT and LHP approximations are small and comparable to those obtained by the use of electronic ADFT. In addition, sample calculations performed on (HF)n chains disclosed that the combined ADFT/LHP approach scales cubically with system size (n) as opposed to the quartic scaling of Hartree-Fock/LHP or DFT/LHP methods. Even for medium size molecules the improved scaling of the ADFT/LHP approach resulted in speedups of at least 5x with respect to Hartree-Fock/LHP calculations. The ADFT/LHP method opens up the possibility of studying nuclear quantum effects on large size systems that otherwise would be impractical.
Extreme reaction times determine fluctuation scaling in human color vision
NASA Astrophysics Data System (ADS)
Medina, José M.; Díaz, José A.
2016-11-01
In modern mental chronometry, human reaction time defines the time elapsed from stimulus presentation until a response occurs and represents a reference paradigm for investigating stochastic latency mechanisms in color vision. Here we examine the statistical properties of extreme reaction times and whether they support fluctuation scaling in the skewness-kurtosis plane. Reaction times were measured for visual stimuli across the cardinal directions of the color space. For all subjects, the results show that very large reaction times deviate from the right tail of reaction time distributions suggesting the existence of dragon-kings events. The results also indicate that extreme reaction times are correlated and shape fluctuation scaling over a wide range of stimulus conditions. The scaling exponent was higher for achromatic than isoluminant stimuli, suggesting distinct generative mechanisms. Our findings open a new perspective for studying failure modes in sensory-motor communications and in complex networks.
A kilobyte rewritable atomic memory
NASA Astrophysics Data System (ADS)
Kalff, Floris; Rebergen, Marnix; Fahrenfort, Nora; Girovsky, Jan; Toskovic, Ranko; Lado, Jose; FernáNdez-Rossier, JoaquíN.; Otte, Sander
The ability to manipulate individual atoms by means of scanning tunneling microscopy (STM) opens op opportunities for storage of digital data on the atomic scale. Recent achievements in this direction include data storage based on bits encoded in the charge state, the magnetic state, or the local presence of single atoms or atomic assemblies. However, a key challenge at this stage is the extension of such technologies into large-scale rewritable bit arrays. We demonstrate a digital atomic-scale memory of up to 1 kilobyte (8000 bits) using an array of individual surface vacancies in a chlorine terminated Cu(100) surface. The chlorine vacancies are found to be stable at temperatures up to 77 K. The memory, crafted using scanning tunneling microscopy at low temperature, can be read and re-written automatically by means of atomic-scale markers, and offers an areal density of 502 Terabits per square inch, outperforming state-of-the-art hard disk drives by three orders of magnitude.
Bellamy, Rob; Chilvers, Jason; Vaughan, Naomi E.
2014-01-01
Appraisals of deliberate, large-scale interventions in the earth’s climate system, known collectively as ‘geoengineering’, have largely taken the form of narrowly framed and exclusive expert analyses that prematurely ‘close down’ upon particular proposals. Here, we present the findings from the first ‘upstream’ appraisal of geoengineering to deliberately ‘open up’ to a broader diversity of framings, knowledges and future pathways. We report on the citizen strand of an innovative analytic–deliberative participatory appraisal process called Deliberative Mapping. A select but diverse group of sociodemographically representative citizens from Norfolk (United Kingdom) were engaged in a deliberative multi-criteria appraisal of geoengineering proposals relative to other options for tackling climate change, in parallel to symmetrical appraisals by diverse experts and stakeholders. Despite seeking to map divergent perspectives, a remarkably consistent view of option performance emerged across both the citizens’ and the specialists’ deliberations, where geoengineering proposals were outperformed by mitigation alternatives. PMID:25224904
Collective Framework and Performance Optimizations to Open MPI for Cray XT Platforms
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ladd, Joshua S; Gorentla Venkata, Manjunath; Shamis, Pavel
2011-01-01
The performance and scalability of collective operations plays a key role in the performance and scalability of many scientific applications. Within the Open MPI code base we have developed a general purpose hierarchical collective operations framework called Cheetah, and applied it at large scale on the Oak Ridge Leadership Computing Facility's Jaguar (OLCF) platform, obtaining better performance and scalability than the native MPI implementation. This paper discuss Cheetah's design and implementation, and optimizations to the framework for Cray XT 5 platforms. Our results show that the Cheetah's Broadcast and Barrier perform better than the native MPI implementation. For medium data,more » the Cheetah's Broadcast outperforms the native MPI implementation by 93% for 49,152 processes problem size. For small and large data, it out performs the native MPI implementation by 10% and 9%, respectively, at 24,576 processes problem size. The Cheetah's Barrier performs 10% better than the native MPI implementation for 12,288 processes problem size.« less
Bellamy, Rob; Chilvers, Jason; Vaughan, Naomi E
2016-04-01
Appraisals of deliberate, large-scale interventions in the earth's climate system, known collectively as 'geoengineering', have largely taken the form of narrowly framed and exclusive expert analyses that prematurely 'close down' upon particular proposals. Here, we present the findings from the first 'upstream' appraisal of geoengineering to deliberately 'open up' to a broader diversity of framings, knowledges and future pathways. We report on the citizen strand of an innovative analytic-deliberative participatory appraisal process called Deliberative Mapping. A select but diverse group of sociodemographically representative citizens from Norfolk (United Kingdom) were engaged in a deliberative multi-criteria appraisal of geoengineering proposals relative to other options for tackling climate change, in parallel to symmetrical appraisals by diverse experts and stakeholders. Despite seeking to map divergent perspectives, a remarkably consistent view of option performance emerged across both the citizens' and the specialists' deliberations, where geoengineering proposals were outperformed by mitigation alternatives. © The Author(s) 2014.
Halligan, Brian D.; Geiger, Joey F.; Vallejos, Andrew K.; Greene, Andrew S.; Twigger, Simon N.
2009-01-01
One of the major difficulties for many laboratories setting up proteomics programs has been obtaining and maintaining the computational infrastructure required for the analysis of the large flow of proteomics data. We describe a system that combines distributed cloud computing and open source software to allow laboratories to set up scalable virtual proteomics analysis clusters without the investment in computational hardware or software licensing fees. Additionally, the pricing structure of distributed computing providers, such as Amazon Web Services, allows laboratories or even individuals to have large-scale computational resources at their disposal at a very low cost per run. We provide detailed step by step instructions on how to implement the virtual proteomics analysis clusters as well as a list of current available preconfigured Amazon machine images containing the OMSSA and X!Tandem search algorithms and sequence databases on the Medical College of Wisconsin Proteomics Center website (http://proteomics.mcw.edu/vipdac). PMID:19358578
Integration of a neuroimaging processing pipeline into a pan-canadian computing grid
NASA Astrophysics Data System (ADS)
Lavoie-Courchesne, S.; Rioux, P.; Chouinard-Decorte, F.; Sherif, T.; Rousseau, M.-E.; Das, S.; Adalat, R.; Doyon, J.; Craddock, C.; Margulies, D.; Chu, C.; Lyttelton, O.; Evans, A. C.; Bellec, P.
2012-02-01
The ethos of the neuroimaging field is quickly moving towards the open sharing of resources, including both imaging databases and processing tools. As a neuroimaging database represents a large volume of datasets and as neuroimaging processing pipelines are composed of heterogeneous, computationally intensive tools, such open sharing raises specific computational challenges. This motivates the design of novel dedicated computing infrastructures. This paper describes an interface between PSOM, a code-oriented pipeline development framework, and CBRAIN, a web-oriented platform for grid computing. This interface was used to integrate a PSOM-compliant pipeline for preprocessing of structural and functional magnetic resonance imaging into CBRAIN. We further tested the capacity of our infrastructure to handle a real large-scale project. A neuroimaging database including close to 1000 subjects was preprocessed using our interface and publicly released to help the participants of the ADHD-200 international competition. This successful experiment demonstrated that our integrated grid-computing platform is a powerful solution for high-throughput pipeline analysis in the field of neuroimaging.
NASA Technical Reports Server (NTRS)
Jain, Raj K.; Flood, Dennis J.
1990-01-01
Excellent radiation resistance of indium phosphide solar cells makes them a promising candidate for space power applications, but the present high cost of starting substrates may inhibit their large scale use. Thin film indium phosphide cells grown on Si or GaAs substrates have exhibited low efficiencies, because of the generation and propagation of large number of dislocations. Dislocation densities were calculated and its influence on the open circuit voltage, short circuit current, and efficiency of heteroepitaxial indium phosphide cells was studied using the PC-1D. Dislocations act as predominant recombination centers and are required to be controlled by proper transition layers and improved growth techniques. It is shown that heteroepitaxial grown cells could achieve efficiencies in excess of 18 percent AMO by controlling the number of dislocations. The effect of emitter thickness and surface recombination velocity on the cell performance parameters vs. dislocation density is also studied.
Two-stage opening of the Dover Strait and the origin of island Britain
Gupta, Sanjeev; Collier, Jenny S.; Garcia-Moreno, David; Oggioni, Francesca; Trentesaux, Alain; Vanneste, Kris; De Batist, Marc; Camelbeeck, Thierry; Potter, Graeme; Van Vliet-Lanoë, Brigitte; Arthur, John C. R.
2017-01-01
Late Quaternary separation of Britain from mainland Europe is considered to be a consequence of spillover of a large proglacial lake in the Southern North Sea basin. Lake spillover is inferred to have caused breaching of a rock ridge at the Dover Strait, although this hypothesis remains untested. Here we show that opening of the Strait involved at least two major episodes of erosion. Sub-bottom records reveal a remarkable set of sediment-infilled depressions that are deeply incised into bedrock that we interpret as giant plunge pools. These support a model of initial erosion of the Dover Strait by lake overspill, plunge pool erosion by waterfalls and subsequent dam breaching. Cross-cutting of these landforms by a prominent bedrock-eroded valley that is characterized by features associated with catastrophic flooding indicates final breaching of the Strait by high-magnitude flows. These events set-up conditions for island Britain during sea-level highstands and caused large-scale re-routing of NW European drainage. PMID:28375202
Two-stage opening of the Dover Strait and the origin of island Britain
NASA Astrophysics Data System (ADS)
Gupta, Sanjeev; Collier, Jenny S.; Garcia-Moreno, David; Oggioni, Francesca; Trentesaux, Alain; Vanneste, Kris; de Batist, Marc; Camelbeeck, Thierry; Potter, Graeme; van Vliet-Lanoë, Brigitte; Arthur, John C. R.
2017-04-01
Late Quaternary separation of Britain from mainland Europe is considered to be a consequence of spillover of a large proglacial lake in the Southern North Sea basin. Lake spillover is inferred to have caused breaching of a rock ridge at the Dover Strait, although this hypothesis remains untested. Here we show that opening of the Strait involved at least two major episodes of erosion. Sub-bottom records reveal a remarkable set of sediment-infilled depressions that are deeply incised into bedrock that we interpret as giant plunge pools. These support a model of initial erosion of the Dover Strait by lake overspill, plunge pool erosion by waterfalls and subsequent dam breaching. Cross-cutting of these landforms by a prominent bedrock-eroded valley that is characterized by features associated with catastrophic flooding indicates final breaching of the Strait by high-magnitude flows. These events set-up conditions for island Britain during sea-level highstands and caused large-scale re-routing of NW European drainage.
Halligan, Brian D; Geiger, Joey F; Vallejos, Andrew K; Greene, Andrew S; Twigger, Simon N
2009-06-01
One of the major difficulties for many laboratories setting up proteomics programs has been obtaining and maintaining the computational infrastructure required for the analysis of the large flow of proteomics data. We describe a system that combines distributed cloud computing and open source software to allow laboratories to set up scalable virtual proteomics analysis clusters without the investment in computational hardware or software licensing fees. Additionally, the pricing structure of distributed computing providers, such as Amazon Web Services, allows laboratories or even individuals to have large-scale computational resources at their disposal at a very low cost per run. We provide detailed step-by-step instructions on how to implement the virtual proteomics analysis clusters as well as a list of current available preconfigured Amazon machine images containing the OMSSA and X!Tandem search algorithms and sequence databases on the Medical College of Wisconsin Proteomics Center Web site ( http://proteomics.mcw.edu/vipdac ).
NASA Technical Reports Server (NTRS)
Stothers, Richard B.
1991-01-01
This study presents the results of 14 tests for the presence of convective overshooting in large convecting stellar cores for stars with masses of 4-17 solar masses which are members of detached close binary systems and of open clusters in the Galaxy. A large body of theoretical and observational data is scrutinized and subjected to averaging in order to minimize accidental and systematic errors. A conservative upper limit of d/HP less than 0.4 is found from at least four tests, as well as a tighter upper limit of d/HP less than 0.2 from one good test that is subject to only mild restrictions and is based on the maximum observed effective temperature of evolved blue supergiants. It is concluded that any current uncertainty about the distance scale for these stars is unimportant in conducting the present tests for convective core overshooting. The correct effective temperature scale for the B0.5-B2 stars is almost certainly close to one of the proposed hot scales.
Time-Resolved Small-Angle X-ray Scattering Reveals Millisecond Transitions of a DNA Origami Switch.
Bruetzel, Linda K; Walker, Philipp U; Gerling, Thomas; Dietz, Hendrik; Lipfert, Jan
2018-04-11
Self-assembled DNA structures enable creation of specific shapes at the nanometer-micrometer scale with molecular resolution. The construction of functional DNA assemblies will likely require dynamic structures that can undergo controllable conformational changes. DNA devices based on shape complementary stacking interactions have been demonstrated to undergo reversible conformational changes triggered by changes in ionic environment or temperature. An experimentally unexplored aspect is how quickly conformational transitions of large synthetic DNA origami structures can actually occur. Here, we use time-resolved small-angle X-ray scattering to monitor large-scale conformational transitions of a two-state DNA origami switch in free solution. We show that the DNA device switches from its open to its closed conformation upon addition of MgCl 2 in milliseconds, which is close to the theoretical diffusive speed limit. In contrast, measurements of the dimerization of DNA origami bricks reveal much slower and concentration-dependent assembly kinetics. DNA brick dimerization occurs on a time scale of minutes to hours suggesting that the kinetics depend on local concentration and molecular alignment.
Wittman; Tyson; Kirkman; Dell'Antonio; Bernstein
2000-05-11
Most of the matter in the Universe is not luminous, and can be observed only through its gravitational influence on the appearance of luminous matter. Weak gravitational lensing is a technique that uses the distortions of the images of distant galaxies as a tracer of dark matter: such distortions are induced as the light passes through large-scale distributions of dark matter in the foreground. The patterns of the induced distortions reflect the density of mass along the line of sight and its distribution, and the resulting 'cosmic shear' can be used to distinguish between alternative cosmologies. But previous attempts to measure this effect have been inconclusive. Here we report the detection of cosmic shear on angular scales of up to half a degree using 145,000 galaxies and along three separate lines of sight. We find that the dark matter is distributed in a manner consistent with either an open universe, or a flat universe that is dominated by a cosmological constant. Our results are inconsistent with the standard cold-dark-matter model.
Coal resources, reserves and peak coal production in the United States
Milici, Robert C.; Flores, Romeo M.; Stricker, Gary D.
2013-01-01
In spite of its large endowment of coal resources, recent studies have indicated that United States coal production is destined to reach a maximum and begin an irreversible decline sometime during the middle of the current century. However, studies and assessments illustrating coal reserve data essential for making accurate forecasts of United States coal production have not been compiled on a national basis. As a result, there is a great deal of uncertainty in the accuracy of the production forecasts. A very large percentage of the coal mined in the United States comes from a few large-scale mines (mega-mines) in the Powder River Basin of Wyoming and Montana. Reported reserves at these mines do not account for future potential reserves or for future development of technology that may make coal classified currently as resources into reserves in the future. In order to maintain United States coal production at or near current levels for an extended period of time, existing mines will eventually have to increase their recoverable reserves and/or new large-scale mines will have to be opened elsewhere. Accordingly, in order to facilitate energy planning for the United States, this paper suggests that probabilistic assessments of the remaining coal reserves in the country would improve long range forecasts of coal production. As it is in United States coal assessment projects currently being conducted, a major priority of probabilistic assessments would be to identify the numbers and sizes of remaining large blocks of coal capable of supporting large-scale mining operations for extended periods of time and to conduct economic evaluations of those resources.
NASA Technical Reports Server (NTRS)
Srinivasan, K. V.
1986-01-01
The design and development of a large diameter high pressure quick acting propulsion valve and valve actuator is described. The valve is the heart of a major test facility dedicated to conducting full scale performance tests of aircraft landing systems. The valve opens in less than 300 milliseconds releasing a 46-centimeter- (18-in.-) diameter water jet and closes in 300 milliseconds. The four main components of the valve, i.e., valve body, safety shutter, high speed shutter, and pneumatic-hydraulic actuator, are discussed. This valve is unique and may have other aerospace and industrial applications.
NASA Technical Reports Server (NTRS)
Srinivasan, K. V.
1986-01-01
This paper describes the design and development of a large diameter high pressure quick acting propulsion valve and valve actuator. The valve is the heart of a major test facility dedicated to conducting full scale performance tests of aircraft landing gear systems. The valve opens in less than 300 milliseconds releasing a 46 cm (18 in) diameter water jet and closes in 300 milliseconds. The four main components of the valve, i.e., valve body, safety shutter, high speed shutter, and pneumatic-hydraulic actuator, are discussed. This valve is unique and may have other aerospace and industrial applications.
Cumulative effects of wildfires on forest dynamics in the eastern Cascade Mountains, USA.
Reilly, Matthew J; Elia, Mario; Spies, Thomas A; Gregory, Matthew J; Sanesi, Giovanni; Lafortezza, Raffaele
2018-03-01
Wildfires pose a unique challenge to conservation in fire-prone regions, yet few studies quantify the cumulative effects of wildfires on forest dynamics (i.e., changes in structural conditions) across landscape and regional scales. We assessed the contribution of wildfire to forest dynamics in the eastern Cascade Mountains, USA from 1985 to 2010 using imputed maps of forest structure (i.e., tree size and canopy cover) and remotely sensed burn severity maps. We addressed three questions: (1) How do dynamics differ between the region as a whole and the unburned portion of the region? (2) How do dynamics vary among vegetation zones differing in biophysical setting and historical fire frequency? (3) How have forest structural conditions changed in a network of late successional reserves (LSRs)? Wildfires affected 10% of forests in the region, but the cumulative effects at this scale were primarily slight losses of closed-canopy conditions and slight gains in open-canopy conditions. In the unburned portion of the region (the remaining 90%), closed-canopy conditions primarily increased despite other concurrent disturbances (e.g., harvest, insects). Although the effects of fire were largely dampened at the regional scale, landscape scale dynamics were far more variable. The warm ponderosa pine and cool mixed conifer zones experienced less fire than the region as a whole despite experiencing the most frequent fire historically. Open-canopy conditions increased slightly in the mixed conifer zone, but declined across the ponderosa pine zone even with wildfires. Wildfires burned 30% of the cold subalpine zone, which experienced the greatest increase in open-canopy conditions and losses of closed-canopy conditions. LSRs were more prone to wildfire than the region as a whole, and experienced slight declines in late seral conditions. Despite losses of late seral conditions, wildfires contributed to some conservation objectives by creating open habitats (e.g., sparse early seral and woodland conditions) that otherwise generally decreased in unburned landscapes despite management efforts to increase landscape diversity. This study demonstrates the potential for wildfires to contribute to regional scale conservation objectives, but implications for management and biodiversity at landscape scales vary geographically among biophysical settings, and are contingent upon historical dynamics and individual species habitat preferences. © 2017 by the Ecological Society of America.
Basin scale permeability and thermal evolution of a magmatic hydrothermal system
NASA Astrophysics Data System (ADS)
Taron, J.; Hickman, S. H.; Ingebritsen, S.; Williams, C.
2013-12-01
Large-scale hydrothermal systems are potentially valuable energy resources and are of general scientific interest due to extreme conditions of stress, temperature, and reactive chemistry that can act to modify crustal rheology and composition. With many proposed sites for Enhanced Geothermal Systems (EGS) located on the margins of large-scale hydrothermal systems, understanding the temporal evolution of these systems contributes to site selection, characterization and design of EGS. This understanding is also needed to address the long-term sustainability of EGS once they are created. Many important insights into heat and mass transfer within natural hydrothermal systems can be obtained through hydrothermal modeling assuming that stress and permeability structure do not evolve over time. However, this is not fully representative of natural systems, where the effects of thermo-elastic stress changes, chemical fluid-rock interactions, and rock failure on fluid flow and thermal evolution can be significant. The quantitative importance of an evolving permeability field within the overall behavior of a large-scale hydrothermal system is somewhat untested, and providing such a parametric understanding is one of the goals of this study. We explore the thermal evolution of a sedimentary basin hydrothermal system following the emplacement of a magma body. The Salton Sea geothermal field and its associated magmatic system in southern California is utilized as a general backdrop to define the initial state. Working within the general framework of the open-source scientific computing initiative OpenGeoSys (www.opengeosys.org), we introduce full treatment of thermodynamic properties at the extreme conditions following magma emplacement. This treatment utilizes a combination of standard Galerkin and control-volume finite elements to balance fluid mass, mechanical deformation, and thermal energy with consideration of local thermal non-equilibrium (LTNE) between fluids and solids. Permeability is allowed to evolve under several constitutive models tailored to both porous media and fractures, considering the influence of both mechanical stress and diagenesis. In this first analysis, a relatively simple mechanical model is used; complexity will be added incrementally to represent specific characteristics of the Salton Sea hydrothermal field.
On the Fluctuating Component of the Sun's Large-Scale Magnetic Field
NASA Astrophysics Data System (ADS)
Wang, Y.-M.; Sheeley, N. R., Jr.
2003-06-01
The Sun's large-scale magnetic field and its proxies are known to undergo substantial variations on timescales much less than a solar cycle but longer than a rotation period. Examples of such variations include the double activity maximum inferred by Gnevyshev, the large peaks in the interplanetary field strength observed in 1982 and 1991, and the 1.3-1.4 yr periodicities detected over limited time intervals in solar wind speed and geomagnetic activity. We consider the question of the extent to which these variations are stochastic in nature. For this purpose, we simulate the evolution of the Sun's equatorial dipole strength and total open flux under the assumption that the active region sources (BMRs) are distributed randomly in longitude. The results are then interpreted with the help of a simple random walk model including dissipation. We find that the equatorial dipole and open flux generally exhibit multiple peaks during each 11 yr cycle, with the highest peak as likely to occur during the declining phase as at sunspot maximum. The widths of the peaks are determined by the timescale τ~1 yr for the equatorial dipole to decay through the combined action of meridional flow, differential rotation, and supergranular diffusion. The amplitudes of the fluctuations depend on the strengths and longitudinal phase relations of the BMRs, as well as on the relative rates of flux emergence and decay. We conclude that stochastic processes provide a viable explanation for the ``Gnevyshev gaps'' and for the existence of quasi periodicities in the range ~1-3 yr.
Geophysical Potential for Wind Energy over the Open Oceans
NASA Astrophysics Data System (ADS)
Possner, A.; Caldeira, K.
2017-12-01
Wind turbines continuously remove kinetic energy from the lower troposphere thereby reducing the wind speed near hub height. The rate of electricity generation in large wind farms containing multiple wind arrays is therefore constrained by the rate of kinetic energy replenishment from the atmosphere above. In particular, this study focuses on the maximum sustained transport of kinetic energy through the troposphere to the lowest hundreds of meters above the surface. In recent years, a growing body of research argues that the rate of generated power is limited to around 1.5 Wm-2 within large wind farms. However, in this study we demonstrate that considerably higher power generation rates may be sustainable over some open ocean areas in giant wind farms. We find that in the North Atlantic maximum extraction rates of up to 6.7 Wm-2 may be sustained by the atmosphere in the annual mean over giant wind farm areas approaching the size of Greenland. In contrast, only a third of this rate is sustained on land for areas of equivalent size. Our simulations indicate a fundamental difference in response of the troposphere and its vertical kinetic energy flux to giant near-surface wind farms. We find that the surface heat flux from the oceans to the atmosphere may play an important role in creating regions where large sustained rates of downward transport of kinetic energy and thus rates of kinetic energy extraction may be geophysically possible. While no commercial-scale deep-water wind turbines yet exist, our results suggest that such technologies, if they became technically and economically feasible, could potentially provide civilization-scale power.
Hollow microcarriers for large-scale expansion of anchorage-dependent cells in a stirred bioreactor.
YekrangSafakar, Ashkan; Acun, Aylin; Choi, Jin-Woo; Song, Edward; Zorlutuna, Pinar; Park, Kidong
2018-03-26
With recent advances in biotechnology, mammalian cells are used in biopharmaceutical industries to produce valuable protein therapeutics and investigated as effective therapeutic agents to permanently degenerative diseases in cell based therapy. In these exciting and actively expanding fields, a reliable, efficient, and affordable platform to culture mammalian cells on a large scale is one of the most vital necessities. To produce and maintain a very large population of anchorage-dependent cells, a microcarrier-based stirred tank bioreactor is commonly used. In this approach, the cells are exposed to harmful hydrodynamic shear stress in the bioreactor and the mass transfer rates of nutrients and gases in the bioreactor are often kept below an optimal level to prevent cellular damages from the shear stress. In this paper, a hollow microcarrier (HMC) is presented as a novel solution to protect cells from shear stress in stirred bioreactors, while ensuring sufficient and uniform mass transfer rate of gases and nutrients. HMC is a hollow microsphere and cells are cultured on its inner surface to be protected, while openings on the HMC provide sufficient exchange of media inside the HMC. As a proof of concept, we demonstrated the expansion of fibroblasts, NIH/3T3 and the expansion and cardiac differentiation of human induced pluripotent stem cells, along with detailed numerical analysis. We believe that the developed HMC can be a practical solution to enable large-scale expansion of shear-sensitive anchorage-dependent cells in an industrial scale with stirred bioreactors. © 2018 Wiley Periodicals, Inc.
Dynamics of passive and active particles in the cell nucleus.
Hameed, Feroz M; Rao, Madan; Shivashankar, G V
2012-01-01
Inspite of being embedded in a dense meshwork of nuclear chromatin, gene loci and large nuclear components are highly dynamic at 37°C. To understand this apparent unfettered movement in an overdense environment, we study the dynamics of a passive micron size bead in live cell nuclei at two different temperatures (25 and 37°C) with and without external force. In the absence of a force, the beads are caged over large time scales. On application of a threshold uniaxial force (about 10(2) pN), the passive beads appear to hop between cages; this large scale movement is absent upon ATP-depletion, inhibition of chromatin remodeling enzymes and RNAi of lamin B1 proteins. Our results suggest that the nucleus behaves like an active solid with a finite yield stress when probed at a micron scale. Spatial analysis of histone fluorescence anisotropy (a measure of local chromatin compaction, defined as the volume fraction of tightly bound chromatin) shows that the bead movement correlates with regions of low chromatin compaction. This suggests that the physical mechanism of the observed yielding is the active opening of free-volume in the nuclear solid via chromatin remodeling. Enriched transcription sites at 25°C also show caging in the absence of the applied force and directed movement beyond a yield stress, in striking contrast with the large scale movement of transcription loci at 37°C in the absence of a force. This suggests that at physiological temperatures, the loci behave as active particles which remodel the nuclear mesh and reduce the local yield stress.
Klukas, Christian; Chen, Dijun; Pape, Jean-Michel
2014-01-01
High-throughput phenotyping is emerging as an important technology to dissect phenotypic components in plants. Efficient image processing and feature extraction are prerequisites to quantify plant growth and performance based on phenotypic traits. Issues include data management, image analysis, and result visualization of large-scale phenotypic data sets. Here, we present Integrated Analysis Platform (IAP), an open-source framework for high-throughput plant phenotyping. IAP provides user-friendly interfaces, and its core functions are highly adaptable. Our system supports image data transfer from different acquisition environments and large-scale image analysis for different plant species based on real-time imaging data obtained from different spectra. Due to the huge amount of data to manage, we utilized a common data structure for efficient storage and organization of data for both input data and result data. We implemented a block-based method for automated image processing to extract a representative list of plant phenotypic traits. We also provide tools for build-in data plotting and result export. For validation of IAP, we performed an example experiment that contains 33 maize (Zea mays ‘Fernandez’) plants, which were grown for 9 weeks in an automated greenhouse with nondestructive imaging. Subsequently, the image data were subjected to automated analysis with the maize pipeline implemented in our system. We found that the computed digital volume and number of leaves correlate with our manually measured data in high accuracy up to 0.98 and 0.95, respectively. In summary, IAP provides a multiple set of functionalities for import/export, management, and automated analysis of high-throughput plant phenotyping data, and its analysis results are highly reliable. PMID:24760818
Lucas, Jessica S.; Loeb, Susan C.; Jodice, Patrick G.R.
2015-01-01
Although several studies have described roost use by Rafinesque's big-eared bats (Corynorhinus rafinesquii), few studies have examined roost selection. We examined roost use and selection by Rafinesque's big-eared bat at the tree, stand, and landscape scales during the maternity season in pristine old-growth habitat in the Coastal Plain of South Carolina. We located 43 roosts (14 maternity, 29 solitary) through cavity searches and radio-telemetry. Maternity colonies and solitary individuals selected roosts based on similar characteristics. The best model explaining roost selection by all bats included tree and stand characteristics; landscape factors had little influence on roost use. Bats selected large diameter trees in areas with a high density of trees with cavities in the surrounding area. Most roosts (67.4%) were in water tupelo (Nyssa aquatica) in semi-permanently flooded and saturated areas. Half of maternity roost cavities had upper bole openings whereas only 25.8% of solitary roosts had upper bole openings. Bats that roosted with maternity groups stayed in roosts for significantly shorter periods of time (1.3 ± 0.1 days) and used significantly more roost trees (5.0 ± 0.6 roosts) than adult males (3.8 ± 1.10 days, 2.3 ± 0.4 roosts, respectively). Maternity colony use of cavities with upper bole openings and shorter residency times suggest that predator avoidance may have been an important factor governing roosting behavior of maternity colonies in this area. Our results suggest that retention of large diameter, hollow trees in wetland areas will benefit Rafinesque's big-eared bat individuals and maternity colonies in this area.
Remote sensing and field test capabilities at U.S. Army Dugway Proving Ground
NASA Astrophysics Data System (ADS)
Pearson, James T.; Herron, Joshua P.; Marshall, Martin S.
2011-11-01
U.S. Army Dugway Proving Ground (DPG) is a Major Range and Test Facility Base (MRTFB) with the mission of testing chemical and biological defense systems and materials. DPG facilities include state-of-the-art laboratories, extensive test grids, controlled environment calibration facilities, and a variety of referee instruments for required test measurements. Among these referee instruments, DPG has built up a significant remote sensing capability for both chemical and biological detection. Technologies employed for remote sensing include FTIR spectroscopy, UV spectroscopy, Raman-shifted eye-safe lidar, and other elastic backscatter lidar systems. These systems provide referee data for bio-simulants, chemical simulants, toxic industrial chemicals (TICs), and toxic industrial materials (TIMs). In order to realize a successful large scale open-air test, each type of system requires calibration and characterization. DPG has developed specific calibration facilities to meet this need. These facilities are the Joint Ambient Breeze Tunnel (JABT), and the Active Standoff Chamber (ASC). The JABT and ASC are open ended controlled environment tunnels. Each includes validation instrumentation to characterize simulants that are disseminated. Standoff systems are positioned at typical field test distances to measure characterized simulants within the tunnel. Data from different types of systems can be easily correlated using this method, making later open air test results more meaningful. DPG has a variety of large scale test grids available for field tests. After and during testing, data from the various referee instruments is provided in a visual format to more easily draw conclusions on the results. This presentation provides an overview of DPG's standoff testing facilities and capabilities, as well as example data from different test scenarios.
Remote sensing and field test capabilities at U.S. Army Dugway Proving Ground
NASA Astrophysics Data System (ADS)
Pearson, James T.; Herron, Joshua P.; Marshall, Martin S.
2012-05-01
U.S. Army Dugway Proving Ground (DPG) is a Major Range and Test Facility Base (MRTFB) with the mission of testing chemical and biological defense systems and materials. DPG facilities include state-of-the-art laboratories, extensive test grids, controlled environment calibration facilities, and a variety of referee instruments for required test measurements. Among these referee instruments, DPG has built up a significant remote sensing capability for both chemical and biological detection. Technologies employed for remote sensing include FTIR spectroscopy, UV spectroscopy, Raman-shifted eye-safe lidar, and other elastic backscatter lidar systems. These systems provide referee data for bio-simulants, chemical simulants, toxic industrial chemicals (TICs), and toxic industrial materials (TIMs). In order to realize a successful large scale open-air test, each type of system requires calibration and characterization. DPG has developed specific calibration facilities to meet this need. These facilities are the Joint Ambient Breeze Tunnel (JABT), and the Active Standoff Chamber (ASC). The JABT and ASC are open ended controlled environment tunnels. Each includes validation instrumentation to characterize simulants that are disseminated. Standoff systems are positioned at typical field test distances to measure characterized simulants within the tunnel. Data from different types of systems can be easily correlated using this method, making later open air test results more meaningful. DPG has a variety of large scale test grids available for field tests. After and during testing, data from the various referee instruments is provided in a visual format to more easily draw conclusions on the results. This presentation provides an overview of DPG's standoff testing facilities and capabilities, as well as example data from different test scenarios.
Evaluating Model-Driven Development for large-scale EHRs through the openEHR approach.
Christensen, Bente; Ellingsen, Gunnar
2016-05-01
In healthcare, the openEHR standard is a promising Model-Driven Development (MDD) approach for electronic healthcare records. This paper aims to identify key socio-technical challenges when the openEHR approach is put to use in Norwegian hospitals. More specifically, key fundamental assumptions are investigated empirically. These assumptions promise a clear separation of technical and domain concerns, users being in control of the modelling process, and widespread user commitment. Finally, these assumptions promise an easy way to model and map complex organizations. This longitudinal case study is based on an interpretive approach, whereby data were gathered through 440h of participant observation, 22 semi-structured interviews and extensive document studies over 4 years. The separation of clinical and technical concerns seemed to be aspirational, because both designing the technical system and modelling the domain required technical and clinical competence. Hence developers and clinicians found themselves working together in both arenas. User control and user commitment seemed not to apply in large-scale projects, as modelling the domain turned out to be too complicated and hence to appeal only to especially interested users worldwide, not the local end-users. Modelling proved to be a complex standardization process that shaped both the actual modelling and healthcare practice itself. A broad assemblage of contributors seems to be needed for developing an archetype-based system, in which roles, responsibilities and contributions cannot be clearly defined and delimited. The way MDD occurs has implications for medical practice per se in the form of the need to standardize practices to ensure that medical concepts are uniform across practices. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Korchagova, V. N.; Kraposhin, M. V.; Marchevsky, I. K.; Smirnova, E. V.
2017-11-01
A droplet impact on a deep pool can induce macro-scale or micro-scale effects like a crown splash, a high-speed jet, formation of secondary droplets or thin liquid films, etc. It depends on the diameter and velocity of the droplet, liquid properties, effects of external forces and other factors that a ratio of dimensionless criteria can account for. In the present research, we considered the droplet and the pool consist of the same viscous incompressible liquid. We took surface tension into account but neglected gravity forces. We used two open-source codes (OpenFOAM and Gerris) for our computations. We review the possibility of using these codes for simulation of processes in free-surface flows that may take place after a droplet impact on the pool. Both codes simulated several modes of droplet impact. We estimated the effect of liquid properties with respect to the Reynolds number and Weber number. Numerical simulation enabled us to find boundaries between different modes of droplet impact on a deep pool and to plot corresponding mode maps. The ratio of liquid density to that of the surrounding gas induces several changes in mode maps. Increasing this density ratio suppresses the crown splash.
Salerno, Aurelio; Domingo, Concepción
2014-09-01
Open-pore biodegradable microparticles are object of considerable interest for biomedical applications, particularly as cell and drug delivery carriers in tissue engineering and health care treatments. Furthermore, the engineering of microparticles with well definite size distribution and pore architecture by bio-safe fabrication routes is crucial to avoid the use of toxic compounds potentially harmful to cells and biological tissues. To achieve this important issue, in the present study a straightforward and bio-safe approach for fabricating porous biodegradable microparticles with controlled morphological and structural features down to the nanometer scale is developed. In particular, ethyl lactate is used as a non-toxic solvent for polycaprolactone particles fabrication via a thermal induced phase separation technique. The used approach allows achieving open-pore particles with mean particle size in the 150-250 μm range and a 3.5-7.9 m(2)/g specific surface area. Finally, the combination of thermal induced phase separation and porogen leaching techniques is employed for the first time to obtain multi-scaled porous microparticles with large external and internal pore sizes and potential improved characteristics for cell culture and tissue engineering. Samples were characterized to assess their thermal properties, morphology and crystalline structure features and textural properties. Copyright © 2014 Elsevier B.V. All rights reserved.
Metallicity calibrations for dwarf stars and giants in the Geneva photometric system
NASA Astrophysics Data System (ADS)
Netopil, Martin
2017-08-01
We use the most homogeneous Geneva seven-colour photometric system to derive new metallicity calibrations for early A- to K-type stars that cover both, dwarf stars and giants. The calibrations are based on several spectroscopic data sets that were merged to a common scale, and we applied them to open cluster data to obtain an additional proof of the metallicity scale and accuracy. In total, metallicities of 54 open clusters are presented. The accuracy of the calibrations for single stars is in general below 0.1 dex, but for the open cluster sample with mean values based on several stars we find a much better precision, a scatter as low as about 0.03 dex. Furthermore, we combine the new results with another comprehensive photometric data set to present a catalogue of mean metallicities for more than 3000 F- and G-type dwarf stars with σ ˜ 0.06 dex. The list was extended by more than 1200 hotter stars up to about 8500 K (or spectral type A3) by taking advantage of their almost reddening free characteristic in the new Geneva metallicity calibrations. These two large samples are well suited as primary or secondary calibrators of other data, and we already identified about 20 spectroscopic data sets that show offsets up to about 0.4 dex.
Large-scale deep learning for robotically gathered imagery for science
NASA Astrophysics Data System (ADS)
Skinner, K.; Johnson-Roberson, M.; Li, J.; Iscar, E.
2016-12-01
With the explosion of computing power, the intelligence and capability of mobile robotics has dramatically increased over the last two decades. Today, we can deploy autonomous robots to achieve observations in a variety of environments ripe for scientific exploration. These platforms are capable of gathering a volume of data previously unimaginable. Additionally, optical cameras, driven by mobile phones and consumer photography, have rapidly improved in size, power consumption, and quality making their deployment cheaper and easier. Finally, in parallel we have seen the rise of large-scale machine learning approaches, particularly deep neural networks (DNNs), increasing the quality of the semantic understanding that can be automatically extracted from optical imagery. In concert this enables new science using a combination of machine learning and robotics. This work will discuss the application of new low-cost high-performance computing approaches and the associated software frameworks to enable scientists to rapidly extract useful science data from millions of robotically gathered images. The automated analysis of imagery on this scale opens up new avenues of inquiry unavailable using more traditional manual or semi-automated approaches. We will use a large archive of millions of benthic images gathered with an autonomous underwater vehicle to demonstrate how these tools enable new scientific questions to be posed.
Rodríguez-Gómez, Francisco; Romero-Gil, Verónica; Arroyo-López, Francisco N; Roldán-Reyes, Juan C; Torres-Gallardo, Rosa; Bautista-Gallego, Joaquín; García-García, Pedro; Garrido-Fernández, Antonio
2017-01-01
This work studies the inoculation conditions for allowing the survival/predominance of a potential probiotic strain ( Lactobacillus pentosus TOMC-LAB2) when used as a starter culture in large-scale fermentations of green Spanish-style olives. The study was performed in two successive seasons (2011/2012 and 2012/2013), using about 150 tons of olives. Inoculation immediately after brining (to prevent wild initial microbiota growth) followed by re-inoculation 24 h later (to improve competitiveness) was essential for inoculum predominance. Processing early in the season (September) showed a favorable effect on fermentation and strain predominance on olives (particularly when using acidified brines containing 25 L HCl/vessel) but caused the disappearance of the target strain from both brines and olives during the storage phase. On the contrary, processing in October slightly reduced the target strain predominance on olives (70-90%) but allowed longer survival. The type of inoculum used (laboratory vs. industry pre-adapted) never had significant effects. Thus, this investigation discloses key issues for the survival and predominance of starter cultures in large-scale industrial fermentations of green Spanish-style olives. Results can be of interest for producing probiotic table olives and open new research challenges on the causes of inoculum vanishing during the storage phase.
Experimental evidence for herbivore limitation of the treeline.
Speed, James D M; Austrheim, Gunnar; Hester, Alison J; Mysterud, Atle
2010-11-01
The treeline ecotone divides forest from open alpine or arctic vegetation states. Treelines are generally perceived to be temperature limited. The role of herbivores in limiting the treeline is more controversial, as experimental evidence from relevant large scales is lacking. Here we quantify the impact of different experimentally controlled herbivore densities on the recruitment and survival of birch Betula pubescens tortuosa along an altitudinal gradient in the mountains of southern Norway. After eight years of summer grazing in large-scale enclosures at densities of 0, 25, and 80 sheep/km2, birch recruited within the whole altitudinal range of ungrazed enclosures, but recruitment was rarer in enclosures with low-density sheep and was largely limited to within the treeline in enclosures with high-density sheep. In contrast, the distribution of saplings (birch older than the experiment) did not differ between grazing treatments, suggesting that grazing sheep primarily limit the establishment of new tree recruits rather than decrease the survival of existing individuals. This study provides direct experimental evidence that herbivores can limit the treeline below its potential at the landscape scale and even at low herbivore densities in this climatic zone. Land use changes should thus be considered in addition to climatic changes as potential drivers of ecotone shifts.
Agartan, Tuba I
2015-10-01
Explaining policy change has been one of the major concerns of the health care politics and policy development literature. This article aims to explain the specific dynamics of large-scale reforms introduced within the framework of the Health Transformation Program in Turkey. It argues that confluence of the three streams - problem, policy, and politics - with the exceptional political will of the Justice and Development Party's (JDP) leaders opened up a window of opportunity for a large-scale policy change. The article also underscores the contribution of recent ideational perspectives that help explain "why" political actors in Turkey would focus on health care reform, given that there are a number of issues waiting to be addressed in the policy agenda. Examining how political actors framed problems and policies deepens our understanding of the content of the reform initiatives as well as the construction of the need to reform. The article builds on the insights of both the ideational and institutionalist perspectives when it argues that the interests, aspirations, and fears of the JDP, alongside the peculiar characteristics of the institutional context, have shaped its priorities and determination to carry out this reform initiative. Copyright © 2015 by Duke University Press.
Azad, Ariful; Ouzounis, Christos A; Kyrpides, Nikos C; Buluç, Aydin
2018-01-01
Abstract Biological networks capture structural or functional properties of relevant entities such as molecules, proteins or genes. Characteristic examples are gene expression networks or protein–protein interaction networks, which hold information about functional affinities or structural similarities. Such networks have been expanding in size due to increasing scale and abundance of biological data. While various clustering algorithms have been proposed to find highly connected regions, Markov Clustering (MCL) has been one of the most successful approaches to cluster sequence similarity or expression networks. Despite its popularity, MCL’s scalability to cluster large datasets still remains a bottleneck due to high running times and memory demands. Here, we present High-performance MCL (HipMCL), a parallel implementation of the original MCL algorithm that can run on distributed-memory computers. We show that HipMCL can efficiently utilize 2000 compute nodes and cluster a network of ∼70 million nodes with ∼68 billion edges in ∼2.4 h. By exploiting distributed-memory environments, HipMCL clusters large-scale networks several orders of magnitude faster than MCL and enables clustering of even bigger networks. HipMCL is based on MPI and OpenMP and is freely available under a modified BSD license. PMID:29315405
CERN data services for LHC computing
NASA Astrophysics Data System (ADS)
Espinal, X.; Bocchi, E.; Chan, B.; Fiorot, A.; Iven, J.; Lo Presti, G.; Lopez, J.; Gonzalez, H.; Lamanna, M.; Mascetti, L.; Moscicki, J.; Pace, A.; Peters, A.; Ponce, S.; Rousseau, H.; van der Ster, D.
2017-10-01
Dependability, resilience, adaptability and efficiency. Growing requirements require tailoring storage services and novel solutions. Unprecedented volumes of data coming from the broad number of experiments at CERN need to be quickly available in a highly scalable way for large-scale processing and data distribution while in parallel they are routed to tape for long-term archival. These activities are critical for the success of HEP experiments. Nowadays we operate at high incoming throughput (14GB/s during 2015 LHC Pb-Pb run and 11PB in July 2016) and with concurrent complex production work-loads. In parallel our systems provide the platform for the continuous user and experiment driven work-loads for large-scale data analysis, including end-user access and sharing. The storage services at CERN cover the needs of our community: EOS and CASTOR as a large-scale storage; CERNBox for end-user access and sharing; Ceph as data back-end for the CERN OpenStack infrastructure, NFS services and S3 functionality; AFS for legacy distributed-file-system services. In this paper we will summarise the experience in supporting LHC experiments and the transition of our infrastructure from static monolithic systems to flexible components providing a more coherent environment with pluggable protocols, tuneable QoS, sharing capabilities and fine grained ACLs management while continuing to guarantee dependable and robust services.
Azad, Ariful; Pavlopoulos, Georgios A.; Ouzounis, Christos A.; ...
2018-01-05
Biological networks capture structural or functional properties of relevant entities such as molecules, proteins or genes. Characteristic examples are gene expression networks or protein–protein interaction networks, which hold information about functional affinities or structural similarities. Such networks have been expanding in size due to increasing scale and abundance of biological data. While various clustering algorithms have been proposed to find highly connected regions, Markov Clustering (MCL) has been one of the most successful approaches to cluster sequence similarity or expression networks. Despite its popularity, MCL’s scalability to cluster large datasets still remains a bottleneck due to high running times andmore » memory demands. In this paper, we present High-performance MCL (HipMCL), a parallel implementation of the original MCL algorithm that can run on distributed-memory computers. We show that HipMCL can efficiently utilize 2000 compute nodes and cluster a network of ~70 million nodes with ~68 billion edges in ~2.4 h. By exploiting distributed-memory environments, HipMCL clusters large-scale networks several orders of magnitude faster than MCL and enables clustering of even bigger networks. Finally, HipMCL is based on MPI and OpenMP and is freely available under a modified BSD license.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Azad, Ariful; Pavlopoulos, Georgios A.; Ouzounis, Christos A.
Biological networks capture structural or functional properties of relevant entities such as molecules, proteins or genes. Characteristic examples are gene expression networks or protein–protein interaction networks, which hold information about functional affinities or structural similarities. Such networks have been expanding in size due to increasing scale and abundance of biological data. While various clustering algorithms have been proposed to find highly connected regions, Markov Clustering (MCL) has been one of the most successful approaches to cluster sequence similarity or expression networks. Despite its popularity, MCL’s scalability to cluster large datasets still remains a bottleneck due to high running times andmore » memory demands. In this paper, we present High-performance MCL (HipMCL), a parallel implementation of the original MCL algorithm that can run on distributed-memory computers. We show that HipMCL can efficiently utilize 2000 compute nodes and cluster a network of ~70 million nodes with ~68 billion edges in ~2.4 h. By exploiting distributed-memory environments, HipMCL clusters large-scale networks several orders of magnitude faster than MCL and enables clustering of even bigger networks. Finally, HipMCL is based on MPI and OpenMP and is freely available under a modified BSD license.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hoye, Robert L. Z., E-mail: rlzh2@cam.ac.uk, E-mail: jld35@cam.ac.uk; Ievskaya, Yulia; MacManus-Driscoll, Judith L., E-mail: rlzh2@cam.ac.uk, E-mail: jld35@cam.ac.uk
2015-02-01
Electrochemically deposited Cu{sub 2}O solar cells are receiving growing attention owing to a recent doubling in efficiency. This was enabled by the controlled chemical environment used in depositing doped ZnO layers by atomic layer deposition, which is not well suited to large-scale industrial production. While open air fabrication with atmospheric pressure spatial atomic layer deposition overcomes this limitation, we find that this approach is limited by an inability to remove the detrimental CuO layer that forms on the Cu{sub 2}O surface. Herein, we propose strategies for achieving efficiencies in atmospherically processed cells that are equivalent to the high values achievedmore » in vacuum processed cells.« less
NASA Astrophysics Data System (ADS)
Doulamis, A.; Doulamis, N.; Ioannidis, C.; Chrysouli, C.; Grammalidis, N.; Dimitropoulos, K.; Potsiou, C.; Stathopoulou, E.-K.; Ioannides, M.
2015-08-01
Outdoor large-scale cultural sites are mostly sensitive to environmental, natural and human made factors, implying an imminent need for a spatio-temporal assessment to identify regions of potential cultural interest (material degradation, structuring, conservation). On the other hand, in Cultural Heritage research quite different actors are involved (archaeologists, curators, conservators, simple users) each of diverse needs. All these statements advocate that a 5D modelling (3D geometry plus time plus levels of details) is ideally required for preservation and assessment of outdoor large scale cultural sites, which is currently implemented as a simple aggregation of 3D digital models at different time and levels of details. The main bottleneck of such an approach is its complexity, making 5D modelling impossible to be validated in real life conditions. In this paper, a cost effective and affordable framework for 5D modelling is proposed based on a spatial-temporal dependent aggregation of 3D digital models, by incorporating a predictive assessment procedure to indicate which regions (surfaces) of an object should be reconstructed at higher levels of details at next time instances and which at lower ones. In this way, dynamic change history maps are created, indicating spatial probabilities of regions needed further 3D modelling at forthcoming instances. Using these maps, predictive assessment can be made, that is, to localize surfaces within the objects where a high accuracy reconstruction process needs to be activated at the forthcoming time instances. The proposed 5D Digital Cultural Heritage Model (5D-DCHM) is implemented using open interoperable standards based on the CityGML framework, which also allows the description of additional semantic metadata information. Visualization aspects are also supported to allow easy manipulation, interaction and representation of the 5D-DCHM geometry and the respective semantic information. The open source 3DCityDB incorporating a PostgreSQL geo-database is used to manage and manipulate 3D data and their semantics.
Quantifying large-scale historical formation of accommodation in the Mississippi Delta
Morton, Robert A.; Bernier, Julie C.; Kelso, Kyle W.; Barras, John A.
2010-01-01
Large volumes of new accommodation have formed within the Mississippi Delta plain since the mid-1950s in association with rapid conversion of coastal wetlands to open water. The three-dimensional aspects and processes responsible for accommodation formation were quantified by comparing surface elevations, water depths, and vertical displacements of stratigraphic contacts that were correlated between short sediment cores. Integration of data from remotely sensed images, sediment cores, and water-depth surveys at 10 geologically diverse areas in the delta plain provided a basis for estimating the total volume of accommodation formed by interior-wetland subsidence and subsequent erosion. Results indicate that at most of the study areas subsidence was a greater contributor than erosion to the formation of accommodation associated with wetland loss. Tens of millions of cubic meters of accommodation formed rapidly at each of the large open-water bodies that were formerly continuous interior delta-plain marsh. Together the individual study areas account for more than 440 × 106 × m3 of new accommodation that formed as holes in the Mississippi River delta-plain fabric between 1956 and 2004. This large volume provides an estimate of the new sediment that would be needed just at the study areas to restore the delta-plain wetlands to their pre-1956 areal extent and elevations.
Simulating large atmospheric phase screens using a woofer-tweeter algorithm.
Buscher, David F
2016-10-03
We describe an algorithm for simulating atmospheric wavefront perturbations over ranges of spatial and temporal scales spanning more than 4 orders of magnitude. An open-source implementation of the algorithm written in Python can simulate the evolution of the perturbations more than an order-of-magnitude faster than real time. Testing of the implementation using metrics appropriate to adaptive optics systems and long-baseline interferometers show accuracies at the few percent level or better.
Remote Sensing of Salinity and Overview of Results from Aquarius
NASA Technical Reports Server (NTRS)
Le Vine, D. M.; Dinnat, E. P.; Meissner, T.; Wentz, F.; Yueh, S. H.; Lagerloef, G. S. E.
2015-01-01
Aquarius is a combined active/passive microwave (L-band) instrument designed to map the salinity of global oceans from space. The specific goal of Aquarius is to monitor the seasonal and interannual variation of the large scale features of the sea surface salinity (SSS) field of the open ocean (i.e. away from land). The instrumentation has been designed to provide monthly maps with a spatial resolution of 150 km and an accuracy of 0.2 psu
Cloud-Scale Numerical Modeling of the Arctic Boundary Layer
NASA Technical Reports Server (NTRS)
Krueger, Steven K.
1998-01-01
The interactions between sea ice, open ocean, atmospheric radiation, and clouds over the Arctic Ocean exert a strong influence on global climate. Uncertainties in the formulation of interactive air-sea-ice processes in global climate models (GCMs) result in large differences between the Arctic, and global, climates simulated by different models. Arctic stratus clouds are not well-simulated by GCMs, yet exert a strong influence on the surface energy budget of the Arctic. Leads (channels of open water in sea ice) have significant impacts on the large-scale budgets during the Arctic winter, when they contribute about 50 percent of the surface fluxes over the Arctic Ocean, but cover only 1 to 2 percent of its area. Convective plumes generated by wide leads may penetrate the surface inversion and produce condensate that spreads up to 250 km downwind of the lead, and may significantly affect the longwave radiative fluxes at the surface and thereby the sea ice thickness. The effects of leads and boundary layer clouds must be accurately represented in climate models to allow possible feedbacks between them and the sea ice thickness. The FIRE III Arctic boundary layer clouds field program, in conjunction with the SHEBA ice camp and the ARM North Slope of Alaska and Adjacent Arctic Ocean site, will offer an unprecedented opportunity to greatly improve our ability to parameterize the important effects of leads and boundary layer clouds in GCMs.
Staghorn: An Automated Large-Scale Distributed System Analysis Platform
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gabert, Kasimir; Burns, Ian; Elliott, Steven
2016-09-01
Conducting experiments on large-scale distributed computing systems is becoming significantly easier with the assistance of emulation. Researchers can now create a model of a distributed computing environment and then generate a virtual, laboratory copy of the entire system composed of potentially thousands of virtual machines, switches, and software. The use of real software, running at clock rate in full virtual machines, allows experiments to produce meaningful results without necessitating a full understanding of all model components. However, the ability to inspect and modify elements within these models is bound by the limitation that such modifications must compete with the model,more » either running in or alongside it. This inhibits entire classes of analyses from being conducted upon these models. We developed a mechanism to snapshot an entire emulation-based model as it is running. This allows us to \\freeze time" and subsequently fork execution, replay execution, modify arbitrary parts of the model, or deeply explore the model. This snapshot includes capturing packets in transit and other input/output state along with the running virtual machines. We were able to build this system in Linux using Open vSwitch and Kernel Virtual Machines on top of Sandia's emulation platform Firewheel. This primitive opens the door to numerous subsequent analyses on models, including state space exploration, debugging distributed systems, performance optimizations, improved training environments, and improved experiment repeatability.« less
High-yield production of graphene by liquid-phase exfoliation of graphite.
Hernandez, Yenny; Nicolosi, Valeria; Lotya, Mustafa; Blighe, Fiona M; Sun, Zhenyu; De, Sukanta; McGovern, I T; Holland, Brendan; Byrne, Michele; Gun'Ko, Yurii K; Boland, John J; Niraj, Peter; Duesberg, Georg; Krishnamurthy, Satheesh; Goodhue, Robbie; Hutchison, John; Scardaci, Vittorio; Ferrari, Andrea C; Coleman, Jonathan N
2008-09-01
Fully exploiting the properties of graphene will require a method for the mass production of this remarkable material. Two main routes are possible: large-scale growth or large-scale exfoliation. Here, we demonstrate graphene dispersions with concentrations up to approximately 0.01 mg ml(-1), produced by dispersion and exfoliation of graphite in organic solvents such as N-methyl-pyrrolidone. This is possible because the energy required to exfoliate graphene is balanced by the solvent-graphene interaction for solvents whose surface energies match that of graphene. We confirm the presence of individual graphene sheets by Raman spectroscopy, transmission electron microscopy and electron diffraction. Our method results in a monolayer yield of approximately 1 wt%, which could potentially be improved to 7-12 wt% with further processing. The absence of defects or oxides is confirmed by X-ray photoelectron, infrared and Raman spectroscopies. We are able to produce semi-transparent conducting films and conducting composites. Solution processing of graphene opens up a range of potential large-area applications, from device and sensor fabrication to liquid-phase chemistry.
The PREP pipeline: standardized preprocessing for large-scale EEG analysis.
Bigdely-Shamlo, Nima; Mullen, Tim; Kothe, Christian; Su, Kyung-Min; Robbins, Kay A
2015-01-01
The technology to collect brain imaging and physiological measures has become portable and ubiquitous, opening the possibility of large-scale analysis of real-world human imaging. By its nature, such data is large and complex, making automated processing essential. This paper shows how lack of attention to the very early stages of an EEG preprocessing pipeline can reduce the signal-to-noise ratio and introduce unwanted artifacts into the data, particularly for computations done in single precision. We demonstrate that ordinary average referencing improves the signal-to-noise ratio, but that noisy channels can contaminate the results. We also show that identification of noisy channels depends on the reference and examine the complex interaction of filtering, noisy channel identification, and referencing. We introduce a multi-stage robust referencing scheme to deal with the noisy channel-reference interaction. We propose a standardized early-stage EEG processing pipeline (PREP) and discuss the application of the pipeline to more than 600 EEG datasets. The pipeline includes an automatically generated report for each dataset processed. Users can download the PREP pipeline as a freely available MATLAB library from http://eegstudy.org/prepcode.
Pilot study of large-scale production of mutant pigs by ENU mutagenesis
Hai, Tang; Cao, Chunwei; Shang, Haitao; Guo, Weiwei; Mu, Yanshuang; Yang, Shulin; Zhang, Ying; Zheng, Qiantao; Zhang, Tao; Wang, Xianlong; Liu, Yu; Kong, Qingran; Li, Kui; Wang, Dayu; Qi, Meng; Hong, Qianlong; Zhang, Rui; Wang, Xiupeng; Jia, Qitao; Wang, Xiao; Qin, Guosong; Li, Yongshun; Luo, Ailing; Jin, Weiwu; Yao, Jing; Huang, Jiaojiao; Zhang, Hongyong; Li, Menghua; Xie, Xiangmo; Zheng, Xuejuan; Guo, Kenan; Wang, Qinghua; Zhang, Shibin; Li, Liang; Xie, Fei; Zhang, Yu; Weng, Xiaogang; Yin, Zhi; Hu, Kui; Cong, Yimei; Zheng, Peng; Zou, Hailong; Xin, Leilei; Xia, Jihan; Ruan, Jinxue; Li, Hegang; Zhao, Weiming; Yuan, Jing; Liu, Zizhan; Gu, Weiwang; Li, Ming; Wang, Yong; Wang, Hongmei; Yang, Shiming; Liu, Zhonghua; Wei, Hong; Zhao, Jianguo; Zhou, Qi; Meng, Anming
2017-01-01
N-ethyl-N-nitrosourea (ENU) mutagenesis is a powerful tool to generate mutants on a large scale efficiently, and to discover genes with novel functions at the whole-genome level in Caenorhabditis elegans, flies, zebrafish and mice, but it has never been tried in large model animals. We describe a successful systematic three-generation ENU mutagenesis screening in pigs with the establishment of the Chinese Swine Mutagenesis Consortium. A total of 6,770 G1 and 6,800 G3 pigs were screened, 36 dominant and 91 recessive novel pig families with various phenotypes were established. The causative mutations in 10 mutant families were further mapped. As examples, the mutation of SOX10 (R109W) in pig causes inner ear malfunctions and mimics human Mondini dysplasia, and upregulated expression of FBXO32 is associated with congenital splay legs. This study demonstrates the feasibility of artificial random mutagenesis in pigs and opens an avenue for generating a reservoir of mutants for agricultural production and biomedical research. DOI: http://dx.doi.org/10.7554/eLife.26248.001 PMID:28639938
NASA Astrophysics Data System (ADS)
Schartmann, M.; Meisenheimer, K.; Klahr, H.; Camenzind, M.; Wolf, S.; Henning, Th.
Recently, the MID-infrared Interferometric instrument (MIDI) at the VLTI has shown that dust tori in the two nearby Seyfert galaxies NGC 1068 and the Circinus galaxy are geometrically thick and can be well described by a thin, warm central disk, surrounded by a colder and fluffy torus component. By carrying out hydrodynamical simulations with the help of the TRAMP code \\citep{schartmann_Klahr_99}, we follow the evolution of a young nuclear star cluster in terms of discrete mass-loss and energy injection from stellar processes. This naturally leads to a filamentary large scale torus component, where cold gas is able to flow radially inwards. The filaments open out into a dense and very turbulent disk structure. In a post-processing step, we calculate observable quantities like spectral energy distributions or images with the help of the 3D radiative transfer code MC3D \\citep{schartmann_Wolf_03}. Good agreement is found in comparisons with data due to the existence of almost dust-free lines of sight through the large scale component and the large column densities caused by the dense disk.
Drive-by large-region acoustic noise-source mapping via sparse beamforming tomography.
Tuna, Cagdas; Zhao, Shengkui; Nguyen, Thi Ngoc Tho; Jones, Douglas L
2016-10-01
Environmental noise is a risk factor for human physical and mental health, demanding an efficient large-scale noise-monitoring scheme. The current technology, however, involves extensive sound pressure level (SPL) measurements at a dense grid of locations, making it impractical on a city-wide scale. This paper presents an alternative approach using a microphone array mounted on a moving vehicle to generate two-dimensional acoustic tomographic maps that yield the locations and SPLs of the noise-sources sparsely distributed in the neighborhood traveled by the vehicle. The far-field frequency-domain delay-and-sum beamforming output power values computed at multiple locations as the vehicle drives by are used as tomographic measurements. The proposed method is tested with acoustic data collected by driving an electric vehicle with a rooftop-mounted microphone array along a straight road next to a large open field, on which various pre-recorded noise-sources were produced by a loudspeaker at different locations. The accuracy of the tomographic imaging results demonstrates the promise of this approach for rapid, low-cost environmental noise-monitoring.
Allan deviation analysis of financial return series
NASA Astrophysics Data System (ADS)
Hernández-Pérez, R.
2012-05-01
We perform a scaling analysis for the return series of different financial assets applying the Allan deviation (ADEV), which is used in the time and frequency metrology to characterize quantitatively the stability of frequency standards since it has demonstrated to be a robust quantity to analyze fluctuations of non-stationary time series for different observation intervals. The data used are opening price daily series for assets from different markets during a time span of around ten years. We found that the ADEV results for the return series at short scales resemble those expected for an uncorrelated series, consistent with the efficient market hypothesis. On the other hand, the ADEV results for absolute return series for short scales (first one or two decades) decrease following approximately a scaling relation up to a point that is different for almost each asset, after which the ADEV deviates from scaling, which suggests that the presence of clustering, long-range dependence and non-stationarity signatures in the series drive the results for large observation intervals.
DupTree: a program for large-scale phylogenetic analyses using gene tree parsimony.
Wehe, André; Bansal, Mukul S; Burleigh, J Gordon; Eulenstein, Oliver
2008-07-01
DupTree is a new software program for inferring rooted species trees from collections of gene trees using the gene tree parsimony approach. The program implements a novel algorithm that significantly improves upon the run time of standard search heuristics for gene tree parsimony, and enables the first truly genome-scale phylogenetic analyses. In addition, DupTree allows users to examine alternate rootings and to weight the reconciliation costs for gene trees. DupTree is an open source project written in C++. DupTree for Mac OS X, Windows, and Linux along with a sample dataset and an on-line manual are available at http://genome.cs.iastate.edu/CBL/DupTree
Ohue, Masahito; Shimoda, Takehiro; Suzuki, Shuji; Matsuzaki, Yuri; Ishida, Takashi; Akiyama, Yutaka
2014-11-15
The application of protein-protein docking in large-scale interactome analysis is a major challenge in structural bioinformatics and requires huge computing resources. In this work, we present MEGADOCK 4.0, an FFT-based docking software that makes extensive use of recent heterogeneous supercomputers and shows powerful, scalable performance of >97% strong scaling. MEGADOCK 4.0 is written in C++ with OpenMPI and NVIDIA CUDA 5.0 (or later) and is freely available to all academic and non-profit users at: http://www.bi.cs.titech.ac.jp/megadock. akiyama@cs.titech.ac.jp Supplementary data are available at Bioinformatics online. © The Author 2014. Published by Oxford University Press.
Pyrolysis of corn stalk biomass briquettes in a scaled-up microwave technology.
Salema, Arshad Adam; Afzal, Muhammad T; Bennamoun, Lyes
2017-06-01
Pyrolysis of corn stalk biomass briquettes was carried out in a developed microwave (MW) reactor supplied with 2.45GHz frequency using 3kW power generator. MW power and biomass loading were the key parameters investigated in this study. Highest bio-oil, biochar, and gas yield of 19.6%, 41.1%, and 54.0% was achieved at different process condition. In terms of quality, biochar exhibited good heating value (32MJ/kg) than bio-oil (2.47MJ/kg). Bio-oil was also characterised chemically using FTIR and GC-MS method. This work may open new dimension towards development of large-scale MW pyrolysis technology. Copyright © 2017 Elsevier Ltd. All rights reserved.
Price, Erika; Ottati, Victor; Wilson, Chase; Kim, Soyeon
2015-11-01
The present research conceptualizes open-minded cognition as a cognitive style that influences how individuals select and process information. An open-minded cognitive style is marked by willingness to consider a variety of intellectual perspectives, values, opinions, or beliefs-even those that contradict the individual's opinion. An individual's level of cognitive openness is expected to vary across domains (such as politics and religion). Four studies develop and validate a novel measure of open-minded cognition, as well as two domain-specific measures of religious and political open-minded cognition. Exploratory and confirmatory factor analysis (controlling for acquiescence bias) are used to develop the scales in Studies 1 to 3. Study 4 demonstrates that these scales possess convergent and discriminant validity. Study 5 demonstrates the scale's unique predictive validity using the outcome of Empathic Concern (Davis, 1980). Study 6 demonstrates the scale's unique predictive validity using the outcomes of warmth toward racial, religious, and sexual minorities. © 2015 by the Society for Personality and Social Psychology, Inc.
NASA Astrophysics Data System (ADS)
Ercolano, Barbara; Weber, Michael L.; Owen, James E.
2018-01-01
Circumstellar discs with large dust depleted cavities and vigorous accretion on to the central star are often considered signposts for (multiple) giant planet formation. In this Letter, we show that X-ray photoevaporation operating in discs with modest (factors 3-10) gas-phase depletion of carbon and oxygen at large radii ( > 15 au) yields the inner radius and accretion rates for most of the observed discs, without the need to invoke giant planet formation. We present one-dimensional viscous evolution models of discs affected by X-ray photoevaporation assuming moderate gas-phase depletion of carbon and oxygen, well within the range reported by recent observations. Our models use a simplified prescription for scaling the X-ray photoevaporation rates and profiles at different metallicity, and our quantitative result depends on this scaling. While more rigorous hydrodynamical modelling of mass-loss profiles at low metallicities is required to constrain the observational parameter space that can be explained by our models, the general conclusion that metal sequestering at large radii may be responsible for the observed diversity of transition discs is shown to be robust. Gap opening by giant planet formation may still be responsible for a number of observed transition discs with large cavities and very high accretion rate.